Search Results

Search found 15605 results on 625 pages for 'cached exchange mode'.

Page 55/625 | < Previous Page | 51 52 53 54 55 56 57 58 59 60 61 62  | Next Page >

  • Google Chrome in Incognito Mode still logging visits history

    - by casey_miller
    I am using Incognito Mode and today I have noticed that when I frequently visit some site in that mode browser logs it and even on not incognito mode it autofills in address bar making Incognito mode useless in my case. Another fun thing is that the item is not in History so I can't manually remove it. Couldn't find anything in Settings. Why is this happening? BTW, instant search is disabled in my case. Using the latest version to date. Extensions installed List item Send to Google Docs Google Translate Eye Dropper Delicious Readabiliy Pagespeed

    Read the article

  • Nginx proxy upstream cached?

    - by Julian H. Lam
    Attempting to resolve an issue that's been annoying me for a bit. I've distilled the symptoms into a set of reproducible steps: I have two sites, siteA, and siteB. They are both Node.js applications running on different ports (for the sake of example, 4567 and 4568) Both applications have their own file in sites_available (plus a symlink from sites_enabled), which contain the directives proxy_pass http://node_siteA/ and proxy_pass http://node_siteB/ respectively, inside of a location block. They also each have an upstream block (defined globally?): upstream node_siteA { upstream node_siteB { server 127.0.0.1:4567; server 127.0.0.1:4568; } } Site A and Site B have nothing to do with each other. Yes, I am restarting (reloading, actually) nginx every time I make a change. If I take down site B and attempt to access it via the web, I am served site A. Why is this? Thoughts Other times, when I create a new Site C, for example, nginx refuses to show me anything except "Welcome to nginx!" for ~5 minutes. This suggests a resolver timeout, perhaps? When I access Site B after its config has been deleted, and it sends me to Site A, this sounds like nginx sending me to servers in a round-robin fashion...

    Read the article

  • Why is my ethernet interface in promiscuous mode

    - by nhed
    I read that seeing a flag of M in netstat -i is the way to tell which of your interfaces is in promiscuous mode I run it and I see that eth1 is in promiscuous mode $ netstat -i Kernel Interface table Iface MTU Met RX-OK RX-ERR RX-DRP RX-OVR TX-OK TX-ERR TX-DRP TX-OVR Flg eth1 1500 0 1770161198 0 0 0 57446481 0 0 0 BMRU lo 16436 0 97501566 0 0 0 97501566 0 0 0 LRU This seems to be the case on all the machines I checked (All Centos6.0, both virtual and physical), any idea why ethernet devices would be in such a mode unless someone was running any pcap based app (sudo lsof | grep pcap shows nothing)? I did not see any mention of promiscuous in any of the config files (sudo grep -r promis /etc) Any ideas what puts the interface into that mode and why? p.s. most of the posts I see seem to be security related, this is not that

    Read the article

  • Windows 7 - XP Mode - Apache

    - by Howard
    I've setup Virtual PC and XP Mode on my Windows 7 Pro. Using Apache 2.0.52 I have no problems having my website up and running on the Windows 7 machine. But Under VPC/XP Mode the best I can do is Localhost mode. What do I need to do to enable http connections? I need the XP Mode as besides the website I also run a Web BBS and a Dos based (via telnet) BBS. Some of the apps in the Dos BBS just won't work under 64 bit, no matter what setting (capability) are used. Thanks in advance...

    Read the article

  • what to use instead of laptop-mode?

    - by playcat
    hello, i have ubuntu 10.10 64bit on hp 6735s (turion processor). it overheats, and i'm forced to use turion power control in order to keep core temperature to a reasonable level. one more measure that i use is putting my processors to conservative mode. that way, i'm perfectly happy with its performance, and heat is where it should be. however, after my latest upgrade, something happened - cores are back to ondemand by default, and i'm not sure if turionpowercontrol is working any more (ps axu | grep urion shows no process). in addition, i read somewhere that laptop-mode uses hdd spindown for preserving data/energy, and that hdds have only a limited amount of those spindowns, so laptop-mode usage can actually shorten the life of my hdd. i'm wondering if there is a good way to set my cores to automatically go to conservative mode? also, what's the good way to see what is the voltage my cores use? on windows i use cpuz tools. thx and sorry for the long explanation.

    Read the article

  • Cached css/javascript files on Sun Java System Web Server

    - by Derp
    I'm doing front-end web development in a Solaris 10 / Sun Java System Web Server 7.0U2 environment. I have noticed that changes to static css or javascript files often do not take effect immediately, whereas changes to static html files always do. My best guess is that a default setting in the web server causes it to cache certain file types in order to provide reasonable performance out of the box. I don't have the admin server running--I'll need to edit the config files by hand. What change(s) can I make so that all of my css and javascript edits take effect immediately? Thanks!

    Read the article

  • how to make vhdresizer work on XP Mode VHD files

    - by A_M
    Hi, I'm trying to shrink an Windows 7 XP Mode VHD file with little success. I've been trying to use VhdResizer. When I select my VHD file, it says "VhdExpand only supports fixed and dynamic VHD files". My XP Mode VHDs are dynamic files. Does anyone have any idea why it is failing? Failing that, does anyone have a process which I can use to shrink my XP mode VHD files? Thanks.

    Read the article

  • SATA Devices not showing up when in UEFI mode

    - by Dan Barzilay
    I'm trying to install Windows and the bios should be set to UEFI mode. The problem is that all SATA devices aren't showing up (shows as if there aren't any) so I can't boot from the installation CD (it's just not there). The weird thing is that when set to LEGACY mode they all show up.. SATA mode is set to AHCI and I'm on Lenovo Y510P. I have a Linux OS installed that is accessible only when BIOS is in LEGACY mode (otherwise the hard drive it's on is not available) I also tried reseting the BIOS settings which didn't help.. Comment please if more details needed Extra details: Computer model: Lenovo IdeaPad Y510P (not overcloacked) Installed Linux OS version: Linux 3.7-trunk-amd64 x86_64 Trying to install Windows: Windows 7 Ultimate 64bit BIOS Information: Vendor: LENOVO Version: 74CN26WW(V1.07) Update: Using user1608638 answer and suggestion of using the USB flash drive as the boot device instead of the CD/DVD method I succeeded in installing Windows 7! (Thanks alot user1608638)

    Read the article

  • Determining where a file is being cached

    - by RoadieRich
    I've got a debian webserver running apache, from which I regularly download an executable using IE 8 on an XP virtual machine (over LAN). However, I realised that somewhere along the line that I've been repeatedly running the same version (and wondering why my changes weren't being displayed). A Ctrl+F5 in IE will let me download the new version (although the page is always updated with a simple F5). This makes me suspect that the caching is happening in windows/IE, but I'm not certain. Wherever it's happening, is there an easy way to prevent it at the server? Eventually, we're hoping to offer the software to the entire company, and we'd like to avoid having to tell everyone to do a Ctrl+F5 every time there's an updated version.

    Read the article

  • Virtualbox in Headless mode

    - by ask
    I used the virtual machines in virtualbox in a "headless" mode instead of a GUI mode. what are the advantages of using it in a headless mode?? by headless does it mean that the server doesnt have a keyboard or monitor attached or does it mean that no window will "pop up" , denoting that it is ON(or any other status), when a virtual machine is worked with? what exactly does it mean? pls reply...

    Read the article

  • Go Directly to Desktop Mode in Windows 8 on Login (Without Installing Extra Software)

    - by Asian Angel
    A lot of people are unhappy with being forced to interact with the new Start Screen in Windows 8 first thing once they have logged into their system. But their is a quick and simple work-around to go directly to Desktop Mode that does not require installing extra software or making changes to your system. The first thing that you will need to do is make sure that the Desktop Tile is in the left uppermost position on the Start Screen as seen here. Once the tile has been moved to that position you will need to restart/reboot your system. Once your system has restarted and you are back at the Login Screen, type in your password but do NOT click on the Arrow Button or tap the Enter Key. Instead of tapping the Enter Key simply press down on it and hold it down until you see the regular desktop. Keep in mind that you may see the Start Screen become visible for just a short moment as it is being bypassed for the desktop. How to Use an Xbox 360 Controller On Your Windows PC Download the Official How-To Geek Trivia App for Windows 8 How to Banish Duplicate Photos with VisiPic

    Read the article

  • Unity stuck in 2D mode, Nvidia Quadro graphics "unknown", Nvidia-Current active but not in use

    - by Jordan Lund
    I've seen this problem reported under several questions, but I haven't been able to resolve any of it so I thought I'd bring it all in under one umbrella. I started a new job and was given a Dell Precision M6400 laptop with Nvidia Quadro FX 2700M graphics card. It had a previous version of Ubuntu on it, but nobody had any passwords for it so I wiped the drive and did a fresh install of 11.10 from scratch. I didn't do any updates during installation, preferring to do them after boot. Updates ran fine and the system works... except Unity is in 2D mode. System Settings - Additional Drivers shows that Nvidia-Current is active but not in use. System Settings - System Info shows Graphics Driver Unknown, Experience Standard Nvidia X Server Settings is installed and working, re-writing the xorg.conf did nothing. /usr/lib/nux/unity_support_test -p OpenGL vendor string: NVIDIA Corporation OpenGL renderer string: Quadro FX 2700M/PCI/SSE2 OpenGL version string: 3.3.0 NVIDIA 285.05.09 Not software rendered: yes Not blacklisted: yes GLX fbconfig: yes GLX texture from pixmap: yes GL npot or rect textures: yes GL vertex program: yes GL fragment program: yes GL vertex buffer object: yes GL framebuffer object: yes GL version is 1.4+: yes Unity 3D supported: yes One suggestion was to do a sudo apt-get --purge remove nvidia* and that resulted in a scrambled screen on boot and a non-bootable installation. Pressing the Delete key on boot allowed me to access the recovery console and do a sudo apt-get install nvidia-current, which brought me back to a working, bootable system. Another suggestion was to edit /etc/default/grub and change the line reading "GRUB_CMDLINE_LINUX_DEFAULT="quiet splash" to read "GRUB_CMDLINE_LINUX_DEFAULT="quiet splash vmalloc=192MB" thus allocating more video RAM. I did that, followed by a sudo update-grub and a re-boot. No change. Created a brand new standard user and logged on with that account, no change.

    Read the article

  • SQL SERVER – How to Set Variable and Use Variable in SQLCMD Mode

    - by Pinal Dave
    Here is the question which I received the other day on SQLAuthority Facebook page. Social media is a wonderful thing and I love the active conversation between blog readers and myself – actually I think social media adds lots of human factor to any conversation. Here is the question - “I am using sqlcmd in SSMS – I am not sure how to declare variable and pass it, for example I have a database and it has table, how can I make the table variable dynamic and pass different value everytime?” Fantastic question, and here is its very simple answer. First of all, enable sqlcmd mode in SQL Server Management Studio as described in following image. Now in query editor type following SQL. :SETVAR DatabaseName “AdventureWorks2012″ :SETVAR SchemaName “Person” :SETVAR TableName “EmailAddress“ USE $(DatabaseName); SELECT * FROM $(SchemaName).$(TableName); Note that I have set the value of the database, schema and table as a sqlcmd variable and I am executing the query using the same parameters. Well, that was it, sqlcmd is a very simple language to master and it also aids in doing various tasks easily. If you have any other sqlcmd tips, please leave a comment and I will publish it with due credit. Reference: Pinal Dave (http://blog.sqlauthority.com) Filed under: PostADay, SQL, SQL Authority, SQL Query, SQL Server, SQL Server Management Studio, SQL Tips and Tricks, T SQL, Technology Tagged: sqlcmd

    Read the article

  • System in low-graphics mode

    - by Artem Moskalev
    I am totally new to Linux, and Ubuntu. I bought ASUS X53E Notebook, erased Windows and installed Ubuntu. First it worked ok. Then when I started working with it, i opened the terminal, entered sudo chmod 666 usr and then all the icons from the main panel disappeared + the whole system stopped responding. I decided to restart the system. When restarted, a message appears: the system is running in low graphics mode and below it: Your screen, graphics card and input device setting could not be detected correctly. You will need to configure it yourself. But the "OK" button is disabled and if I press any buttons nothing happens. If I enter Ctrl-Alt-F2 it opens the bash terminal. But there commands sudo or apt-get are not found and it says that permission denied if i try to enter any folder like cd /usr If I enter the su command it asks for the password I don't know. When encountering this problem first, I reinstalled the whole Ubuntu. but today it happened again just the same. What shall I do? maybe there is something wrong with the hardware? If I need to install another distribution of Linux could you recommend one? but I'd rather stick to Debian releases like Ubuntu. so how do I fix the problem? PS: Please give answers in simple terms because am a newbie so i don't know what goes where yet.

    Read the article

  • Solutions for "Maintenance Mode"

    - by Ka Lyse
    Given a web application running across 10+ servers, what techniques have you put in place for doing things like altering the state of your website so that you can implement certain features. For instance, you might want to: Restrict Logins/Disable Certain Features Turn Site to "Read Only" Turn Site to Single "Maintenance Mode" page. Doing any of the above is pretty trivial. You can throw a particular "flag" in an .ini file, or add a row/value to a site_options table in your database and just read that value and do the appropriate thing. But these solutions have their problems. Disadvantages/Problems For instance, if you use a file for your application, and you want to switch off a certain feature temporarily, then you need to update this file on all servers. So then you might want to look at running something like ZooKeeper, but you are probably overcomplicating things. So then, you might decide that you want to store these "feature" flags in a database. But then you are obviously adding unncessary queries to each page request. So you think to yourself, that you will throw memcached in to the mix and just cache the query. Then you just retrieve all of your "Features" from memcached and add a 2ms~ latency to your application on every page. So to get around this, you decide to use a two tier-cache system, whereby you use an inmemory cache on each machine, (like Apc/Redis etc). This would work, but then it gets complicated, because you would have to set the key/hash life to perhaps 60 seconds, so that when you purge/invalidate the memcached object storing your "Features" result, your on machine cache is prompt enough to get the the new states. What suggestions might you have? Keeping in mind that optimization/efficiency is the priority here.

    Read the article

  • "drm:[drm_crtc_helper_set_config] *ERROR* failed to set mode on [CRTC6]" during install

    - by Warez J. Coxtrong
    I have a desktop PC which I built a while back. It has an Athlon XP 2500+, 2.5GB RAM, and an Nvidia (PNY Verto) Geforce 6200. I removed the CD drive to add another hard drive so I install from a USB flash drive. When I try to install 12.04 it seems to work just fine. The GUI boots up and wifi even connects to my router. I go through the language screen, the partition screen, the keyboard screen, the location screen, and the import my Windows settings screen just fine. But as soon as I'm done with all that the next screen goes black and displays this message: drm:[drm_crtc_helper_set_config] *ERROR* failed to set mode on [CRTC6] Since the GUI starts just fine earlier during the install, I know I shouldn't be having any display problems, but I am. How do I get Ubuntu to install properly? I have four IDE hard drives, all Western Digital: a 250GB, a 160GB, a 40GB occupied by Windows XP, and a 120GB that I'm trying to use for Ubuntu. 3GB of the 120GB drive are formatted for swap. I have checked the MD5sum of install image file and it all matches up. d791352694374f1c478779f7f4447a3f ubuntu-12.04-desktop-i386.iso

    Read the article

  • Single Item Recovery

    One of the more obscure parts of Exchange Server is the Dumpster. This allows the recovery of one or more deleted emails, in much the same way as the recycle bin in Windows, even if the user has purged them. Although it is seen as an alternative to backup, or as a means to document retention, it really provides a separate function, undoing an accidental deletion. Elie explains how to configure deleted retention items and how to recover a purged item.

    Read the article

  • A Deep Dive into Transport Queues (Part 2)

    Johan Veldhuis completes his 'Deep Dive' by plunging even deeper into the mysteries of MS Exchange's Transport queues that are used to temporarily store messages which are waiting until they are passed through to the next stage, and explains how to change the way they work via configuration settings.

    Read the article

  • Implementing Cluster Continuous Replication, Part 3

    Cluster continuous replication (CCR) uses log shipping and failover to provide a more resilient email system with faster recovery. Once it is installed, a clustered server requires different management routines. These are done either with a GUI tool, The Failover Cluster Management Console, or the Exchange Management Shell. You can use Powershell as well for some tasks. Confused? Not for long, since Brien Posey is once more here to help.

    Read the article

  • Implementing Cluster Continuous Replication, Part 2

    Cluster continuous replication (CCR) helps to provide a more resilient email system with faster recovery. It was introduced in Microsoft Exchange Server 2007 and uses log shipping and failover. configuring Cluster Continuous Replication on a Windows Server 2008 requires different techniques to Windows Server 2003. Brien Posey explains all.

    Read the article

  • Problems setting up Davmail in Kubuntu 11.04

    - by jjesse
    Haven't had a chance to setup Davmail per the earlier discussion around connecting to my Exchange box w/ Kontact but tonight I had some free time. So I have Davmail installed and working I thinking, but the problem that I am running into is the icon doesn't not show up in the system tray. The documentation says right-click o the tray, but there are no icons to right click on. Any help would be greatly appreciated. Running Kubuntu 11.04 Thanks, Jonathn

    Read the article

  • A Deep Dive into Transport Queues - Part 1

    Submission queues? Poison message queues? Johan Veldhuis unlocks the mysteries of MS Exchange's Transport queues that used to temporarily store messages waiting until they are passed through to the next stage, and explains how to manage these queues.

    Read the article

  • AVAudioPlayer not unloading cached memory after each new allocation

    - by Rob
    I am seeing in Instruments that when I play a sound via the standard "AddMusic" example method that Apple provides, it allocates 32kb of memory via the prepareToPlay call (which references the AudioToolBox framework's Cache_DataSource::ReadBytes function) each time a new player is allocated (i.e. each time a different sound is played). However, that cached data never gets released. This obviously poses a huge problem if it doesn't get released and you have a lot of sound files to play, since it tends to keep allocating memory and eventually crashes if you have enough unique sound files (which I unfortunately do). Have any of you run across this or what am I doing wrong in my code? I've had this issue for a while now and it's really bugging me since my code is verbatim of what Apple's is (I think). How I call the function: - (void)playOnce:(NSString *)aSound { // Gets the file system path to the sound to play. NSString *soundFilePath = [[NSBundle mainBundle] pathForResource:aSound ofType:@"caf"]; // Converts the sound's file path to an NSURL object NSURL *soundURL = [[NSURL alloc] initFileURLWithPath: soundFilePath]; self.soundFileURL = soundURL; [soundURL release]; AVAudioPlayer * newAudio=[[AVAudioPlayer alloc] initWithContentsOfURL: soundFileURL error:nil]; self.theAudio = newAudio; // automatically retain audio and dealloc old file if new m4a file is loaded [newAudio release]; // release the audio safely // this is where the prior cached data never gets released [theAudio prepareToPlay]; // set it up and play [theAudio setNumberOfLoops:0]; [theAudio setVolume: volumeLevel]; [theAudio setDelegate: self]; [theAudio play]; } and then theAudio gets released in the dealloc method of course.

    Read the article

  • NHibernate correct way to reattach cached entity to different session

    - by Chris Marisic
    I'm using NHibernate to query a list of objects from my database. After I get the list of objects over them I iterate over the list of objects and apply a distance approximation algorithm to find the nearest object. I consider this function of getting the list of objects and apply the algorithm over them to be a heavy operation so I cache the object which I find from the algorithm in HttpRuntime.Cache. After this point whenever I'm given the supplied input again I can just directly pull the object from Cache instead of having to hit the database and traverse the list. My object is a complex object that has collections attached to it, inside the query where I return the full list of objects I don't bring back any of the sub collections eagerly so when I read my cached object I need lazy loading to work correctly to be able to display the object fully. Originally I tried using this to re-associate my cached object back to a new session _session.Lock(obj, LockMode.None); However when accessing the page concurrently from another instance I get the error Illegal attempt to associate a collection with two open sessions I then tried something different with _session.Merge(obj); However watching the output of this in NHProf shows that it is deleting and re-associating my object's contained collections with my object, which is not what I want although it seems to work fine. What is the correct way to do this? Neither of these seem to be right.

    Read the article

  • Keeping cached browser data inside ASP update panel textboxes/dropdowns for browser back click

    - by pmlevere
    I'm new in VB.net/asp and am running a VB web application in a visual database program called IronSpeed designer. I'm primarily using IronSpeed in this case for its login/role security features. I have a basic two page setup for this app. The user logs in then is taken to AccountEntry.aspx, they enter data into textboxes and select some dropdown values that are linked to a sql database, then they click "submit" to move to Results.aspx. On Results.aspx, the user can change data and then generate several types of reports (PDF, Excel, etc). I'm used to setting up ASP controls inside ASPContent areas, and in these areas if a user performs a browser back click the previously entered data will still be on the page for potential user modification. However in this web app, IronSpeed is setting up the page and asp controls inside an asp update panel. It appears inside an asp update panel, cached values can't be seen on a browser back click. In this case, it's important that the orginally entered values still be there for the user experience if the user advances to Results.aspx then clicks browser back to modify a value on AccountEntry.aspx. If I have to I'll setup Session Variables and disable browser clicking, but that is last resort. Is there any way to save cached data inside an asp update panel and have it there for a browser back click?

    Read the article

< Previous Page | 51 52 53 54 55 56 57 58 59 60 61 62  | Next Page >