Search Results

Search found 15129 results on 606 pages for 'orientation changes'.

Page 352/606 | < Previous Page | 348 349 350 351 352 353 354 355 356 357 358 359  | Next Page >

  • Determining where a file is being cached

    - by RoadieRich
    I've got a debian webserver running apache, from which I regularly download an executable using IE 8 on an XP virtual machine (over LAN). However, I realised that somewhere along the line that I've been repeatedly running the same version (and wondering why my changes weren't being displayed). A Ctrl+F5 in IE will let me download the new version (although the page is always updated with a simple F5). This makes me suspect that the caching is happening in windows/IE, but I'm not certain. Wherever it's happening, is there an easy way to prevent it at the server? Eventually, we're hoping to offer the software to the entire company, and we'd like to avoid having to tell everyone to do a Ctrl+F5 every time there's an updated version.

    Read the article

  • Only show changed files while syncing from ext4 to NTFS

    - by qox
    I would like rsync to print modified and deleted files. The verbose option (-v) does print modified files but also the list of subdirectories, maybe because touched directories are considered modified. Since I sync a lot of files from a lot of subdirectories, it's impossible to see the actual changes. So, is there a way to not print directories using rsync ? Im not looking for grep -v "*/$" kind of answers since it would also exclude new directories. Command I am using: rsync -avh --delete /media/data/src /media/data/bkp And everytime it prints the list of all directories: src/dir1/ src/dir1/sdir1/ src/dir1/sdir2/ src/dir2/ ..... Thanks for your help. EDIT: Ok, after some intensive tests .. It doesn't print all directories when syncing from an ext4 partition to an ext4 and from NTFS to NTFS. It only does when syncing from ext4 to NTFS .. And options '-c' or '--omit-dir-times' don't change that.

    Read the article

  • Nginx config with try_files and rewrite : precedence?

    - by Penegal
    Good morning, everybody. Firstly, this question may have been already asked, but I searched ServerFault during about 15 minutes without finding it, so, if it was already asked, please accept my apologies. I'm trying to rationalize my Nginx server config, but I have a rather dumb question that I couldn't solve, even with extensive Web search, except if I totally f*cked my search. Here is the question : is try_files parsed before or after rewrite ? Asked differently, Do I have to put try_files after all rewrite directives, or is Nginx config parser smart enough to evaluate try_files after all relevant rewrite directives ? The link with the config rationalization is that the answer to this question will change the organisation of the config, ie if config file order of try_files and rewrite changes the config behaviour, it will force me to disperse my includes, some of them containing try_files and other ones containing rewrite, because I also have rewrite directly in nginx.conf. Hoping you can help me, Regards.

    Read the article

  • Clarify git stash for me in switching branches

    - by EmmyS
    I've been working on branch A. My work there is not finished, but I need to switch to branch B for a while. It looks like stash is the command to use. I've found a number of references showing how to use stash to save your changes, but I'm a bit confused. All of the references say something like, when you're ready to go back, just do git stash pop. They don't, however, tell me if I need to switch back to branch A before doing that, though. So, do I manually go back to branch A before running stash pop, or do I stay in branch B, and the actual act of running stash pop will send me back to branch A where I left off with it?

    Read the article

  • What are the design principles that promote testable code? (designing testable code vs driving design through tests)

    - by bot
    Most of the projects that I work on consider development and unit testing in isolation which makes writing unit tests at a later instance a nightmare. My objective is to keep testing in mind during the high level and low level design phases itself. I want to know if there are any well defined design principles that promote testable code. One such principle that I have come to understand recently is Dependency Inversion through Dependency injection and Inversion of Control. I have read that there is something known as SOLID. I want to understand if following the SOLID principles indirectly results in code that is easily testable? If not, are there any well-defined design principles that promote testable code? I am aware that there is something known as Test Driven Development. Although, I am more interested in designing code with testing in mind during the design phase itself rather than driving design through tests. I hope this makes sense. One more question related to this topic is whether it's alright to re-factor an existing product/project and make changes to code and design for the purpose of being able to write a unit test case for each module?

    Read the article

  • CI - How long is continous?

    - by Andy
    We currently are using CCNet as our continous integration server. Most projects check for changes every 30 seconds (the default) and if needed perform a build (unit tests, stylecop, fxcop, etc). We've gotten quite a few projects now, and the server spends most of its time near 100% cpu utilization. This has alarmed some of the development team, even though the server is responsive and builds are still about the same length of time they've always been. Its been suggested that we lower the check interval to about five minutes. To me that seems too long, and we risk people committing code and then going home for the weekend and now there's a broken build possibly holding up others. In response, the suggestion is that if someone needs to know the results they can force the build. But that seems to defeat the purpose of CI, as I thought it was supposed to be automated. My proposed solution is just to get another build server and split the builds amongst the servers. Am I thinking about this the wrong way, or is there a point where if integration isn't often enough you're not really doing CI anymore?

    Read the article

  • Sync 2 hard drives without being connected at the same time (Mac)

    - by Jef
    Hi, I've got a 2TB hard drive at home and one at work. I use my laptop to connect to them. I want these 2 hard drives to be exactly the same. But these hard drives will never be connected at the same time. Is this possible? It means that my laptop needs to track the changes AND copy all new files to its hard drive temporarily. My laptop should act as an intermediate between the 2 external hard drives. There are lots of syncing options if both drives are connected at the same time, but there seems to be no solution for my specific problem. Any ideas? FYI: I use a Mac.

    Read the article

  • Error 1069 the service did not start due to a logon failure

    - by Si.
    Our CruiseControl.NET service on Win2003 Server (VMWare Virtual) was recently changed from a service account to a user account to allow for a new part of our build process to work. The new user has "Log on as a service" rights, verified by checking Local Security Settings - Local Policies - User Rights Assignment, and the user password is set to never expire. The problem I'm facing is every time the service is restarted, I get the 1069 error as described in this questions subject. I have to go into the properties of the service (log on tab) and re-enter the password, even though it hasn't changed, and the user already has the appropriate rights. Once I enter the password apply the changes, a prompt appears telling me that the user has been granted log on as a service rights. The service will then start will no problems. Not a show stopper, but a pain none-the-less. Why isn't the password persisting with the service?

    Read the article

  • Edit write-protected files by breaking hard links

    - by Taymon
    A directory which I own and can write to contains hard links to files that I don't own and don't have write permission for. I want to open and edit these files in Emacs. When I save my changes, Emacs should rename the existing hard link by appending ~, then write my new version of the file as a new file owned by me. I was under the impression that Emacs could just do this (because of the way it does backups), but it's not working; when I save, it attempts to change the file's permissions in order to write to it (and fails because I don't own the file). How do I make this happen?

    Read the article

  • Issue with domain redirection

    - by phphunger
    I have a flash games website which is hosted at godaddy server. But because of heavy traffic on my site the godaddy server gets down. So I changed the hosting to Midphase server. The strange thing is i have created new name servers and new database in my mid phase server but still the web site is coming from the godaddy server. When i do any modifications from the godaddy ftp i am getting the modifications but when i do any modifications in the mid phase server. No changes are happening. One strange thing is the who.is is showing the new name servers and new server details but not getting the new servers contents. Can anyone help me in this regard?

    Read the article

  • Do not get 'Safely remove' option in Tray for USB storage devices.

    - by blitzkriegz
    I had done some tweaks in service settings (Disabled some as I thought it is not needed). Now, I am not getting the option to safely remove mass storage device when I click the icon on the system tray. I tried enabling some of the services, but it didn't help. Moreover I'm not very sure if this anomaly is because of my changes in services. Any idea how to make those 'safely remove' options appear when I click the USB icon on the tray. Right now nothing is happening when I click.

    Read the article

  • Replication with AK8

    - by Steve Tunstall
    Hello folks, This came up today and I want to make sure it's clear. Remember the "deferred update" I spoke about in my "Upgrade to AK8" entry just a bit ago? It's important to understand that this deferred update changes the way replication works. It is necessary that systems with the deferred update applied only replicate with other systems that have also had this deferred update applied. So if you apply it, your system can NOT replicate with ANY other system that has NOT had it applied, even if that other system is running AK8!!! Got it??? Remember, we do have a new version of the 2011 code for the older systems that do not want to upgrade to AK8. This 2011.1.8 code ALSO HAS this same deferred update in it. So, if you upgrade your system to AK8, and then apply the deferred update, and you have another system running either 2011.1.8 or AK8, you can replicate with them again once they apply the deferred update for multiple initiator groups. Yes, even if you're not using LUNs. Here is what it looks like if you try. It will fail.

    Read the article

  • Using the Coherence ConcurrentMap Interface (Locking API)

    - by jpurdy
    For many developers using Coherence, the first place they look for concurrency control is the com.tangosol.util.ConcurrentMap interface (part of the NamedCache interface). The ConcurrentMap interface includes methods for explicitly locking data. Despite the obvious appeal of a lock-based API, these methods should generally be avoided for a variety of reasons: They are very "chatty" in that they can't be bundled with other operations (such as get and put) and there are no collection-based versions of them. Locks do directly not impact mutating calls (including puts and entry processors), so all code must make explicit lock requests before modifying (or in some cases reading) cache entries. They require coordination of all code that may mutate the objects, including the need to lock at the same level of granularity (there is no built-in lock hierarchy and thus no concept of lock escalation). Even if all code is properly coordinated (or there's only one piece of code), failure during updates that may leave a collection of changes to a set of objects in a partially committed state. There is no concept of a read-only lock. In general, use of locking is highly discouraged for most applications. Instead, the use of entry processors provides a far more efficient approach, at the cost of some additional complexity.

    Read the article

  • Mirroring a drive in Debian

    - by James Willson
    I have a drive with data on it. I dont want to use RAID, instead I want to do hourly backups to a second drive. I basically want to mirror the data drive and resync every hour. It is inefficient to re-move the data each time so really I only want to move across what changes. I.e if I add a new file to the data drive only that file will be moved across. What tools are there for doing this on the command line? I used to use luckybackup on Ubuntu but now im on commandline debian.

    Read the article

  • How do you know which domain owns the hosting?

    - by BubbleStalker
    For example if I have 1) host adress 2) login 3) password, I am entering by SSH on ruby on rails hosting, then how can i be sured that this hosting belongs to a specific domain? for example how can I know if www.site.com - belongs to some specific hosting to which I have access. I am asking this because I have access to hosting of ruby on rails, and when i modify files, there is no changes, i've tried to use the files "script", "serv", "restart.txt" - by ssh: touch tmp/restart.txt ./serv restart script restart nothing of the above helped...and I don't know what to do, any ideas?

    Read the article

  • Editing remotely the PHP files on a Centos server

    - by Alex2012
    I have a intranet web server (Centos 6, Apache, PHP) to which I would like to give access to a developer. He will connect by remote desktop from Windows 7 to Ubuntu 12.4 and from here by SSH to /var/www/html folder where it has to create and edit the files. This solution was chosen because: - I could not make a remote desktop connection from Windows to Centos - The web developer need some editor for PHP files and is not allowed to install software on Windows 7 machine - it is more a test solution ( we are all learning to use Linux). When the developer is connected from Ubuntu to Centos by SSH (SFTP) he could save the changes only if on Centos the account used to connect has ownership to that folder. Can you please tell how can I give all required rights. I tried different solutions found on Internet but without to much success. Are there other way to connect to Centos server?

    Read the article

  • Writing good looking documentation in pdf format on a mac [closed]

    - by Matthew
    I'd like to write some documentation that will be written in sections (1.1, 1.2, 2.1, 2.2, 2.3, etc). I'd like to use something that will generate the table of contents in the end or at least have it "linked" up so that when I make changes everything is still accurate (if I merge section 2.1 and 2.2 or something). The end result needs to be in pdf. I don't want to pay $350 for Acrobat Pro either... I've got microsoft word, but every little thing is a pain (for example, having one background of a page be a different color than another). Any ideas on how to create quality documentation in a pdf format?

    Read the article

  • Release Notes for 5/31/2012

    Here are the notes for this week’s release: Pull Requests We’ve moved the pull request creation UI from a modal dialog to a full page to make room for more context on what you’re sending with your pull request. In this week’s changes, you’ll see the list of commits; look for diffs to appear in the coming weeks. Corrected several issues where the line that had comments associated was not being properly differentiated in the code snippet previews on pull request discussions. Corrected an issue where pull request line comments would not appear to be associated with the proper line of code. Fixed an issue where accepting pull requests was not working properly. Bug Fixes Fixed an issue where code snippets could not be inserted using Opera. Fixed an issue where Internet Explorer displayed an error message when browsing the site using https. Fixed an issue where the CAPTCHA was not displaying properly on the Contact User page. Fixed an issue where paging was not working properly in the “Recent Release in Followed Projects” sections for logged in users. Other Updated the activity stream to no longer auto-load new items when scrolling down. Added an explicit download link at the end of the download countdown in case certain browsers/scenarios don’t automatically start downloading. Have ideas on how to improve CodePlex? Visit our ideas page! Vote for your favorite ideas or submit a new one. Got Twitter? Follow us and keep apprised of the latest releases and service status at @codeplex.

    Read the article

  • Release Notes for 5/18/2012

    Here are the notes for this week’s release: Pull Requests We’ve added the ability to see the snippets of code where a user commented inline in the discussion of pull requests. You can also add another line comment directly from the discussion area, rather than navigating to the code diff viewer. Note that there’s currently a known issue where the line associated with the comment isn’t being properly differentiated for existing pull requests (the line in the middle of each diff preview should be bolded). Apologies for the inconvenience! As part of this work, we also took some time to clean up our diff viewer UI to remove the dots and introduce a new color scheme where green is used for added lines. Bug Fixes Fixed an issue affecting the ability to assign pull requests. Fixed an issue where managing various team resources for a project was not working in Chrome or Firefox. Fixed an issue where a project’s RSS subscribe dialog popped up in the wrong place. Fixed an issue where editing wiki anchor links would insert extra characters, resulting in broken links. Fixed an issue where project logos did not display correctly when browsing the site with https in Chrome or Firefox. Fixed an issue where users could encounter errors when deleting remote Git branches. Fixed an issue affecting the ability of fork collaborators to push changes to the fork. Fixed an issue where the advanced work item filters would not persist when navigating through result pages. Fixed an issue where the issue tracker notifications link was not clickable in Chrome. Fixed an issue where pull request comments with line breaks would not be formatted properly when viewing the pull request. Other We upgraded our Git servers to version 1.7.10.1. Have ideas on how to improve CodePlex? Visit our ideas page! Vote for your favorite ideas or submit a new one. Got Twitter? Follow us and keep apprised of the latest releases and service status at @codeplex.

    Read the article

  • How to get the finer resolution for asus eee pc 1000hd after installing windows7?

    - by vij
    Hi i have installed windows 7 on my 2 year old ASUS EEE PC 1000HD. The default Operating system it came with was windows xp. Now the system is working fine but there are some display problems which i am going to state now: Firstly, during booting windows 7 logo is missing. There are no options for resolution changes. The only option possible is 800x600. The boot time is quite long when compared to xp. The graphics driver used in this system is STANDARD VGA GRAPHICS ADAPTER. The processor is intel celeron with 1GB RAM. I think with these requirements it is possible to get windows 7 with finer resolution. But why these problems are occurring?I even tried to install graphics driver provided by intel for this system. But i ve got the error message "this system doesn't meet the necessary requirements". So how to overcome this problem??

    Read the article

  • Ubuntu server 9.10 freezes up after ~10 minutes - unconventional

    - by L33tCh
    There is a topic by essentially the same name, but my answer leads to another question which I would like answered so... My case is as follows: I connect through SSH to my server... no problems for many months on Ubuntu Server 9.04. Eventually I decided to updgrade and straight away I get the 10 minutes freeze problem. I went through all the possible fixes (many can be found here but my quick fix, the only that worked, leaves me with no real answer as to why and how to properly correct it... As ridiculous as it sounds... while I leave a keyboard plugged in... it runs without flaw. Can anyone make sense of this? For the sake of completeness, (although none of the following made any difference upon subsequent completion), some of the permanent changes I have made while trying to correct the issue are: replacing samba config file with 9.04 backup, booting with noapic and modifying sysctl.conf, just uncommenting additional settings for enhanced security.

    Read the article

  • Why would an ext3 filsystem be rolled back on a Debian VM running in VirtualBox after loss of power to the host

    - by Sevas
    A Debian Virtual machine runs as a Guest VirtualBox VM. It's filesystem is EXT3. The host system loses power and after booting up the host system and guest VM, I find that the VM's filesystem has been rolled back to a previous state, losing changes made to the filesystem some time before losing power. The operations that were rolled back had been fully completed before the loss of power (files fully copied, file handles closed, etc.), but it's possible and even likely that other write operations were occuring on the VM at the point of the crash. So I am trying to figure out if it's the filesystem recovery process that rolls back filesystem operations after encountering corruption post power loss, or is it possibly related to VirtualBox and the way it ignores flush requests for performance gains by default (discussed here) Are there any other factors that would result in the filesystem being rolled back after losing power?

    Read the article

  • Desktop Properties Background selection problems

    - by Gazza
    From the Desktop of my XP Home Edn SP3 system, under the Right-Click Menu Properties, the [Desktop] tab, of the 'Display Properties' window shows a selection list for the Desktop's Background image. For me this list has become frozen and its slider (at right) will not move. The [Browse] button and the 'Position' dropdown list (whether the selected image is centred, stretched, etc) are also greyed out. I can still use both the Color dropdown list at the right of the Background list and the [Customize Desktop] button below it, and make any changes as desired. What's gone wrong. I'm guessing maybe a registery setting or I've deleted a file somewhere, but haven't found the cause as yet. Any suggestions?? Thanks.

    Read the article

  • How should I implement Transaction database EJB 3.0

    - by JamesBoyZ
    In the CustomerTransactions entity, I have the following field to record what the customer bought: @ManyToMany private List<Item> listOfItemsBought; When I think more about this field, there's a chance it may not work because merchants are allowed to change item's information (e.g. price, discount, etc...). Hence, this field will not be able to record what the customer actually bought when the transaction occurred. At the moment, I can only think of 2 ways to make it work. I will record the transaction details into a String field. I feel that this way would be messy if I need to extract some information about the transaction later on. Whenever the merchant changes an item's information, I will not update directly to that item's fields. Instead, I will create another new item with all the new information and keep the old item untouched. I feel that this way is better because I can easily extract information about the transaction later on. However, the bad side is that my Item table may contain a lot of rows. I'd be very grateful if someone could give me an advice on how I should tackle this problem. UPDATE: I'd like to add more information about the current design. public class Customer implements Serializable { @OneToMany private List<CustomerTransactions> listOfTransactions; } public class CustomerTransactions implements Serializable { @ManyToMany private List<Item> listOfItemsBought; } public class Merchant implements Serializable { @OneToMany private List<Item> listOfSellingItems; }

    Read the article

  • Lightweight alternative to R for RHEL?

    - by Eric Rath
    I want to use R for some statistical analysis of logfile information, but found that even the "limited" R-core RPM has a lot of dependencies not already installed. I don't want to install so many packages for a peripheral need. Are there lightweight alternatives for simple statistical analysis on RHEL 6? I have an R script that accepts on stdin a large set of values -- one value per line -- and prints out the min, max, mean, median, 95th percentile, and standard deviation. For more context, I'm using grep and awk to find GET requests for a particular path in our webserver log files, get the response times, and calculate the metrics listed above in order to measure the impact on performance of changes to a web application. I don't need any graphing capabilities, just simple computation. Is there something I've overlooked?

    Read the article

< Previous Page | 348 349 350 351 352 353 354 355 356 357 358 359  | Next Page >