Search Results

Search found 7799 results on 312 pages for 'changing'.

Page 149/312 | < Previous Page | 145 146 147 148 149 150 151 152 153 154 155 156  | Next Page >

  • rsync stuck with the --checksum option

    - by billc.cn
    I use back-in-time to backup my Linux installation. It serves as an advanced wrapper for the rsync command. Today I tried to add /var/log to the list of folders to be backed up and it caused some serious performance problems. The job seems to stuck on a particular file and the CPU usage of the rsync parent process reaches 100%. I then used lsof to see which file caused the problem and it seems to be the /var/log directory. I did some googling and some experiments with the different rsync options and found --checksum to be the offender. Without the parameter, an incremental backup finishes properly in minutes. With it, the process will stuck when rsync tries to sync a constantly changing log file. This kind of make sense, but it still seems to be a bug to me. Am I using the option correctly? Is there a workaround for this?

    Read the article

  • Vista Home Premium won't connect to network.

    - by Ice
    I deinstalled some software (e.g. some Toolbars for IE and Norton security suite, because the free trial period of 60 days is run out and so i want to switch to microsofts free security essentials). But now i cannot connect to the network again: Neither over LAN nor over WiFi. Neither with standard DHCP ON (sets a 169.254.x.y address) nor with fixed IP-Address. No error during changing the settings. Ping to the Gateway e.g. doesn't work and gives "generall failure". The DHCP-Client sets no IP-Address. There is allways only 'local access' on every network-connection. i ran out of ideas, please help. UPDATE per 2011-03-11: I followed each and every hint but nothing changed. I did a research on microsofts homepage through all the sites belonging to network but nothing helped. What can i do next?

    Read the article

  • Dialog boxes font is missing in Windows 7

    - by forgetso
    About 3 years ago I messed up my laptop with some virus removal software. It's obviously changed the registry settings and the text in dialogue boxes in the control panel, run box, etc. is the wrong size. See image:http://i.stack.imgur.com/SlFFU.jpg I've tried changing the MS Shell Dlg registry values to different fonts and restarting. I've made sure that I have the fonts SSERIFE.FON and SERIFE.FON on my system and that they are correctly referenced in the registry. Does anyone know how to fix this? I'd like to be able to see all of the options in the dialog boxes! Cheers

    Read the article

  • Why can't I change the icon being displayed for a desktop shortcut?

    - by Bill
    I wanted to change the icon showing for a shortcut on the desktop, which points to a .Net utility I created for myself. I rebuilt the exe after changing the icon used within the project. In Windows Explorer the new icon shows in all size variations for the exe. In the 'Change Icon' dialog from the Properties option, the correct icon is showing as present, yet when I Apply it remains with the old icon? Even when I delete the shortcut and create a new one pointing to the newly built exe the old icon somehow is assigned?

    Read the article

  • why does my google chat get blocked (by corporate firewall) somedays but not others?

    - by Peter
    I have noticed that some days I am able to chat while using Gmail, and other days I am not. It would make sense to me that I would either always be blocked, or never. But I can't figure out why it seems to change daily or weekly. Is Google constantly changing the URLs involved so that the censoring companies (they use websense where I work) have to play catch up? Or is there some other reason I'm missing? I am more interested in the technical reason it is might be happening than in an actual work around.

    Read the article

  • Are Uninterruptible Power Supply (UPS) at home really a worthwhile investment for home computing? [c

    - by Jian Lin
    Possible Duplicate: Is a UPS a good investment for a home machine? I have 2 Uninterruptible power supplies at home, but they seem to expire after 5 years or so. During this 5 years, it maybe saved the computer from turning off twice. But I wonder, I often save my documents every 10 minutes. So the most I can lose is 10 minute worth of work. Not to mention that, I may not be editing anything when the power is out. Also, when the UPS expires, one of them will give out beeping sound once in a while, probably signaling the changing of the battery. So I can either buy 2 new UPS or just buy 2 power outlets instead. 2 UPS, costing $160 to $250 roughly, can it be really worth it if it can save your computer from powering off for just 2 times? What are the alternatives?

    Read the article

  • The letter 'W' gets rendered oddly

    - by Jeff E
    It seems that with this particular font (it's one of the Ubuntu fonts; I can't tell which one exactly because Chrome's debugger doesn't tell me which font-family won the CSS battle) the letter 'W' (both cases), at the default zoom level and typically with smaller sizes, will render very oddly: (open the image in a new tab) The right-hand side of the letter is totally wrong. Changing the zoom levels will mitigate (but not eliminate) the issue. Does anyone know how to fix this? It doesn't happen on Firefox or IE so I assume it's not an issue with the font itself.

    Read the article

  • Why does my Microsoft Habu force me to unplug and replug in when booting up?

    - by jordsta
    I'm not too sure why, but every single time I boot my computer, I have to unplug and replug my semi-new Microsoft Habu, when I use Windows 7. Why does it do this? I've been through my BIOS, msconfig, and what-not.. It just seems to happen all the time. I've even looked in forums, with people resorting to changing their mouse because of the same issue.. It's stupid. Hopefully someone else can help... hopefully someone with experience with this mouse.

    Read the article

  • How to select all ActiveX objects in an area using a mouse in Excel?

    - by enderland
    Because of this problem with ActiveX objects changing size, I am not grouping my ActiveX objects in my Excel worksheet. Grouping them causes my solution hack to not work which is quite annoying. However, I often times want to be able to essentially use the mouse and select a region and then select all ActiveX objects contained in the region. This would also be useful for easily selecting objects to group them initially. Basically: Use mouse to select area Automatically select all ActiveX components in region I'm fine with a VBA solution if needed. How can I do this?

    Read the article

  • Why does Excel 2010 automatically reformat numbers?

    - by user179555
    Excel 2010 is automatically reformatting numbers/cells where a specific number format has already been applied. Whenever I access or modify the cell contents, Excel reformats by changing the number of decimal places to 2. I've already gone into file \ options \ advanced \ and made sure that the option to automatically add decimal places is unselected, but Excel still automatically makes the change. Very time consuming to continue reselecting number formats each time the cell contents are changed/accessed. Any ideas? Thx!

    Read the article

  • USB Mouse disconnects ONLY on bootcamp (win7,works fine on OSX) [duplicate]

    - by gourounakis
    This question already has an answer here: Why is my USB mouse disconnecting and reconnecting randomly and often? 7 answers I have a mid 2010 iMac with a Logitech G500 mouse which works fine on OS X. I game on Windows 7 in bootcamp, and for a month now I have been getting random mouse disconnects while gaming. Sometimes none, sometimes 2-3 per minute. The mouse lights go off and I get the disconnect sound from Windows 7 then it connects again after a second or two. I tried changing the port I connect the mouse to, but still the same thing. The only devices on USB are Apple keyboard with Apple extension cord, the mouse, and a Creative SoundBlaster Tactic 3D Alpha USB Gaming Headset. Any ideas?

    Read the article

  • Colliding network addresses

    - by joepd
    A customer is using the same address space as the company network. I need to connect with Cisco VPN Client to the network of the customer, without affecting local connectivity. To keep all working, I connect from a VM to avoid network addressing collisions. This works, but I'm looking for a way to get rid of the VM. Is it possible to connect with Cisco VPN Client 4.8 from Windows 7, without changing the 'normal' routing? How to tell specific applications (most importantly: Putty, VNC, mstsc, psql) to resolve their routes to the VPN instead of the default network interface? Have been hearing about SOCKS-proxies, is that something that could be made to work? I fear that this is a a bit of a crude question, but I'd be happy to specify more details/context.

    Read the article

  • How can I tell when an FTP is complete?

    - by identry
    I have a cron job that processes files that my client's upload via FTP to my FreeBSD server. The cron job runs once an hour, and normally processing each file only takes a few seconds. The cron job looks in the client's upload directory and moves any new files to a tmp directory. It then processes the file(s) and moves them into a final directory where they are then available to the public through a website. The problem is, every once in awhile, the cron job runs just as a new file is being uploaded. It moves the half-uploaded file to the tmp directory, and tries to process it, and fails, of course. Question: how can I determine if the uploaded file is complete? The only thing I can think of is checking the file size to see if it's changing, but that seems like a kludge. Is there some sort of flag or something that is set when the upload is complete?

    Read the article

  • How can I retrieve "remembered"(stored) wi-fi password from a win. 7 device?

    - by user180880
    I have access to PC, and I am a standard user. Everything(incl. "show charecters" tickbox at wiriless menu) requires admin access. Now, that said machine is actually like a big tv with touch. Type-stuff is handled by virtual keyboard of windows. I can reach to c:\ProgramData\Microsoft\Wlansvc\Profiles\Interfaces and can see-open files there, which is I assume where passwords stored are. Now the problem is that these passwords is encypted. I'm ok with if there is a way with changing/resetting admin password as well. Considering this device has nothing but massive amounts of usb(yep, not even cd-dvd // rw) the only way is from inside or with usb without admin rights.

    Read the article

  • Does a hard drive degrade if we always write to the same directory?

    - by code-gijoe
    I have a very heavy I/O application that is constantly receiving data through the network and writing to a specific directory on the HDD. Then, the application need to load the files from that specific place. One of my clients has been experiencing slowness and when I try to access the directory it takes quite long before I can see the content. My gut feeling is the HDD is degrading due to high I/O for a couple of years and I'm thinking of changing the HDD. Is there a benefit to write to multiple directories instead of using always the same? BTW he is using Windows XP.

    Read the article

  • Roughly, what percentage of users will reach changed DNS?

    - by user3722246
    If my main server go offline for some reason for +1hrs, I'm planning to make a DNS change so users will access secondary server. It is not a perfect solution to decrease downtime but it is simple and would work. I'm not sure about its usefulness. So I have a question. If I'm going to make a DNS change to an A record for my domain (changing from one IP to another), what percentage of users are moved over to the new info in 2hrs? (roughly) I know this is a vague question and there are lots of variables but any input is welcomed because I had painful downtime experiences and don't want to experience it again. Thanks

    Read the article

  • Firefox 16: Setting to bookmark always to Toolbar by Shortcut?

    - by Echt Einfach TV
    The question sounds easy but I could not find a solution using google search. Maybe you have an idea: I am searching for the Firefox setting to bookmark always to the toolbar. If I click CTRL+D the default "Folder" is "Bookmarks Menu". I'd like to have it always "Bookmarks Toolbar", without the need of changing it every time. Screenshot: http://i.imgur.com/wzxt5.png PS: I know that you can use the mouse and drag the icon from the URL to the toolbar. But using the keyboard more and more as it is just faster, this is not an option.

    Read the article

  • Sync laptop and desktop solution/ideas?

    - by user1078719
    I have a laptop and a desktop I want to keep files up-to-date on both of them. I was thinking of creating a hard drive on 'server' and then changing "my documents" to that server path. It should work all right when there are two computers but what if a laptop uses wireless access at my house and more importantly, when it's connected wirelessly on a different IP range when I'm at college? For the second case, I guess, creating a VPN/remote server connection to the desktop will work, but if I want to make it automatic (my documents mapping), is there way to do that?

    Read the article

  • Playbook is starting to get mail all of a sudden [closed]

    - by 1.21 gigawatts
    After 6 months my Playbook is just now this morning starting to pull in emails from my GMail account. Why? I didn't change anything on my end. The only difference I can think of (besides something changing on RIMs services end if that has any part in this) may be that I've been on a secure university connection for a few hours. I've been on this network before but the playbook doesn't seem to stay connected to it very long. Maybe it's because I have a lot of emails in my account? Has anyone else noticed this?

    Read the article

  • ODBC in SSIS 2012

    - by jamiet
    In August 2011 the SQL Server client team published a blog post entitled Microsoft is Aligning with ODBC for Native Relational Data Access in which they basically said "OLE DB is the past, ODBC is the future. Deal with it.". From that blog post:We encourage you to adopt ODBC in the development of your new and future versions of your application. You don’t need to change your existing applications using OLE DB, as they will continue to be supported on Denali throughout its lifecycle. While this gives you a large window of opportunity for changing your applications before the deprecation goes into effect, you may want to consider migrating those applications to ODBC as a part of your future roadmap.I recently undertook a project using SSIS2012 and heeded that advice by opting to use ODBC Connection Managers rather than OLE DB Connection Managers. Unfortunately my finding was that the ODBC Connection Manager is not yet ready for primetime use in SSIS 2012. The main issue I found was that you can't populate an Object variable with a recordset when using an Execute SQL Task connecting to an ODBC data source; any attempt to do so will result in an error:"Disconnected recordsets are not available from ODBC connections." I have filed a bug on Connect at ODBC Connection Manager does not have same funcitonality as OLE DB. For this reason I strongly recommend that you don't make the move to ODBC Connection Managers in SSIS just yet - best to wait for the next version of SSIS before doing that.I found another couple of issues with the ODBC Connection Manager that are worth keeping in mind:It doesn't recognise System Data Source Names (DSNs), only User DSNs (bug filed at ODBC System DSNs are not available in the ODBC Connection Manager)  UPDATE: According to a comment on that Connect item this may only be a problem on 64bit.In the OLE DB Connection Manager parameter ordinals are 0-based, in the ODBC Connection Manager they are 1-based (oh I just can't wait for the upgrade mess that ensues from this one!!!)You have been warned!@jamiet

    Read the article

  • Using design-patterns to transform web-service model classes into local model classes and vise versa

    - by Daniil Petrov
    There is a web-application built with play framework 1.2.7. It contains less than 10 model classes. The main purpose of the application is a lightweight access to a complex remote application (more than 50 model classes). The remote application has its own SOAP API and we use it for synchronization of data. There is a scheduled job in the web-app which makes requests to the remote app. It gets bunches of objects from the remote model and populates corresponding objects of the local model. Currently, there are two groups of classes - the local model and the remote model (generated from wsdl schema). It is not allowed to make any modifications to the remote model. Transformations are being made in the scheduled job class. When it gets objects from the remote app it creates local objects. Recently, it was decided to add a possibility to modify the remote objects. It requires more transformations on our side. We need to transform from remote to local model when reading objects and from local to remote when changing objects. I wonder if this would be possible to use some design-patterns to reduce a number of transformations?

    Read the article

  • How do I make an on-screen HUD in libgdx?

    - by Devin Carless
    I'm new to libgdx, and I am finding I am getting stumped by the simplest of things. It seems to want me to do things a specific way, but the documentation won't tell me what that is. I want to make a very simple 2d game in which the player controls a spaceship. The mouse wheel will zoom in and out, and information and controls are displayed on the screen. But I can't seem to make the mouse wheel NOT zoom the UI. I've tried futzing with the projection matrices in between Here's my (current) code: public class PlayStage extends Stage { ... public void draw() { // tell the camera to update its matrices. camera.update(); // tell the SpriteBatch to render in the // coordinate system specified by the camera. spriteBatch.setProjectionMatrix(camera.combined); spriteBatch.begin(); aButton.draw(spriteBatch, 1F); playerShip.draw(spriteBatch, 1F); spriteBatch.end(); } } camera.zoom is set by scrolled(int amount). I've tried about a dozen variations on the theme of changing the camera's projection matrix after the button is drawn but before the ship is, but no matter what I do, the same things happen to both the button and the ship. So: What is the usual libgdx way of implementing an on-screen UI that isn't transformed by the camera's projection matrix/zoom?

    Read the article

  • [Visual Studio Extension Of The Day] Test Scribe for Visual Studio Ultimate 2010 and Test Professional 2010

    - by Hosam Kamel
      Test Scribe is a documentation power tool designed to construct documents directly from the TFS for test plan and test run artifacts for the purpose of discussion, reporting etc... . Known Issues/Limitations Customizing the generated report by changing the template, adding comments, including attachments etc… is not supported While opening a test plan summary document in  Office 2007, if you get the warning: “The file Test Plan Summary cannot be opened because there are problems with the contents” (with Details: ‘The file is corrupt and cannot be opened’), click ‘OK’. Then, click ‘Yes’ to recover the contents of the document. This will then open the document in Office 2007. The same problem is not found in Office 2010. Generated documents are stored by default in the “My documents” folder. The output path of the generated report cannot be modified. Exporting word documents for individual test suites or test cases in a test plan is not supported. Download it from Visual Studio Extension Manager Originally posted at "Hosam Kamel| Developer & Platform Evangelist" http://blogs.msdn.com/hkamel

    Read the article

  • OpenGL extension vs OpenGL core

    - by user209347
    I was doubting: I'm writing a cross-platform engine OpenGL C++, I figured out windows forces the developers to access OpenGL features above 1.1 through extensions. Now the thing is, on Linux, I know that I can directly access functions if the version supports it through glext.h and opengl version. The problem is that if on Linux, the core doesn't support it, is it possible there is an extensions that supports the same functionality, in my case vertex buffer objects? I'm doing something like this: Windows: (hashdeck) define glFunction functionpointer_to_the_extension (apparently the layout changes font size if I use #) Linux: Since glext already defined glFunction, I can write in client code glFunction, and compile it both on Windows AND Linux without changing a single line in my client code using the engine (my goal). Now the thing is, I saw a tutorial use only the extension on Linux, and not checking for the opengl implementation version. If the functionality is available in the core, is it also available as extension (VBO's e.g.)? Or is an extension something you never know is available? I want to write an engine that gets all the possibilities on hardware, so I need to check (on Linux) for extensions as well as core version for possible functionality implementation.

    Read the article

  • Parallelism in .NET – Part 2, Simple Imperative Data Parallelism

    - by Reed
    In my discussion of Decomposition of the problem space, I mentioned that Data Decomposition is often the simplest abstraction to use when trying to parallelize a routine.  If a problem can be decomposed based off the data, we will often want to use what MSDN refers to as Data Parallelism as our strategy for implementing our routine.  The Task Parallel Library in .NET 4 makes implementing Data Parallelism, for most cases, very simple. Data Parallelism is the main technique we use to parallelize a routine which can be decomposed based off data.  Data Parallelism refers to taking a single collection of data, and having a single operation be performed concurrently on elements in the collection.  One side note here: Data Parallelism is also sometimes referred to as the Loop Parallelism Pattern or Loop-level Parallelism.  In general, for this series, I will try to use the terminology used in the MSDN Documentation for the Task Parallel Library.  This should make it easier to investigate these topics in more detail. Once we’ve determined we have a problem that, potentially, can be decomposed based on data, implementation using Data Parallelism in the TPL is quite simple.  Let’s take our example from the Data Decomposition discussion – a simple contrast stretching filter.  Here, we have a collection of data (pixels), and we need to run a simple operation on each element of the pixel.  Once we know the minimum and maximum values, we most likely would have some simple code like the following: for (int row=0; row < pixelData.GetUpperBound(0); ++row) { for (int col=0; col < pixelData.GetUpperBound(1); ++col) { pixelData[row, col] = AdjustContrast(pixelData[row, col], minPixel, maxPixel); } } .csharpcode, .csharpcode pre { font-size: small; color: black; font-family: consolas, "Courier New", courier, monospace; background-color: #ffffff; /*white-space: pre;*/ } .csharpcode pre { margin: 0em; } .csharpcode .rem { color: #008000; } .csharpcode .kwrd { color: #0000ff; } .csharpcode .str { color: #006080; } .csharpcode .op { color: #0000c0; } .csharpcode .preproc { color: #cc6633; } .csharpcode .asp { background-color: #ffff00; } .csharpcode .html { color: #800000; } .csharpcode .attr { color: #ff0000; } .csharpcode .alt { background-color: #f4f4f4; width: 100%; margin: 0em; } .csharpcode .lnum { color: #606060; } This simple routine loops through a two dimensional array of pixelData, and calls the AdjustContrast routine on each pixel. As I mentioned, when you’re decomposing a problem space, most iteration statements are potentially candidates for data decomposition.  Here, we’re using two for loops – one looping through rows in the image, and a second nested loop iterating through the columns.  We then perform one, independent operation on each element based on those loop positions. This is a prime candidate – we have no shared data, no dependencies on anything but the pixel which we want to change.  Since we’re using a for loop, we can easily parallelize this using the Parallel.For method in the TPL: Parallel.For(0, pixelData.GetUpperBound(0), row => { for (int col=0; col < pixelData.GetUpperBound(1); ++col) { pixelData[row, col] = AdjustContrast(pixelData[row, col], minPixel, maxPixel); } }); Here, by simply changing our first for loop to a call to Parallel.For, we can parallelize this portion of our routine.  Parallel.For works, as do many methods in the TPL, by creating a delegate and using it as an argument to a method.  In this case, our for loop iteration block becomes a delegate creating via a lambda expression.  This lets you write code that, superficially, looks similar to the familiar for loop, but functions quite differently at runtime. We could easily do this to our second for loop as well, but that may not be a good idea.  There is a balance to be struck when writing parallel code.  We want to have enough work items to keep all of our processors busy, but the more we partition our data, the more overhead we introduce.  In this case, we have an image of data – most likely hundreds of pixels in both dimensions.  By just parallelizing our first loop, each row of pixels can be run as a single task.  With hundreds of rows of data, we are providing fine enough granularity to keep all of our processors busy. If we parallelize both loops, we’re potentially creating millions of independent tasks.  This introduces extra overhead with no extra gain, and will actually reduce our overall performance.  This leads to my first guideline when writing parallel code: Partition your problem into enough tasks to keep each processor busy throughout the operation, but not more than necessary to keep each processor busy. Also note that I parallelized the outer loop.  I could have just as easily partitioned the inner loop.  However, partitioning the inner loop would have led to many more discrete work items, each with a smaller amount of work (operate on one pixel instead of one row of pixels).  My second guideline when writing parallel code reflects this: Partition your problem in a way to place the most work possible into each task. This typically means, in practice, that you will want to parallelize the routine at the “highest” point possible in the routine, typically the outermost loop.  If you’re looking at parallelizing methods which call other methods, you’ll want to try to partition your work high up in the stack – as you get into lower level methods, the performance impact of parallelizing your routines may not overcome the overhead introduced. Parallel.For works great for situations where we know the number of elements we’re going to process in advance.  If we’re iterating through an IList<T> or an array, this is a typical approach.  However, there are other iteration statements common in C#.  In many situations, we’ll use foreach instead of a for loop.  This can be more understandable and easier to read, but also has the advantage of working with collections which only implement IEnumerable<T>, where we do not know the number of elements involved in advance. As an example, lets take the following situation.  Say we have a collection of Customers, and we want to iterate through each customer, check some information about the customer, and if a certain case is met, send an email to the customer and update our instance to reflect this change.  Normally, this might look something like: foreach(var customer in customers) { // Run some process that takes some time... DateTime lastContact = theStore.GetLastContact(customer); TimeSpan timeSinceContact = DateTime.Now - lastContact; // If it's been more than two weeks, send an email, and update... if (timeSinceContact.Days > 14) { theStore.EmailCustomer(customer); customer.LastEmailContact = DateTime.Now; } } Here, we’re doing a fair amount of work for each customer in our collection, but we don’t know how many customers exist.  If we assume that theStore.GetLastContact(customer) and theStore.EmailCustomer(customer) are both side-effect free, thread safe operations, we could parallelize this using Parallel.ForEach: Parallel.ForEach(customers, customer => { // Run some process that takes some time... DateTime lastContact = theStore.GetLastContact(customer); TimeSpan timeSinceContact = DateTime.Now - lastContact; // If it's been more than two weeks, send an email, and update... if (timeSinceContact.Days > 14) { theStore.EmailCustomer(customer); customer.LastEmailContact = DateTime.Now; } }); Just like Parallel.For, we rework our loop into a method call accepting a delegate created via a lambda expression.  This keeps our new code very similar to our original iteration statement, however, this will now execute in parallel.  The same guidelines apply with Parallel.ForEach as with Parallel.For. The other iteration statements, do and while, do not have direct equivalents in the Task Parallel Library.  These, however, are very easy to implement using Parallel.ForEach and the yield keyword. Most applications can benefit from implementing some form of Data Parallelism.  Iterating through collections and performing “work” is a very common pattern in nearly every application.  When the problem can be decomposed by data, we often can parallelize the workload by merely changing foreach statements to Parallel.ForEach method calls, and for loops to Parallel.For method calls.  Any time your program operates on a collection, and does a set of work on each item in the collection where that work is not dependent on other information, you very likely have an opportunity to parallelize your routine.

    Read the article

< Previous Page | 145 146 147 148 149 150 151 152 153 154 155 156  | Next Page >