Search Results

Search found 4580 results on 184 pages for 'faster'.

Page 84/184 | < Previous Page | 80 81 82 83 84 85 86 87 88 89 90 91  | Next Page >

  • Mouse for a lefty?

    - by Kyralessa
    As a programmer, I'm fairly particular about my keyboard and mouse. One thing I've only recently noticed is that the mice I tend to prefer don't work well for my five-year-old. He's left-handed, but ends up mousing with his right hand on the computer because (a) that's where the mouse is, and (b) most mice are designed for right-handed people anyway. So, a couple of questions: If you're left-handed and you mouse with your left hand, do you have any recommendations on good mice? If you have a lefty in your household (whether or not it's you), is there an easy way to swap settings between left-handed and right-handed mouse buttons? One way is Control Panel Mouse Check the box on the first tab OK. Do you know of anything faster? Or, better yet, what I'd really like is a way to reverse those settings on only one particular mouse, which would be designated as my son's mouse.

    Read the article

  • HD movies stutter.

    - by Absolute0
    I just put together a new system build in the hopes that all of my daily tasks would run smoothly and without any hiccups. Unfortunately I am still seeing some sound and sometimes video stuttering when playing HD movies in VLC (no problems with xvid/divx files). My setup is as follows: Intel core i5 750 quad core 2.66mhz 4GB ram asus p55 motherboard radeon hd 5570 video card 650gb 7200rpm western digital sata HDD 23" Nec ea23wmi monitor Operating System: Windows 7 What might be the main bottleneck that needs upgrading to fix my delays? Seems like the hard drive might be the problem but anything faster than 7200rpm is beyond my budget for a decent hard drive. Could it be anything else?

    Read the article

  • Opening offline version of Microsoft Books Online in browser

    - by ercan
    I often use the MSDN website for language reference. In order to make navigation faster, I downloaded the offline version of SQL Server 2005 Books Online from here: http://www.microsoft.com/downloads/details.aspx?familyid=be6a2c5d-00df-4220-b133-29c1e0b6585f&displaylang=en The reason why it is 137MB is that it comes with its own GUI, which, not surprisingly, is rather poor! Apparently though, the pages are written in html. The URIs look like: ms-help://MS.SQLCC.v9/sqlcc9/html/674933a8-e423-4d44-a39b-2a997e2c2333.htm . I can open the URI in IE, but with errors. Do you know if I can open them with Firefox and how? Or is there a simple HTML version of "MS Books Online", for example in a ZIP file?

    Read the article

  • Should I use my ISP's DNS, or Google's 8.8.8.8?

    - by Torben Gundtofte-Bruun
    It seems like a good idea to use Google's public DNS 8.8.8.8 and 8.8.4.4 because it's really fast -- much faster than my own ISP's DNS! -- and probably more reliable, too. That seems like a ridiculously quick win for me, and much easier to remember. Assuming we're not all "tin foil hat" about Google, why shouldn't everybody use Google DNS? Note: I've seen this question, but I don't want a comparison to OpenDNS. This is about everyday use by everyday people in their homes. Update: I seem to have put my hand in a wasps' nest of privacy concerns. I appreciate the issue, but I was expecting a more technology-oriented discussion...

    Read the article

  • Directly editing IIS 7 applicationHost.config configuration file

    - by lunadesign
    I know that IIS 7+ now uses XML config files instead of the metabase. I also know that if I edit a web.config file for a given site, IIS automagically detects the changes and implements any corresponding config changes. However, does this also apply to the server-level applicationHost.config settings file? (Its usually located in C:\windows\system32\inetsrv\config.) Specifically, is it safe to carefully edit this file instead of using IIS Manager or the appcmd command line utility? I couldn't find anything in the documentation that said it was okay or not okay to do this. I'm curious because I have to change the bindings for numerous sites from one IP to another. It would be much faster to simply do a global search and replace for the IP address in the config file instead of manually editing a few dozen sites in the GUI.

    Read the article

  • Reducing latency for different geographic regions on Amazon Cloud

    - by Shoaibi
    I have got an application which has three components Application code : Amazon EC2 US-EAST-1 instance Application images, and other static data : Amazon S3 with CloudFront Application Database : Amazon RDS In short i need something like Cloud Front for EC2. In long, people using this application from a different region say middle east will have faster static content downloading due to Cloud Front but there would be a lot of latency in communicating to EC2 instance. I want to use a budget friendly way of enhancing this. Launching Amazon Instances in every region that offer is sure a choice, but isn't really cheap, so would try to avoid it unless its last resort. Also say if my clients also need to communicate to the RDS database directly, is there some kind of solution which gives that kind of functionality mentioned above, but for RDS?

    Read the article

  • zeroing a disk with dd vs Disk Utility

    - by jdizzle
    I'm attempting to zero a disk on my Mac OS X machine. I'm going for complete zeros and unformatted, so I think of dd. Unfortunately the maximum throughput I've managed to get out of dd is 7MB/s. Just for grins I tried disk utility and it has a throughput of 19MB/s. What gives? I've tried changing the bs option on dd to all sorts of values, but it still hovers around 7MB/s. Why is disk utility so much faster?

    Read the article

  • What is the best file system and allocation size for a USB flash drive?

    - by e-t172
    I'm considering using my 4 GB Kingston DataTraveler USB stick to store my Firefox and Thunderbird profiles for my laptop and desktop PCs. I want to maximize performance when using Firefox. The question is: what is the best file system and allocation size for the fastest Firefox profile operation on a USB flash drive? I'm using Windows 7 on both machines and I don't care about compatibility or the drive's lifetime. I just want to maximize performance. I could even use ext2 with the Ext2 IFS driver if that means it'll be faster. I'm assuming (perhaps I'm wrong) that putting a Firefox profile on a USB stick would be a "lots of small files" usage. In that case, it seems that NTFS would perform best, but I'm not sure. Besides I found nothing regarding the best allocation size to use. Considering that the default allocation size is designed for hard drives (which have different characteristics), I'm assuming that the default allocation size is not the best.

    Read the article

  • PostgreSQL: performance descrease due to index bloatper

    - by Henry-Nicolas Tourneur
    I'm running a PgSQL 8.1 on a CentOS 4.4 system (not upgradable unfortunately). There's a Java app running on top of the PgSQL daemon and we got to reindex the database every 2 months or so. Also important: the database isn't growing. It looks like the bloat is now coming faster than before and this tends to increase. My config is available here, autovacuum daemon is enabled and running quite often: pastebin.com/RytNj7dK You can also find the output of this query wiki.postgresql.org/wiki/Show_database_bloat 3 hours after running reindex: http://pastebin.com/raw.php?i=75fybKyd 72 hours after running reindex: http://pastebin.com/raw.php?i=89VKd7PC Does anyone have any idea what should I tweak to get rid of that growing bloat? Thanks for your help PS: due to antispam prevention system, I had to remove the first 2 http:// prefixes for my two first links.

    Read the article

  • Computer becomes very slow (permanently) after running a bunch of apps

    - by djzmo
    Hello there, My computer with Windows XP installed becomes very slow after I ran some heavy tasks at a time. (play a full 3D online game while extracting a 4GB RAR archive) It freezes for about 200~500ms every few seconds, and this always happens if I do heavy tasks at once in my computer (for example, installing a program while playing games), and the lag remains permanently (even a reboot won't make it better) unless I repair-install the Windows. Since I have a low-end computer: Intel(R) Pentium(R) 4 CPU 2.00 GHz, 512 MB of RAM ATI RADEON 9550 AGP 256 MB And the only way I used everytime to fix this problem is by repair-installing my Windows XP. So that I won't lose any data or installed programs. But I believe there's a better and faster way to fix this without repair-installing the Windows. Any idea?

    Read the article

  • VirtualBox in production?

    - by MrG
    I'm planning to move a service which is currently powered by Debian into a VirtualBox. That would allow us to easily port it i.e. to a faster machine if required. The setup would be: debian host > Virtual Box #1 > debian instance #1 running Apache & application > Virtual Box #2 > debian instance #2 containing database Do you have any experience with a production setup based on Virtual Box? Is it stable and fast enough? Many thanks!

    Read the article

  • PowerShell filter vs. function

    - by Marcel Janus
    I'm reading currently the Windows PowerShell 3.0 Step by Step book to get some more insights to PowerShell. On page 201 the author demonstrates that a filter is faster than the function with the same functionally. This script takes 2.6 seconds on his computer: MeasureaddOneFilter.ps1 Filter AddOne { "add one filter" $_ + 1 } and this one 4.6 seconds MeasureaddOneFunction.ps1 Function AddOne { "Add One Function" While ($input.moveNext()) { $input.current + 1 } } If I run this code is get the exact opposite of his result: .\MeasureAddOneFilter.ps1 Days : 0 Hours : 0 Minutes : 0 Seconds : 0 Milliseconds : 226 Ticks : 2266171 TotalDays : 2,62288310185185E-06 TotalHours : 6,29491944444444E-05 TotalMinutes : 0,00377695166666667 TotalSeconds : 0,2266171 TotalMilliseconds : 226,6171 .\MeasureAddOneFunction.ps1 Days : 0 Hours : 0 Minutes : 0 Seconds : 0 Milliseconds : 93 Ticks : 933649 TotalDays : 1,08061226851852E-06 TotalHours : 2,59346944444444E-05 TotalMinutes : 0,00155608166666667 TotalSeconds : 0,0933649 TotalMilliseconds : 93,3649 Can someone explain this to me?

    Read the article

  • Very low throughput on 10GbE network

    - by aix
    I have two Linux machines, each equipped with a Solarflare SFN5122F 10GbE NIC. The two NICs are connected together with an SFP+ Direct Attach cable. I am using netperf to measure TCP throughput between the two machines. On one box, I run: netserver and on the other: netperf -t TCP_STREAM -H 192.168.x.x -- -m 32768 I get: MIGRATED TCP STREAM TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 192.168.x.x (192.168.x.x) port 0 AF_INET Recv Send Send Socket Socket Message Elapsed Size Size Size Time Throughput bytes bytes bytes secs. 10^6bits/sec 87380 16384 32768 10.02 1321.34 The measured throughput is 1.3Gb/s. This is 7.5x below the theoretical maximum, and only 30% faster than 1GbE. What steps can I take to troubleshoot this?

    Read the article

  • Website content hosted with Google. Good or bad?

    - by user305052
    I recently decided to host my styles.css and various scripts on Google Docs and link them into my website. I also have all my images hosted through Picasa so that they too will load much faster and consistently across users. My site has most of its traffic from Japan, Africa, and South America, so I assume there will be a performance boost for my users since my server is hosted in Hong Kong. I (in Canada) have measured my load times to be half of what they used to be. Basically it's a free CDN for my personal stuff. I'm not too sure about all of this yet, so here's my question: what are the caveats of this setup? EDIT: So after rummaging through the ToS of both Picasa and Docs, there doesn't seem to be anything wrong with this kind of use.

    Read the article

  • when to use squid on server side?

    - by ajsie
    so i have set up apache serving my php pages. i read about squid but don't understand why/how i should use it to speed up my web server. from what i've learned squid is located in same network (or another) and caches content requested by the web browsers, and then when another web browser wants a same page, squid returns that page cached locally, so it never sends a request to the apache server (faster response time for the client, and reduced load for the server). so it seems that squid is for the client side (web browser), and has nothing to do with the server side (apache). but then some people tell others how they have speeded up apache using squid. so im confused. could squid be used on the server side too? and how will it work?

    Read the article

  • OSX Time Machine: deletion of backup folders

    - by jml
    I saw this question and was hoping that someone could expand upon the chosen answer (which I understood): Can you sudo mv Time Machine backup files as sudo from the trash to their original locations? I have tried doing this as root to no avail (operation not permitted). If not, can you successfully rm them via the trash via the terminal, faster than what the endless 'preparing to empty the trash' dialog suggests, and If you get the files back out of the trash can you tell if they are intact via disk utility (and how) Can you force indexing on a Time Machine drive in the same way that you would a normal drive to rebuild the TM index? I realize that a single answer could clarify all of the above, but I wanted to include details to be clear on what I am asking. Thanks for any help.

    Read the article

  • Fedora12 Slow USB 2.0 Write Speed, ehci_hcd module is missing

    - by MA1
    I am using Fedora 12, the problem I am facing is USB 2.0 write speed. I have a dual boot system with Windows XP and Fedora 12. USB 2.0 write speed in Windows XP is much faster then what I am getting in Fedora 12. After searching Google I came to know that ehci_hcd module is missing/not present in my system. ehci_hcd module is neither loaded nor it is present in the available list of modules. Can someone guide me how to fix this issue? Does ehci_hcd have something to do with USB 2.0 write speed? Do I have to recompile the kernel and add/enable the ehci_hcd module?

    Read the article

  • SQL Server 2008 Optimization

    - by hgulyan
    I've learned today, if you append to your query OPTION (MAXDOP 0) your query will run on multiple processors and if it's huge query, query will perform faster. I know general guidelines on query optimizations (using indexes, selecting only needed fields etc.), my question is about SQL Server optimization. Maybe changing some options in configurations or anything else. What guidelines are there for SQL Server Optimization? Thank you. P.S. I suppose, this is not the right place to ask server related questions. Should I delete it or maybe it can be migrated to serverfault?

    Read the article

  • Accessing through VPN, which internet connection would be used

    - by Sriram
    I've a data card which has a limit of 2 GB up/download traffic per month. I've an office internet line which has an unlimited up/download. I've successfully connected to VPN using my data card and by changing certain configurations like DNS etc., have also been able to use my office line for internet (verified by doing a IP check - returns the static IP of our office). Now to my question - is it just a NAT which is happening or am I actually using my office line for all communication. Which one would reflect the usage/trace. The data card usage log at this moment does not reflect any usage (which is confusing since the VPN is over the data card connection). Further more (theoretically), would the net be any faster if my office line is let us say 8Mbps and the data card is 512kbps by doing this as against directly connecting the internet using the data card.

    Read the article

  • ASUS EAH4670 vs. ASUS EAH4670 V2 -- Whats the difference?

    - by roosteronacid
    I've been offered to buy a used ASUS EAH4670/DI/1GD3 graphics card. I went price-scouting to see if the offer I was given was fair, and I found out that there's a similar card, labelled ASUS EAH4670/DI/1GD3 V2. Question is; What's the difference? What's with the V2? What does it mean? Is it just a BIOS upgrade that I can do myself? Updated driver-software which I can download myself? Or is the card actually a better version--faster, more reliable (physical changes to the print), etc.?

    Read the article

  • Building a new cluster for mathematical calculations (Win/Lin)

    - by Muhammad Farhan
    I would like to build a new cluster to perform heavy mathematical calculations in Matlab and Abaqus. One of my friend told me that distributed computing is way faster than parallel computing, which is very true after reading a bit on the internet. However, I have never clustered before. Current workstation I own: Dell Precision T5400 2 x Intel Xeon 2.5 GHz 16 GB RAM (2GB x 8) 1 x Western Digital 1TB HDD 7200 rpm 1 x nVidia Quadro FX4600 768MB GPU 1 x 870W PSU OS: Windows 7 Ultimate 64-bit 2nd WS: I can buy another WS similar configuration to the one I own I am not bothered about OS, I am willing to cluster with either Windows or Linux. However, my software are compatible with windows 64-bit only. Please help me setup a cluster. Thank you.

    Read the article

  • how to optimize virtual box shared folders

    - by Nrew
    This is really pissing me off. No matter how much memory I put into the guest os(windows xp). It still hangs for about 365 days before you can access the file you want to access from the shared folder. What do I do to make things faster? Because after it hangs and not respond for 365 days. It will do it again for another 250 days. Ive even set the shared folder to permanent. This is a fairly decent machine: 2.50ghz processor(x64 architecture, but I have only 2Gb of memory so my host os is just 32 bit windows 7) hdd has much space left: 156 Gb free of 250Gb

    Read the article

  • Java: how to do fast copy of a BufferedImage's pixels? (include unit test)

    - by WizardOfOdds
    I want to do a copy (of a rectangle area) of the ARGB values from a source BufferedImage into a destination BufferedImage. No compositing should be done: if I copy a pixel with an ARGB value of 0x8000BE50 (alpha value at 128), then the destination pixel must be exactly 0x8000BE50, totally overriding the destination pixel. I've got a very precise question and I made a unit test to show what I need. The unit test is fully functional and self-contained and is passing fine and is doing precisely what I want. However, I want a faster and more memory efficient method to replace copySrcIntoDstAt(...). That's the whole point of my question: I'm not after how to "fill" the image in a faster way (what I did is just an example to have a unit test). All I want is to know what would be a fast and memory efficient way to do it (ie fast and not creating needless objects). The proof-of-concept implementation I've made is obviously very memory efficient, but it is slow (doing one getRGB and one setRGB for every pixel). Schematically, I've got this: (where A indicates corresponding pixels from the destination image before the copy) AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA And I want to have this: AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAAAAAAAAA AAAAAAAAAAAAABBBBAAA AAAAAAAAAAAAABBBBAAA AAAAAAAAAAAAAAAAAAAA where 'B' represents the pixels from the src image. I'm looking for an exact replacement of the method, not for an API link/quote. import org.junit.Test; import java.awt.image.BufferedImage; import static org.junit.Assert.*; public class TestCopy { private static final int COL1 = 0x8000BE50; // alpha at 128 private static final int COL2 = 0x1732FE87; // alpha at 23 @Test public void testPixelsCopy() { final BufferedImage src = new BufferedImage( 5, 5, BufferedImage.TYPE_INT_ARGB ); final BufferedImage dst = new BufferedImage( 20, 20, BufferedImage.TYPE_INT_ARGB ); convenienceFill( src, COL1 ); convenienceFill( dst, COL2 ); copySrcIntoDstAt( src, dst, 3, 4 ); for (int x = 0; x < dst.getWidth(); x++) { for (int y = 0; y < dst.getHeight(); y++) { if ( x >= 3 && x <= 7 && y >= 4 && y <= 8 ) { assertEquals( COL1, dst.getRGB(x,y) ); } else { assertEquals( COL2, dst.getRGB(x,y) ); } } } } // clipping is unnecessary private static void copySrcIntoDstAt( final BufferedImage src, final BufferedImage dst, final int dx, final int dy ) { // TODO: replace this by a much more efficient method for (int x = 0; x < src.getWidth(); x++) { for (int y = 0; y < src.getHeight(); y++) { dst.setRGB( dx + x, dy + y, src.getRGB(x,y) ); } } } // This method is just a convenience method, there's // no point in optimizing this method, this is not what // this question is about private static void convenienceFill( final BufferedImage bi, final int color ) { for (int x = 0; x < bi.getWidth(); x++) { for (int y = 0; y < bi.getHeight(); y++) { bi.setRGB( x, y, color ); } } } }

    Read the article

  • Not working WEP key in Ubuntu

    - by stupid-idiot
    This time I've got some problem with connection to the internet from my old laptop with ubuntu, I already had some problems with setting up the radio on my network card but I think it works now since it finds some wireless networks. The problem is that I can't connect to my router via wi-fi even if I'm hundred-percently sure the wepkey I punch there is correct, it works via LAN though. Also, I'm not able to ping myself with 127.0.0.1. I'm sorry to say it but I don't think I can stand this shitness for much longer. Although it's much much faster than Windows it's useless since i can't download any libraries for development. If anyone has any experience with this, even indirect, please leave your answer or comment thx

    Read the article

  • MacBook Pro - 72k vs 54k rpm drives - Heat and Noise

    - by webworm
    I am looking at a new 15" MacBook Pro for development purposes. I am planning to run a Virtual Machine for about 50% of my work (Windows 7 x64, IIS, SQL Server, and VS 2010). The upgrade from a 54000 rpm drive to a 72000 rpm is only $45. From what I understand the faster rotational speed of the 72000 rpm drive will help virtual machine performance. However, I am concerned that additional heat and fan noise might be an issue. I will be running mostly on A/C power so decreased battery life is not a major concern for me. Since I would be running with a Core i7 processor which gives off a fair amount of heat already I was wondering if it might be best to stay at 54000 rpm for the hard drive. What do you all think? Thanks!

    Read the article

< Previous Page | 80 81 82 83 84 85 86 87 88 89 90 91  | Next Page >