Search Results

Search found 10033 results on 402 pages for 'execution speed'.

Page 216/402 | < Previous Page | 212 213 214 215 216 217 218 219 220 221 222 223  | Next Page >

  • Windows 8 Hidden Downloading? Slow Internet?

    - by EApubs
    After upgrading to Windows 8, im facing a very slow download speed. My router and the NIC is working fine because on the same PC I also have Linux. In Linux, everything works fine. When checking through the task manager, it shows that the system use the full bandwidth 1Mbps even without a download. The process menu doesn't show whats taking much bandwidth. The bitdefender's firewall also shows network activity. Is there any hidden apps downloading? How to fix this?

    Read the article

  • how to design LAN connectivity between private and corporate ?

    - by maruti
    there is a bunch of servers connected to shared storage in a private LAN (10.x.x.x). this privateLAN is managed by a windows server (DHCP, DNS and directory services). how can these hosts be accessed from outside of this privateLAN? Eg. Remote desktop. can the NIC2 on each of the hosts be connected to the other public LAN (compromising speed or security? what are improtant considerations: additional hardware? like switches? routing&DNS software?

    Read the article

  • Backup/Multihomed network connection

    - by J_P
    We have a couple locations that require 24/7 access to Internet and our current provider (AT&T) while mostly good is not always up. My concern would be if I go with another provider (for example Comcast) I'm going to be subject to the same down time if it's in the "last mile". I for the most part don't know where the failure points are on the ISP side but I would imagine the large majority are within the last mile. I'd looked at Mifi or similar solution but have concerns about bandwidth caps and overall speed. Any suggestions would be appreciated.

    Read the article

  • Visual Studio 2010 locking referenced Assembly

    - by cunningdave
    I have a sandbox app that is built from the simple WPF Application template. This sandbox references an assembly that I am also building which contains the definition of a UserControl (WPF). I am instantiating this user control in the sandbox, to test the control's behaviour. The point of all this is to speed up development. This worked fine, but recently the .Vshost.exe paired with the sandbox process won't shut down. This prevents me from recompiling the Controls library, though ironically I can recompile the sandbox application. I can't kill the vshost process with Task Manager... only restarting VS2010 will clear it out. But every time I run the application from VS, the process just hangs there, blocking my workflow. I'm at a loss. Any ideas what could be causing this? Or does someone have any proposed workaround (mega-kill switch, perhaps?)

    Read the article

  • Installing WiFi in a church for internet access

    - by Dave Griffiths
    Hi, a colleague at work asked this question: My local church want to get WiFi enabled in the whole of their offices and main church building. The offices and main building are in the same overall structure and so the signal does not need to travel outside to any other buildings outside this main structure. Where would I start on such a project and what kind of repeaters are needed (and how are they set up) given that some of the walls are pretty thick. Do you still have one ADSL line coming in to router or are there some options of having two lines coming in so that they isn't a single point of failure. Am looking at fibre for the broadband for increased speed. Any starting pointers appreciated.

    Read the article

  • Shortcut with arguments in Debian

    - by Duncan
    I have a volume on a debian server which contains a large number of images at full resolution in various folders. What I'd like to do is have a separate sort of browse proxy folder which contains lower quality browse copies of these to enable users to access them for viewing over lower speed dial in accounts. I'd ideally like these to be created on the fly using ImageMagick so there isnt the need to store the large number of browse copies full time and worry about keeping them up to date etc The way I'd invisaged this happening is the browse proxy folder containing a duplicate file and folder structure but with symlinks pointing to a script to transform them with the file path as an argument. Except I know this isnt possible with symlinks so am wondering if there's another way of doing this on linux. On windows shortcuts can take arguments and I'm wondering how to do the same on a Linux platform? (or perhaps I'm going about this the wrong way?)

    Read the article

  • Pushing image changes to multiple servers

    - by gms8994
    I need the ability to push images out to multiple servers whenever they're updated. I've looked at Network Filesystems, but they're all but worthless due to their speed. Images can be uploaded to any one of 3 servers, and would then need to be copied to the other 2. Any suggestions? I'm open to try just about anything. EDIT: Graphics data (jpg, gif, png, etc). Linux only. We're currently using rsync. But having it work back and forth is getting cumbersome. It's all local network.

    Read the article

  • Rejecting new HTTP requests when server reaches a certain throughput

    - by user56221
    I have a requirement to run an HTTP server that rejects new HTTP requests (with a 503, or similar) when the global transfer rate of current HTTP responses exceeds a certain level. For example, if the web server is transferring at 98Mbps, and a new HTTP request arrives, we would want to reject this (as we couldn't guarantee a good speed). I've had a look at mod_cband for Apache, limit_req for nginx, and lighttpd's rate limiting features, but none of them seem to handle my (rather contrived, granted) use case. I should add that I'm open to using pretty much any web server, and am open to implementing this in iptables rules if someone can craft such a rule! (Refusing the TCP connection is fine, it doesn't have to respond with an HTTP 503). Any suggestions?

    Read the article

  • How does storage spaces decide where to put my files?

    - by George Duckett
    With Windows 8 Storage Spaces, you can lump many hard disks of varying types, speeds and sizes together to use as a single storage space / logical drive. How does Windows decide what to place where? For example will it move files about depending on frequency of access? Maybe splitting files frequently accessed together between hard disks etc. What does it optimize for? Speed, reliability, etc? If the above is asking too much, can I easily see where the files are physically (on which physical disk)?

    Read the article

  • My laptop keeps hard resetting

    - by cgoddard
    I have had this laptop for a long time (five years at least) and it hasn't had the best treatment over the years. But over the last few months, it has been randomly completely shutting down, no blue screening, no steady shutdown, just black screen, then BIOS. I think it might be happening as I do a save, but cant be sure, but recently it has been getting very annoying. I did initially think it was an overheating issue, but the speed that it comes back online is staggeringly fast for it to have cooled down sufficiently. plus it has got incredibly hot before with no issue (since its been doing this). Does anyone know what might be going on? Dell Inspiron 640m, running Windows 7.

    Read the article

  • can I connect two modems and comunicate on two side

    - by GongT
    I just wondering how modems work. I know the process of "modulation" and "demodulation". So I wanto know is "demodulation" are simple revese of "modulation" in real world. The PC can got an IP address when connect as type 1 What will happen when I connect them like type 2? type 1 : [PC] ================= [router] type 2 : [PC] === [m] ------- [m] === [router] [m] : modem(exactly same) === : Ethernet cable --- : DSL cable (phone line, maybe optical fiber, or something else?) ISP has a large number of model, Is them same thing as the one in my home(but with diffrent size/speed/price...etc)? Or it's completely different thing?

    Read the article

  • Why does the heat production increase as the clockrate of a CPU increases?

    - by Nils
    This is probably a bit off-topic, but the whole multi-core debate got me thinking. It's much easier to produce two cores (in one package) then speeding up one core by a factor of two. Why exactly is this? I googled a bit, but found mostly very imprecise answers from over clocking boards which do not explain the underlying Physics. The voltage seems to have the most impact (quadratic), but do I need to run a CPU at higher voltage if I want a faster clock rate? Also I like to know why exactly (and how much) heat a semiconductor circuit produces when it runs at a certain clock speed.

    Read the article

  • Gigabit connection between 2 computers

    - by Alecu
    I am trying to set up a connection between 2 pc's. It's a direct cable connection. The problem that I am having is that the connection isn't gigabit. I know both network cards support gigabit. I checked in the device settings and in one pc I can see in the Speed & Duplex the 1Gbps option but on the other pc I can't see that option despite the fact that the network card is gigabit. I have quite some files to copy from one pc onto another and I really need a good connection between them. I can't really remove the hard drive because it would void the warranty. Could this be a driver issue or a cable issue? On both pc's I have windows 8. The network card on the pc that doesn't have the 1 gbps option is an Atheros ar8161 card.

    Read the article

  • Download resume support blocked by isp?

    - by John Doe
    Can ISP block resume support for downloads ? I'm using IDM (internet download manager) to download of the internet from resume supported websites, yet I am unable to resume downloads. I tried different computers with the same result. Turned off firewall, didn't have any effect i was able to download with no issues until a couple of days ago. Another thing i noticed is that before IDM used to try to connect to several connections to speed up my download, but now it can only connect to one connection. Also i tried to download using my vpn, and i was able to download and resume downloads with no problem.

    Read the article

  • Utilising a Magento server cluster to drive hot reindexing

    - by WOBenji
    We've asked a similar question in the past, basically we have a very large Magento store with 500000 products which are currently reindexed once a day, during the night. We'd like to speed this process up significantly, we're at about 4-5 hours now. The solution was suggested for us to do something like this on a server cluster and replicate the database changes after they've been done on a machine that isn't being bothered with serving customers. But what is the mechanism for that? How do we replicate those changes across to the live site from the server cluster? Can someone point me in the right direction here?

    Read the article

  • Speedup of fixing an openssl bug with 8192 bit key [on hold]

    - by rubo77
    This is related to this Bug-Report https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=747453 OpenSSL contains a set of arbitrary limitations on the size of accepted key parameters that make unrelated software fail to establish secure connections. The problem was found while debugging a XMPP s2s connection issue where two servers with long certificate keys (8192 Bit RSA) failed to establish a secure connection because OpenSSL rejected the handshake. This seems to be a small problem to be fixed but although there is an easy patch available to fix the issue in that bug report, no reactions are noticed so far.. The last patch that broke the 2048 barrier took 2 years to be implemented and only resulted in an increase to 4096bit, which seems to be a bad joke. Where would we have to report this to speed up the implementation for such an issue?

    Read the article

  • Slow network file copy on Windows 7

    - by Jason V.
    I wrote a program that uses xcopy to transfer files (usually between 1KB and 2MB) over our intranet. Usually, I am copying files from my host machine (Windows 7 x64) to a VMWare virtual machine running Windows Server 2008 (the VM is running on my host machine, if that matters). On Windows XP, the file transfers usually only require a few seconds to complete. But on my Windows 7 machine, the transfer of the first file (1.5 MB) takes around 1.5 minutes to complete. This is true whether I use xcopy, robocopy, or programmatically using File.Copy(). I noticed that if I use File.Copy, the first transfer is very slow and subsequent transfers are much faster. Any clue how I can speed up the process? Is there a setting in Windows 7 (or server 2008) that I could try?

    Read the article

  • Windows Command queue?

    - by Stefano
    i'm thinking if does exist some kind of software that can put in a queue a bunch of windows commands... for example i can say to first copy some file somewhere, then rename those, then delete the old files, then edit one of them etc.... without waiting the effective execution of any of those passages.... this could be useful when copying big files that take a lot and i don't want to sit in front of the computer keeping the eyes on the progress bar... does exist anything like this?

    Read the article

  • Adding expire headers to content served from CDN?

    - by mdolon
    I'm using MaxCDN to serve content to my blog using W3 Total Cache. The problem I'm running into when evaluating my site using Google Page Speed and YSlow! is that expire headers are not being sent on content delivered from the CDN, nor are they coming from a cookieless domain. Is this something that is completely in the hands of my CDN or is it something I can fix using my server configuration? Some info about my setup: nginx with php-fastcgi wordpress 3.0 w3 total cache 0.9a (dev release) MaxCDN the site: http://devgrow.com/

    Read the article

  • How can I check that I'm not frying the chipset on my 945 gclf?

    - by Journeyman Geek
    I'm running an old 945 gclf as a home server. Its an odd little system which has a passively cooled single core atom 230 processor, and a small heatsink/fan combo over the chipset. lm-sensors/sensors detects one temperature sensor which I guess is the processor. coretemp-isa-0000 Adapter: ISA adapter Core 0: +38.0°C (crit = +100.0°C) The fan on it is making rather horrible noises and I am considering unplugging it or using a speed reducer. There's also a risk it may fail anyway. Is there any way I can monitor the temperatures of the chipset or be warned of any issues with it? Is there anything specific I can watch out for that would indicate trouble?

    Read the article

  • Mysql skip-name-resolve

    - by user72459
    I had recently purchased a new server, and transferred all my accounts via WHM Transfer. The problem is that when WHM takes a daily backup, it outputs are message such as DBD::mysql::st execute failed: There is no such grant defined for user 'abc' on host 'myhostname' The problem is solved when I remove skip-name-resolve from my my.cnf file. Tough I dont find any differences in the speed (when I dont add it), it is often mentioned in forums that adding skip-name-resolve optimizes Mysql Performance. Does adding skip-name-resolve really help, if one has a Dedicated server?

    Read the article

  • Prioritize file sharing capabilities in Windows Server 2008

    - by cmbrnt
    I've got a server running Windows Server 2008, and use it mainly for sharing files throughout the domain from a number of disks. It's running on VMware ESXi 4.0, in case that matters. My problem is that when I log in to the server to check user permissions etc, the access speed the files on the remote disks almost grinds to a halt. I havn't been able to measure the speeds, but I would guess it slows down to about 100kB/s as soon as I log in. This is on a gigabit network and the problems are equal for all users, even the ones connected to the same switch as the server. I've assigned 2 GB RAM to the server, and reserved it 1,5Ghz processor power. I don't have to do anything special on the server for this halt to occur. How can I make sure file sharing is prioritized on the server, so no matter what applications I'm using it will always make sure file sharing works properly? Could this be a VMware issue?

    Read the article

  • For very beginning startup: home server or EC2?

    - by StCee
    The micro instance of Amazon EC2 only has ram of 613MB, my laptop got 8GB. And I suppose likewise the processing power of my computer would be better than the micro instance. My question is, what are the considerations in deciding to host yourself or on Amazon EC2, especially for a really baby startup? For example, would network speed be a problem? My computer broadband network is 100Mbs up to 1Gbs. What would Amazon compare to this? My site at these moment would just host some images and perform some php requests. I would probably also use cloudflare but seems it increases the dns lookup time considerably... And of course the overall objective is to make the best user experience.

    Read the article

  • How to set a character set per application in *nix?

    - by SimmaDoWN
    I am attempting to set a character set of IBM850 on slackware linux for a particular application (epic5). Im using rxvt-unicode and have setup LANG/LC_*=en_US. Now if I set the encoding to IBM850 in kde's konsole program im able to display certain characters correctly. I'd rather not use IBM850 for everything; is there a way to set/alias a command for a per application execution? Ive tried things like: LC_CTYPE=IBM850 epic5 LC_ALL=IBM850 epic5 No success. Any help would be appreciated

    Read the article

  • Running a cronjob

    - by Ed01
    've been puzzling over cronjobs for the last few hours. I've read documentation and examples. I understand the basics and concepts, but haven't gotten anything to work. So I would appreciate some help with this total noob dilemma. The ultimate goal is to schedule the execution of a django function every day. Before I get that far, I want to know that I can schedule any old script to run, first once, then on a regular basis. So I want to: 1) Write a simple script (perhaps a bash script) that will allow me to determine that yes, it did indeed run successfully, or that it failed. 2) schedule this script to run at the top of the hour I tried writing a bash script that simple output some text to the terminal: #!/bin/bash echo "The script ran" Then I dropped this into a .txt file MAILTO = *****.******@gmail.com 05 * * * * /home/vadmin/development/test.sh But nothing happened. I'm sure I did many things wrong. Where do I start to fix all of this?

    Read the article

< Previous Page | 212 213 214 215 216 217 218 219 220 221 222 223  | Next Page >