Search Results

Search found 9286 results on 372 pages for 'transfer speed'.

Page 41/372 | < Previous Page | 37 38 39 40 41 42 43 44 45 46 47 48  | Next Page >

  • How To Speed Up Adding Column To Large Table In Sql Server

    - by Chris
    I want to add a column to a Sql Server table with about 10M rows. I think this query would eventually finish adding the column I want: alter table T add mycol bit not null default 0 but it's been going for several hours already. Is there any shortcut to get a "not null default 0" column inserted into a large table? Or is this inherently really slow? This is Sql Server 2000. Later on I have to do something similar on Sql Server 2008.

    Read the article

  • Windows 7 Loading speed

    - by Sergiy Byelozyorov
    I have installed Windows 7 64-bit about two week ago. Since then everything was working alright, but yesterday Windows started to load much slower. Main delay comes while Windows logo is glowing -- it used to be around 10-15 seconds and now it's around 2 minutes. As far as I remember, I have only installed Microsoft IntelliMouse driver recently and removed it today. Unfortunately I found out that I have switched off System Restore some time long ago, so can't even test if the problem can be fixed by restoring old version. How can I troubleshoot the problem? Can I somehow find out what's causing the delay? P.S. Please don't suggest to reinstall Windows. I know that will help, but reinstalling all software and settings is a pain and requires a lot of time.

    Read the article

  • NTFS write speed really slow (<15MB/s) on Ubuntu

    - by Zulakis
    When copying large files or testing writespeed with dd, the max writespeed I can get is about 12-15MB/s on drives using the NTFS filesystem. I tested multiple drives (all connected using SATA) which all got writespeeds of 100MB/s+ on Windows or when formatted with ext4, so it's not an alignment or drive issue. top shows high cpu usage for the mount.ntfs process. AMD dual core processor (2.2 GHz) Kernel version: 3.5.0-23-generic Ubuntu 12.04 ntfs-3g version: both 2012.1.15AR.1 (Ubuntu default version) and 2013.1.13AR.2 How can I fix the writespeed?

    Read the article

  • Analyze a wireless network that constantly drops/has speed issues

    - by Eddie Parker
    Hello, I'm curious what the best tools are to use for analyzing problems with a WiFi network. Here's the scenario: I have a WiFi router (Belkin N+) that's setup in AP mode. I have three RT-N13U's that I've purchased to use as 'repeaters', but I've had so many problems when more than one of them is running (bad routes) that I've only got one active. Sometimes certain boxes on my network can't talk to others, and drops are quite frequent and quite aggravating. I'm running Mac, Windows, and Linux (Gentoo) boxes on this network, so any software, or steps I should take that work for any of those boxes should be sufficient. Apologies if this is answered somewhere else - I'll close it as a dupe if so.

    Read the article

  • Baseline upload speed for streaming video on demand?

    - by Peter Turner
    Is there a table that shows what is necessary for streaming video on demand? Specifically what I'm trying to do is stream Video on Demand to 50-100 people. I'm flexible with the format, but I'd like text on the screen to be legible (i.e. not 320x240). If someone knows of a calculator or something that would let me figure out exactly how I need to structure my video and ISP that'd be handy too.

    Read the article

  • How do I upload huge files across the internet without using P2P?

    - by Brien Malone
    I work remotely and have 44GB of media files that I need to send back to my office. There are lots of free services out there that can handle up to 2GB, but I haven't seen talk of anything larger. We both have 50mbps+ connections, so I would rather not mail physical media (though, that is an option). Bittorrent is blocked at my corporate headquarters. We have an FTP server, but the per-user cap is 10GB. I use Citrix, but throughput is throttled to 3mbps. (44gb @ 50mbps = 4 to 5 hours... @3mbps = 5 or 6 days.) Any suggestions appreciated. Windows 7 and Windows 2003 Server are the OSes Involved I have tried JetBytes and it is blocked by our content filter

    Read the article

  • synchronization of file locations between two machines

    - by intuited
    Although similar threads have been asked on this site and its siblings before, I've not managed to glean the answer to this persistent question. Any help is much appreciated. The situation: I've got two laptops; both contain a ton of music. Sometimes I move these music files to different locations, or change the metadata in them, or convert them to a different format. I might do any of these things on either machine. I rarely do all of them at once — ie it's unlikely that I'll convert a file's format and move it to a different location all in one go. I'd like to be able to synchronize these changes without having to sift through everything that was renamed or moved. I'm familiar with rsync but I find it inadequate, because although it can compute checksums, it doesn't have any way to store them. So if a file differs, it can't figure out which side it changed on. This also means that it can't attempt to match a missing file to a new one with the same checksum (ie a move) if the filesize and date are the same, it , so it takes an epoch to do a sync on a large repository. I would like to only check the checksum if the files even if you turn on checksumming, it still doesn't use it intelligently: ie it checksums files even if the sizes differ. IIRC. it's not able to use file metadata as a means of file comparison. this is sort of a wishlist item but it seems doable. I've also looked into rsnapshot, but its requirement to create a full backup is impractical in this situation. I don't need a backup, I just need a record of what file with each hash was where when. Unison seems like it might be able to do something vaguely along these lines, but I'm loathe to spend hours wading through its details only to discover that it's sadly lacking. Plus, it's fun asking questions on here. What I'd like is a tool that does something along these lines: keeps track of file checksums or of actual renames, possibly using inotify to greatly reduce resource consumption/latency stores a database containing this info, along with other pertinencies like the file format and metadata, the actual inode, the filename history, etc. uses this info to provide more-intelligent synchronization with a counterpart on the other side. So for example: if a file has been converted from flac to ogg, but kept the same base filename, or the same metadata, it should be able to send the new version over, and the other side should delete the original. Probably it should actually sequester it somewhere in case they or you screwed up, but that's a detail. And then when the transaction is done, the state is logged so that the next time the two interact they can work out their differences. Maybe all this metadata stuff is a fancy pipe dream. I would actually be pretty happy if there was something out there that could just use checksums in an intelligent way. This would be sort of like having the intelligence of something like git, minus the need to duplicate data in an index/backup/etc (and branching, and checkouts, and all the other great stuff that RCSs do. basically just fast forward commit pushes are all I want, with maybe the option to roll back.) So is there something out there that can do this? If not, can someone suggest a good way to start making it?

    Read the article

  • Exporting Thunderbird from Win7 32bit to Win7 64bit

    - by Muleskinner
    I am trying to export my thunderbird mails and mail accounts from my old win7 32bit to a new win7 64 bit. Both have the same version (16.0.1). Earlier in this situation I just copy the Thunderbird folder found under my user data account (tried both local and roaming), but this is not working in this scenario. I also tried to use Thunderbird export/import function, but it gives me a blank screen. This is what I am talking about:

    Read the article

  • rsync - Exclude files that are over a certain size?

    - by Rory
    I am doing a backup of my desktop to a remote machine. I'm basically doing rsync -a ~ example.com:backup/ However there are loads of large files, e.g. wikipedia dumps etc. Most of the files I care a lot about a small, such as firefox cookie files, or .bashrc. Is there some invocation to rsync that will exclude files that are over a certain size? That way I could copy all files that are less than 10MB first, then do all files. That way I can do a fast backup of the most important files, then a longer backup of everything else.

    Read the article

  • Move files on windows7 ntfs dead slow?

    - by slabo
    Just installed Win7, and i noticed when i move a file/folder, it actually copies, then deletes. I noticed because if i want to move 40 GB, I must have 40 GB free space. On XP, if I am moving folders and files within the same partition, no matter size, are moved instantly. Only the number of files matters. Any way to get around this? Thanks.

    Read the article

  • How to use mencoder to speed up video ?

    - by timday
    There's a webcam streaming WMV video that I can play/record with mplayer/mencoder (Debian Lenny including Debian-multimedia) just fine. I'd like to capture the video such that, when played back, it's faster (so say 100s of real time zooms past in every 1s of video). The "frameskip" filter seems like it ought to do this, but when I try e.g mencoder -o output.avi -ofps 25 -vf framestep=100 -ovc lavc -lavcopts vcodec=mpeg4:... <mms source URL> what I actually get is something which plays for the same length of time as it's allowed to capture for, but which only updates the displayed image every 4 seconds. Is there any way mencoder can be persuaded to do what I want ? (By the way, ideally I'd like to average the frames together to obtain motion-blur instead of just dropping/discarding most of them.).

    Read the article

  • Copying compressed files from Server 2008 R2 network share to XP client via VPN fails

    - by Dejan Janjuševic
    At the first sight the question looks similar to this one. I have experienced an odd behavior while trying to copy a certain file from Windows Server 2008 R2 network share to Windows XP Professional client via VPN. The VPN was set up using RRAS on the server machine. I will try to provide as much informations as possible in order to make the issue more clear. When trying to copy the compressed file sized ~2.5 MB (via Explorer or CMD, doesn't matter), the process stalls after some 20%, producing an error message after few seconds: Cannot copy filename: The specified network name is no longer available. If i start the command ping -t 192.168.2.1 (where the IP address specified belongs to the server) side by side with the copy command, I can clearly see that the ping command times out for few seconds as the copy process stalls. When this happens all network activities are frozen. After a few seconds, the network recovers, ping continues to run normally, however the copy process stands still before it displays the above error message. Copying other files (I tried 4-5 files), of which some are larger and some are smaller, succeeds. Seems to me that I can copy all uncompressed files. As soon as I try to copy an archive, the process freezes. Even a 707 KB large archive can't be copied. I can only reproduce this behavior on 2 machines, both Windows XP Professional, one is w/ SP2 and the other w/ SP3. Other XP clients don't have this problem, neither do Windows 7 clients. If I connect to the server using Remote Desktop Connection without using VPN from either of these 2 machines (using the same user account), I can copy anything I want normally, even these "problematic" files. Does anyone have any clue about what could possibly be going on?

    Read the article

  • Increase backup speed, Backup Exec 2010 - QNAP TS419U+ NAS

    - by user99912
    We have a QNAP NAS and the network shares are being backed up by Backup Exec 2010 over SMB. We can't install the remote agent on the NAS as it has an ARM processor and, as far as I am aware, there is no compatible agent. Do you have any suggestions on any faster method of backing up these shares as opposed to the current scenario? Currently the network bandwidth is not the issue, it seems that this access method is just not able to go any quicker. We've also added the NAS shares to the start of the selection list, but we're still running into 18 hours total backup time (total amount of data on the NAS is roughly 650GB). Any comments and/or suggestions welcome. EDIT: Data is being pulled from the NAS by Backup Exec to a LTO4 tape drive

    Read the article

  • Compressing and copying large files on Windows Server?

    - by Aaron
    I've been having a hard time copying large database backups from the database server to a test box at another site. I'm open to any ideas that would help me get this database moved without having to resort to a USB hard drive and the mail. The database server is running Windows Server 2003 R2 Enterprise, 16 GB of RAM and two quad-core 3.0 GHz Xeon X5450s. Files are SQL Server 2005 backup files between 100 GB and 250 GB. The pipe is not the fastest and SQL Server backup files typically compress down to 10-40% of the original, so it made sense to me to compress the files first. I've tried a number of methods, including: gzip 1.2.4 (UnxUtils) and 1.3.12 (GnuWin) bzip2 1.0.1 (UnxUtils) and 1.0.5 (Cygwin) WinRAR 3.90 7-Zip 4.65 (7za.exe) I've attempted to use WinRAR and 7-Zip options for splitting into multiple segments. 7za.exe has worked well for me for database backups on another server, which has ~50 GB backups. I've also tried splitting the .BAK file first with various utilities and compressing the resulting segments. No joy with that approach either- no matter the tool I've tried, it ends up butting against the size of the file. Especially frustrating is that I've transferred files of similar size on Unix boxes without problems using rsync+ssh. Installing an SSH server is not an option for the situation I'm in, unfortunately. For example, this is how 7-Zip dies: H:\dbatmp>7za.exe a -t7z -v250m -mx3 h:\dbatmp\zip\db-20100419_1228.7z h:\dbatmp\db-20100419_1228.bak 7-Zip (A) 4.65 Copyright (c) 1999-2009 Igor Pavlov 2009-02-03 Scanning Creating archive h:\dbatmp\zip\db-20100419_1228.7z Compressing db-20100419_1228.bak System error: Unspecified error

    Read the article

  • How to avoid damage to ISO archives?

    - by TMRW
    So had a problem where a 16GB ISO was damaged(likely my own fault using standard windows copy dialog instead of proper copy tool like robocopy with verification turned on). It took several hours but i managed to restore the ISO(basicly i rebuilt the damaged parts and recompiled).Namely some .rar archives inside it were unreadable but the ISO itself was readable. So im wondering how can i further protect something like this from happening again?.Obviously proper copy tool but mayble something else?.Perhaps set as "read only" could help?.I generally don't move these files a lot and if i need to acces them then it's only for opening/extracting.

    Read the article

  • Excel Help: Data Input Help

    - by B-Ballerl
    Everyday I download data from a site that will have rows each filled with individual data for clients. I'm able to input the data into excel as a whole but after that I'm having trouble figuring out how to put it into a chart. For example Web visits time. So say Client 1 stayed for 5 min increasing his total time on the site to 20 min and Client 2 stayed for 0 min keeping his time of 10 min and they were both registered on new years eve, and R1's last login was today and R2's was yesterday. (R for some reason repersents Client, no idea why...). Client 3 hasn't been on since he registered keeping his total at 4 min So my data would look something like this for Today (20110104) R1,20101231,20110104,20 R2,20101231,20110103,10 R3,20101231,20101231,4 And this for the day before (201101030), R1,20101231,20110102,15 R2,20101231,20110103,10 R3,20101231,20101231,4 I get about 200+ client rows each day where even the names of the Client list are changing. Is it possible to import the data each day and fill it in a excel sheet where the Client number is off on the left hand side in a table, and the amount of time (Whole Number ex. 4) each day it spends on the site extend to the right under it's specific date see Picture? I've manage to create a manual sheet but have been unsucessful at getting excel to do any of it for me. Here are two pictures:

    Read the article

  • How to fix semaphore timeout period has expired.

    - by Ryan
    I bough a new hard drive and I'm attempting to copy stuff off the old hard drive to the new one. I tried it using Windows Vista and Windows 7 but no go. Then I tried installing some 3rd party updated drivers that were supposed to fix the problem now my Windows Vista disk won't boot. Has anyone had this problem before and how did you fix it?

    Read the article

  • Netgear router-speed problem after XBox use

    - by John Dudley
    When my son is at my place at the weekend, he plays XBox Live over the internet, using my wireless network (Netgear 'g' type router). This usually thrashes & crashes the router to the extent I have to hard-boot it to get it working again. However, after this weekend, on my two laptops, I'm left with the problem that the router is working, but I'm only getting 0.38Mbps out of it, at all times of the day. I've tried hard-booting the router, but no difference. Could this be a knock-on effect of the Xbox use? I can't believe the router could be damaged, but working slower? Is that possible? Tiscali haven't come back to me yet on any 'network' issues. Thanks in advance

    Read the article

  • Optimizing wireless router speed and minimizing interference.

    - by Tchalvak
    I've been experiencing problems with my wireless connectivity lately, and want to make sure that it's not related to the abundance of other wireless routers here in my building. So, what I'm looking for is a method (probably via some application or another) to audit the wireless channels (and other factors that might be important that I don't even know of yet) that are floating through the aether around me. Ubuntu or other linux apps are preferred, but some kind of windows/mac solution is possible, since I do have other OSes around me that I could install & test on. Router: netgear WGT624 v3 Hearsay tells me that channels 1, 6, and 11 are "non-overlapping" (I expect they aren't used for non-wireless-router purposes or something, not sure how they couldn't overlap with other routers using other channels), so perhaps my best choices of channel are limited, so if channels aren't really a big concern, I'd be happy to get links to other optimizations that I should look into.

    Read the article

  • How to restore a Windows Easy Transfer file from a 64bit machine to a 32bit machine?

    - by Kevin Davis
    Using a 32bit laptop, I saved my settings etc. using "Windows Easy Transfer" from the Win7 RC. I set the file destination to a Win2K R2 machine that happened to be 64bit. When I re-installed my laptop and tried to restore my settings from the file I'd saved I was surprised to get an error: "Windows Easy Transfer can't transfer files from a 64-bit computer to a 32-bit computer." Is there a known workaround? Ideas on how to unpack the file and get my stuff?

    Read the article

< Previous Page | 37 38 39 40 41 42 43 44 45 46 47 48  | Next Page >