Search Results

Search found 6355 results on 255 pages for 'slow downs'.

Page 103/255 | < Previous Page | 99 100 101 102 103 104 105 106 107 108 109 110  | Next Page >

  • Zeacom UC Compared To Microsoft UC

    - by Kia
    Which is a better solution? Zeacom's Unified Communications or Microsoft's Unified Communications (UC)? Which one has your company implemented? I heard Microsoft coined the term "Unified Communications" but they were slow to jumpstart it... Other companies such as Zeacom have been working on and improving on their UC product since years ago. But Microsoft is such a standard. Which one would you go with?

    Read the article

  • Does anybody knows a replacement for TorrentSpy?

    - by Ither
    Hi, I've used TorrentSpy since 2004 and it does a pretty good job but it is very slow with torrent files that have hundreds or thousands of files. Does anyboy knows a better tool for the same job in XP? If you didn't know it, TorrentSpy show the data contained in a torrent file in a readable way: URL's tracker, number of files, its size, number of complete and incomplete downloaders, etc. Edit: What I want is something that can be used from explorer right-clicking in the torrent file, like MediaInfo does with media files.

    Read the article

  • Server to server replication and CPU and 32k\ corrupt doc

    - by nick wall
    Summary: if database contains a doc with 32K issue or corrupt, on server to server replication it causes marked increase in CPU in nserver.exe task, which effectively causes our server(s) to slow right down. We have a 5 server cluster (1 "hub" and 4 HTTP servers accessed via reverse proxy and SSO for load balancing and redundancy). All are physically located next to each other on network, they don't have dedicated network\ ports for cluster or replication. I realise IBM recommendation is dedicated port for cluster. Cluster queues are in tolerance and under heavy application user load, i.e. the maximum number of documents are being created, edited, deleted, the replication times between servers are negligible. Normally, all is well. Of the servers in the cluster, 1 is considered the "hub", and imitates a PUSH-PULL replication with it's cluster mates every 60mins, so that the replication load is taken by the hub and not cluster mates. The problem we have: every now and then we get a slow replication time from the hub to a cluster mate, sometimes up to 30mins. This maxes out the nserver.exe task on the "cluster mate" which causes it to respond to http requests very slowly. In the past, we have found that if a corrupt document is in the DB, it can have this affect, but on those occasions, the server log will show the corrupt doc noteId, we run fixup, all well. But we are not now seeing any record of corrupt docs. What we have noticed is if a doc with the 32K issue is present, the same thing can happen. Our only solution in that case is to run a : fixup mydb.nsf -V, which shows it is purging a 32K doc. Luckily we run a reverse proxy, so we can shut HTTP servers down without users noticing, but users do notice when a server has the problem! Has anyone else seen this occur? I have set up DDM event handlers for many of the replication events. I have set the replication time out limit to 5 mins (the max we usually see under full user load is 0.1min), to prevent it rep'ing for 30mins as before. This ia a temporary work around. Does anyone know of a DDM event to trap the 32K issue? we could at least then send alert. Regarding 32K issue: this prob needs another thread, but we are finding this relatively hard to find the source of the issue as the 32K event is fairly rare. Our app is fairly complex, interacting with various other external web services, with 2 way data transfer. But if we do encounter a 32K doc, we can't look at field properties, so we can't work out which field has issue which would give us a clue as to which process is culprit. As above, we run a fixup -V. Any help\ comments on this would be gratefully received.

    Read the article

  • Is my external hard drive dying?

    - by thatotherguy
    I keep getting this error lately when I try to copy large (200+ MB) files over to my external. Following this, the disk becomes unresponsive and I got to unplug it and plug it in again to work. The copy process also is unreasonably slow. It is worth noting that this happens on Windows too, so it's not the notorious "Error -36" bug OS X had prior to 10.6.3. The disk is a Western Digital 3200BMV. Any ideas?

    Read the article

  • Xubuntu Running Slower than windows with constant freezing

    - by Joseph
    So I switched my computer to Ubuntu after some issues with Windows, then switched the GUI to Xubuntu after I noticed Ubuntu and Unity were painfully slow. I'm running an i5 with 8 gigs of ram and I still experience at least one full freeze (where I can't do anything and have to unplug) per week. REISUB does nothing and never has. I am constantly running Virtualbox because I still need Windows. Any thoughts to prevent this annoying freezing?

    Read the article

  • Cannot download Microsoft Project 2010 demo videos

    - by Nam Gi VU
    Hi everyone, I need to download the demo videos at my office so that I can view it later at home since I have slow internet connection there but it seems to be no way to download the videos at [the site][1], and other resources in this site. Using some of the firefox addons to download them is not possible. Do you have any tips for me? Please share. [1]: http://www.microsoft.com/showcase/en/US/search?phrase="project conference 2009"

    Read the article

  • Does urandom share the same entropy of random?

    - by ???
    Does the entropy pool /dev/random used the same to /dev/urandom? I want to mknod /dev/random 1 9 to replace the slow random, I think the current entropy is randomly enough, if urandom is based on the same entropy, and all succeed random numbers are generated based on that entropy, I don't think there'll be any vulnerable.

    Read the article

  • How to make sure all mails using Postfix?

    - by Ngompol2
    I'm using Ubuntu 10.10 32 bit. This is new server with nginx, php-fpm and PHP 5.3 I will install postfix. Currently the server can send mail (maybe sending through sendmail) but very slow until PHP timeout. To install, I will run: sudo apt-get install php-pear sudo pear install mail sudo pear install Net_SMTP sudo pear install Auth_SASL sudo pear install mail_mime sudo apt-get install postfix But after Postfix installed, how to make sure all mails using Postfix?

    Read the article

  • Performing task on remote window server 2003 machine

    - by Vaibhav Jain
    I have to perform various task such as restarting process, monitor process state check disk space, log monitoring on Window server 2003 machine. For this i am using remote desktop access which is very slow. Is there any alternate (Tool or Framework) for windows server where i can execute my script on my machine and the required task will be performed on the remote machine in somewhat interactive manner (like putty in linux)

    Read the article

  • Is it possible to download A torrent file completely and quickly with no uploading, but let the file be uploaded by other users after it has finished?

    - by B-Ballerl
    Is it possible to download A torrent file completely and quickly with no uploading, and then after seed the file to be uploaded by other users? I'm finding that the downloading process is incredibly slow while downloading some torrent file with incredibly high upload rates. Is it possible to speed up the downloading rate by cutting off uploading completley until the file has been completely downloaded??

    Read the article

  • Architectural advice - web camera remote access

    - by Alan Hollis
    I'm looking for architectural advice. I have a client who I've built a website for which essentially allows users to view their web cameras remotely. The current flow of data is as follows: User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms Ftp connection is enabled for the cameras ftp user. Web camera opens ftp connection to server. Web camera begins taking photos. Web camera sends photo to ftp server. On image url request: Server reads latest image on hard drive uploaded via ftp for camera. Server deleted any older images from the server. This is working okay at the moment for a small amount of users/cameras ( about 10 users and around the same amount of cameras), but we're starting to worrying about the scalability of this approach. My original plan was instead of having the files read from the server, the web server would open up an ftp connection to the web server and read the latest images directly from there meaning we should have been able to scale horizontally fairly easily. But ftp connection establishment times were too slow ( mainly due to the fact that PHP out of the ox is unable to persist ftp connections ) and so we abandoned this approach and went straight for reading from the hard drive. The firmware provider for the cameras state they're able to build a http client which instead of using ftp to upload the image could post the image to a web server. This seems plausible enough to me, but I'm looking for some architectural advice. My current thought is a simple Nginx/PHP/Redis stack. Web camera issues post requests of latest image to Nginx/PHP and the latest image for that camera is stored in Redis. The clients can then pull the latest image from Redis which should be extremely quick as the images will always be stored in memory. The data flow would then become: User opens page to view web camera image. Javascript script polls url on server ( appended with unique timestamp ) every 1000ms Camera is sent an http request to start posting images to a provided url Web camera begins taking photos. Web camera sends post requests to server as fast as it can On image url request: Server reads latest image from redis Server tells redis to delete later image My questions are: Are there any greater overheads of transferring images via HTTP instead of FTP? Is there a simple way to calculate how many potential cameras we could have streaming at once? Is there any way to prevent potentially DOS'ing our own servers due to web camera requests? Is Redis a good solution to this problem? Should I abandon PHP/Ngix combination and go for something else? Is this proposed solution actually any good? Will adding HTTPs to the mix cause posting the image to become too slow? Thanks in advance Alan

    Read the article

  • Checking if Intel VT-x acceleration is enabled from inside a VMware virtual machine?

    - by user269950
    My (Fortune 500) company just rolled out new VMs and everyone is complaining they are dog slow. Is there any way I could verify, from inside a VM, whether Intel virtualization (VT-x) acceleration has been properly enabled? The processor claims to be a Xeon E7-2830 but the experience has been more like a first-gen Atom. I'd ask IT directly but I get the impression they're unlikely to respond to any suggestion that they are, in fact, drooling imbeciles.

    Read the article

  • On ubuntu 10.04, what is the recommended RoR stack?

    - by Kurucu
    I can't find clear answers / methods on this. As seen elsewhere, passenger and RoR under apache gobble up ram on my VPS. I've tried a multitude of stacks and implementations, currently resting on a sub optimal apache/cgi/rails configuration, which has swapped my ram usage for CPU time and slow response to requests. Can anyone recommend an efficient and preferably simple to administer method of setting up rails apps in ubuntu 10.04 server?

    Read the article

  • Open Office: How to disable image link updates

    - by Max Kielland
    I'm writing a user manual to a card game and there is a looot of linked images. Open Office is working so slow because every time I flip to a page with linked images it starts to update them. Is it possible to tell Open Office to NOT update the links until I tell it to do so? I would like it to display the same snapshot it showed the last time I initiated link update. I'm using Open Office v3.3.0 // Thank you.

    Read the article

  • Radeon X1300 acceleration

    - by user30966
    Hello, I've just bought a Samsung XL2370 with a native resolution of 1920x1080. Should a Radeon X1300 be capable of shifting around windows on a screen that size? Because maximising and minimising windows, scrolling in Firefox and VS2008 seems very slow and jerky. Does the Radeon X1300 have any hardware accelearation? My old display was only 1028x768 and I never noticed any problems. Maybe is it time to buy a new graphics card? Thanks, AJ

    Read the article

  • Rendering videos slower

    - by asrijaal
    Hi there, I've got some movies which should be used for tutorial aspects. Problem: the shown stuff is too fast, I need to slow down the movies. Does anyone know a tool / programm which can do this? The movies are in mp4 format.

    Read the article

  • Availability of big files on multiple servers

    - by Imises
    I have to handle many (1'000 - 30'000) big files ranging from 200MB up to 2GB. The demand for these files is variable (0 - 300 downloads / file). This is why a single file must saved on 2 or more servers. My servers are placed in different datacenters (France), with different size HDDs (750GB to 4TB). Currently I share the files using PHP and ncftpget / ncftpput, but it's very slow. I need a solution to handle balancing these files across 7+ servers.

    Read the article

  • Partitioning Windows 8.1 in order to have data partition first and system partitions at HDD end

    - by LivingSilver94
    How can I have recovery partition at the HDD end? My aim is to have data partition installed on fastest sectors of my hard drive, as I don't really care if restoring operations are slow... During Windows installation, when I create a partition, I get immediately created EFI and recovery ones just before the actual data partition. EFI position is good, I want my PC boot fast, but I want to move partitions I don't care about speed. I've also considered GParted, but I think I'm not able to use it :P

    Read the article

  • Is there a list of programs for yum

    - by scriptingalias
    Basically I would like to know if there's is an actual web page that can be searched for the programs available under yum. I have yumex and I've tried using it but its super slow to search (sometimes it takes 5 minutes) and I would like some webpage or other method of doing a search. thanks,

    Read the article

< Previous Page | 99 100 101 102 103 104 105 106 107 108 109 110  | Next Page >