Search Results

Search found 4412 results on 177 pages for 'paul est'.

Page 64/177 | < Previous Page | 60 61 62 63 64 65 66 67 68 69 70 71  | Next Page >

  • Unable to read data from the transport connection: An existing connection was forcibly closed by the remote host.

    - by Paul J. Warner
    I am having an issue with a program where after 6 mins +- 5 secs we get the above exception. Some more info about the exception stacktrace is below. This all happens pretty religiously, 6 mins goes by and bam the following 3 exeptions. We have the application installed in 2 other environments and it is working fine there. I am hoping to find some server settings either IIS 6 or Server 2003 settings that may be causing this issue to occur. I have reviewed some of the similar questions and don't see very many answers. I am hoping that maybe the information I have provided may help a little bit. 208741,Exception,,,,2011-06-21 00:30:14.193,SERVERNAME,2624,1,CLIENTNAME,The underlying connection was closed: An unexpected error occurred on a receive. , at System.Web.Services.Protocols.WebClientProtocol.GetWebResponse(WebRequest request) at System.Web.Services.Protocols.HttpWebClientProtocol.GetWebResponse(WebRequest request) at Microsoft.Web.Services3.WebServicesClientProtocol.GetResponse(WebRequest request, IAsyncResult result) at System.Web.Services.Protocols.SoapHttpClientProtocol.Invoke(String methodName, Object[] parameters) at System.Net.Sockets.NetworkStream.Read(Byte[] buffer, Int32 offset, Int32 size) at System.Net.FixedSizeReader.ReadPacket(Byte[] buffer, Int32 offset, Int32 count) at System.Net.Security._SslStream.StartFrameHeader(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest) at System.Net.Security._SslStream.StartReading(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest) at System.Net.Security._SslStream.ProcessRead(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest) at System.Net.TlsStream.Read(Byte[] buffer, Int32 offset, Int32 size) at System.Net.PooledStream.Read(Byte[] buffer, Int32 offset, Int32 size) at System.Net.Connection.SyncRead(HttpWebRequest request, Boolean userRetrievedStream, Boolean probeRead),2004437127,114,1 208742,Exception,,,,2011-06-21 00:30:14.227,SERVERNAME,2624,1,CLIENTNAME,Unable to read data from the transport connection: An existing connection was forcibly closed by the remote host. , at System.Net.Sockets.NetworkStream.Read(Byte[] buffer, Int32 offset, Int32 size) at System.Net.FixedSizeReader.ReadPacket(Byte[] buffer, Int32 offset, Int32 count) at System.Net.Security._SslStream.StartFrameHeader(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest) at System.Net.Security._SslStream.StartReading(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest) at System.Net.Security._SslStream.ProcessRead(Byte[] buffer, Int32 offset, Int32 count, AsyncProtocolRequest asyncRequest) at System.Net.TlsStream.Read(Byte[] buffer, Int32 offset, Int32 size) at System.Net.PooledStream.Read(Byte[] buffer, Int32 offset, Int32 size) at System.Net.Connection.SyncRead(HttpWebRequest request, Boolean userRetrievedStream, Boolean probeRead),2004437127,114,1 208743,Exception,,,,2011-06-21 00:30:14.287,SERVERNAME,2624,1,CLIENTNAME,An existing connection was forcibly closed by the remote host , at System.Net.Sockets.NetworkStream.Read(Byte[] buffer, Int32 offset, Int32 size),-691097507,62,1

    Read the article

  • Running git-svn with cron results in garbage in .git

    - by Paul
    I've setup a git-svn repo with cron to fetch from the svn repo daily. I have a script to do the fetching, and this is what is invoked by cron. Everything is fine with the repo, and the script works fine when executed manually. However, when it runs under cron, empty files get dropped into the .git directory. The files have names that look like they are some base64 output, e.g. juTrvjP6m8 and kcKf3hu3b4. Two of these files show up for every cron run. I thought these might be commit hashes, but they're not, git-show says it's an unknown revision. I set-up the repo as follows: git svn init http://svn.ip.addr/repo git svn fetch svn-remote My script looks like this: cd /gitsvn/dir git svn fetch svn-remote git svn push pub The last line pushes the repo to a separate (bare) public repo from which others can clone. I'm piping the output from the cron job to a file, which looks like this: fatal: unable to run 'git-svn' Counting objects: 21, done. Delta compression using up to 2 threads. Compressing objects: 100% (10/10), done. Writing objects: 100% (11/11), 59.08 KiB, done. Total 11 (delta 8), reused 0 (delta 0) To /gitpub/repo.git 360faf5..a153b0d trunk -> trunk The line "fatal: unable to run 'git-svn'" is alarming, but the fetch seems to go ahead anyway. Any suggestions? Where are these empty garbage files coming from, and how to stop them? Am I in for bigger problems in the future? BTW, I'm using git 1.6.3.3.

    Read the article

  • IMAP mailboxes Backup script...

    - by Paul Stevens
    Hi guys, I need some advise to create a script that runs at night on my IMAP mail server, and backup all the mailboxes (50) and once the backup is done, burn such file in a DVD. What could be the best procedure to accomplish an IMAP backup and what tools should I use?.. Thanks

    Read the article

  • Myvidoop error when trying to log in

    - by paul-werner.mp
    Recently I have not been able to get into my myvidoop username. It won't even bring up the pictures and asks to send the access code to one of my email addresses. I have seen this message before and there was no error previously but now it will not send the access code. I've tried accessing the site through firefox and through IE but I get the same thing (whether I went through the add on to get to the site or just typed in the website). I just want to know if I possibly have something that could be blocking it from sending this out. I use the free version of Avira for protection, run Ad Muncher (but I've also tried to get this access code with Ad Muncher closed), and have Ad Block Pro plugin (but like I've said before, the same thing happens in IE). I know it uses Java to send this access code out and I've even updated Java but still cannot get it to send it to me.

    Read the article

  • How do you set the Scheduled Tasks "Run only if logged in" option from the command line?

    - by Paul K
    I want to create some tasks with the "Run only if logged in" option set, but using a batch file, as you can in the user interface version of the schtasks.exe tool. I'm doing this so I don't have to specify a password, either in the batch file, or when the batch file is run... I would use /ru System but I can't because the tool I'm automating (SyncToy 2.1) won't run with that account because I am hitting a network drive... Also, I noticed that Google Chrome sets up some tasks with this option set during installation, so I'm thinking there must be a way...

    Read the article

  • Process files in a folder that haven't previously been processed

    - by Paul
    I have a series of files in a directory that I need to carry an action out on using a script. Once the action is done, then I want to keep a log that the file has been processed, so that the next time the script is run, it does not attempt to carry out the action again. So lets say I can find all the files that should be processed like this: for i in `find /logfolder -name '20*.log'` ; do process_log $i echo $i >> processedlogsfile done So I have a file containing the logs I have processed, and my goal would be to modify the for loop such that these processed logs are not processed a second time. Doing a manual scan each time seems inefficient, particularly as the processedlogfiles gets bigger: if grep -iq "$i" processdlogfiles ; then continue; fi It would be good if these files could be excluded when setting up the for loop. Note that the OS in question is a linux derivative, a managment appliance, with a limited toolset (no attr command for example) and so no way to install additional utilities (well it is possible but not an option). Most common bash shell commands are available though. Also, the filenames and locations of the processed files must remain where they are - they can't be altered to reflect their processed status

    Read the article

  • Random timeouts in IIS7

    - by Cor-Paul
    Hi, I have a weird problem which I think is caused by my IIS7 installation on Vista 64 bit. I have a bunch of AJAX calls and JS dynamic file loads (~30) in a local application, and I get random timeouts (or so it seems) in my browser. In Chrome it looks like the page just stops loading (no HDD activity), in Firefox/Firebug I can see that some of the files are being loaded but they never actually finish. When I reload the page the same occurs, but for (random) other files that must be loaded. When I try to load one of those JS files concurrently (so during the timeout in FF) in another browser, the file loads there. So I am pretty sure the request can be handled by IIS. I am thinking about a limit or so on simultaneous requests from the same browser which is not working correctly, but I am pretty clueless on how to solve this. Does anyone recognize this problem and know a solution? Thanks!

    Read the article

  • "A device attached to the system is not functioning" when reconnecting to XenApp 6 session

    - by Paul Doom
    A few of our XenApp 6 customers are seeing issue sometimes when users move from one workstation to another and try to connect to their running session. An error box pops up with "A device attached to the system is not functioning" This is on healthy 2008 R2 servers with all Citrix hotfixes applied. After the error, the user's session must be logged out before they can get back in. This was posted here: http://forums.citrix.com/thread.jspa?threadID=273457 , but is not answered. There is nothing telling in the event logs, and it is an inconsistent problem. Has anyone else seen this issue of found a fix?

    Read the article

  • Bought a 1U server, now can't find rails

    - by Paul Tomblin
    I bought a 1U server used on eBay (a Supermicro motherboard, but branded "Asempra"), and I also bought a set of "Supermicro" rails, but unfortunately the rails didn't fit the server. The side of the server looks like: Can you tell me where I can find rails for this machine?

    Read the article

  • removing the .local in local network name serving

    - by Paul Nathan
    I have several local machines; I use a OSX 10.6 machine to do most of the serving. Annoyingly, it postfixes its network name with .local. How would I set up a system so that I could access it by its hostname? server: httpd apache2 default install I am accessing it with a web browser(surprisingly). also when I ping my osx machine as name, it doesn't work; ping name.local does work.

    Read the article

  • Why does my MacBook Pro trackpad sometimes set its tracking speed to slow, all by itself?

    - by Paul D. Waite
    Every now and then, the trackpad on my MacBook Pro will seem to set its own tracking speed to slow. I’ll notice that the cursor is moving slowly, and when I check in System Preferences, the tracking speed is indeed at slow, even though I never set it to slow myself. This might happen before/after switching into a VMWare virtual machine, but I’m not sure. It doesn’t seem to happen on startup or anything, just during use. Anyone else seen this?

    Read the article

  • Windows 7 File Associations .mov

    - by Paul Mung
    I created a new windows 7 SP1 base image. Everything is all fine and dandy with that. So i am now installing standard applications I would like Quicktime to manage .mov files. The only problem is WMP (Windows Media Player) won't give up the association to .mov files. It's driving me crazy... i've been reading threads on how to fix file associations. I would like to do it via registry, powershell or cmdline. I cannot use GPO I've tried the following: assoc .mov=QuickTime.mov ftype QuickTime.mov="%ProgramFiles(x86)\QuickTime\QuickTimePlayer.exe" Reg add HKCU\Software\Microsoft\windows\CurrentVersion\Explorere\FileExts\.mov\UserChoice" /v Progid /d QuickTime.mov /f Reg add HKCU\Software\Microsoft\windows\CurrentVersion\Explorere\FileExts\.mov\OpenWithList" /v a /d QuickTimePlayer.exe /f Reg add "HKCU\Software\Microsoft\windows\CurrentVersion\Explorere\FileExts\.mov\OpenWithList" /v b /d wmplayer.exe /f Reg add HKCU\Software\Microsoft\windows\CurrentVersion\Explorere\FileExts\.mov\OpenWithList" /v MRUList /d ab /f Reg add HKCU\Software\Microsoft\windows\CurrentVersion\Explorere\FileExts\.mov\OpenWithProgids" /v Quicktime.mov /t REG_NONE /d 0000 /f Reg add HKCU\Software\Microsoft\windows\CurrentVersion\Explorere\FileExts\.mov\OpenWithProgids" /v WMP11.AssocFile.MOV /t REG_NONE /d 0000 /f

    Read the article

  • How do I run a stable Windows XP kvm guest on Ubuntu 10.04?

    - by Jean-Paul Calderone
    I have three Windows XP guests running on a recently upgraded 64-bit Ubuntu 10.04 system. Occasionally (on the order of once every few days), one of the guests will become non-responsive and the kvm process on the host which is running that guest will start consuming 100% CPU. It will continue to do so until it is killed. When restarted, it will be fine for a while, and then the issue repeats. The kvm command line used to run all three guests is this: /usr/bin/kvm -S -M pc-0.12 -enable-kvm -m 1024 -smp 1 -name bigdog21vmxp1 \ -uuid ea47ff84-125b-16f7-9a4d-a6d0d8bab46a \ -chardev socket,id=monitor,path=/var/lib/libvirt/qemu/bigdog21vmxp1.monitor,server,nowait \ -monitor chardev:monitor \ -localtime \ -boot c \ -drive file=/var/lib/libvirt/images/windowsxp-1.qcow2,if=ide,index=0,boot=on,format=qcow2 \ -net nic,macaddr=54:52:00:02:06:0e,vlan=0,name=nic.0 \ -net tap,fd=58,vlan=0,name=tap.0 \ -chardev pty,id=serial0 \ -serial chardev:serial0 \ -parallel none \ -usb \ -usbdevice tablet \ -vnc 127.0.0.1:1 \ -k en-us \ -vga cirrus \ -soundhw es1370 Why do the systems misbehave this way sometimes? And what configuration can I change in order to fix it? Or, if the problem is due to a bug in kvm, what is the process for isolating a kvm failure so that the developers have a chance of fixing it?

    Read the article

  • iPhone Missing Apps in some Directories

    - by paul simmons
    Hi, I was moving iPhone applications directory, and ended in a result that some applications are missing from Springboard. As far as I can see, the apps in /Applications/GUID-LIKE-DIRECTORIES are missing like Facebook; however I see they are in /Applications if I check with SSH. Any help?

    Read the article

  • Recommendation for e-mail forwarding

    - by Paul Mrozowski
    We currently have a provider that we use to primarily forward our e-mails to both a hosted Exchange site and to outside e-mail addresses. We have a large # of forwards set-up (maybe 500). They work fine for us but the bandwidth costs are a getting pretty expensive (approx 80GB in email a month) so we're looking for a less expensive alternative (it costs us nearly $150/mo). I've looked at just hosting this here via a PostFix server and one of the nice front-ends available for it. However, I'm not thrilled with the prospect of being responsible for maintaining yet another server, backups, network connectivity, etc. So I'm leaning towards having this hosted by someone else if possible. Does anyone have any recommendations for hosting providers? Are there some I should stay away from?

    Read the article

  • Seeking web-based FTP client for very large file upload

    - by Paul M. Nguyen
    I have looked around for these for some time... the limits imposed by the web server and/or the dynamic programming environment (e.g. PHP) are far too restrictive for the application I'm working on. We need to be able to move large graphics and video files to and from clients (ranging from tens of MB to a few GB in a single file). Plain FTP with a proper desktop client will do the trick, and we're hosting this in Amazon EC2 with EBS. User management will be done from the office via webmin. Users are chroot-jailed into their home dir by proftpd. net2ftp will work for many clients, but we often need to move single files that approach 1GB or exceed 2-3GB which is way out of the range of any http-based uploader. So we turn to Java or Flash - can they do it? From within the web browser establish an FTP connection and grab a huge file? There are licensed applets and such out there, but none seem convincing. Again, I'm looking for some code that can speak FTP and read (& write?) the local disk, that is delivered in a web browser, and can move single files of 2GB+. The reason for having a web-based interface to FTP is to skip the software installation step for our clients. I will consider proper desktop client software as long as it's "portable" and at least Win+Mac and can be easily configured by lay-man users in a hurry.

    Read the article

  • Delay after commit before Redmine comments ticket

    - by Paul
    I was used to having a post-commit hook trigger processing of the commit message by Trac (SVN repository). The comment (or action) was added immediately. Now I switched to Mercurial as a VCS and Redmine as a ticketing system. Redmine does recognize information in changeset/commit messages like "refs #185" - but it takes several minutes to appear in the ticket. Any ideas why this is? Is the routine to grab the changesets called periodically somehow?

    Read the article

  • Font name used in files shared between OSX and Windows

    - by Paul
    Our designer uses OSX, and creates InDesign or AI files. They are then passed to us for changes. When we open the files on Windows, we are told that fonts are missing. In this example, the Futura font is being used. The Windows machine has Futura installed, from BitStream. The nane of the font is "Futura Std", whereas on OSX, it is simply Futura. So InDesign chooses a random font to substitute Futura with on Windows, it does not choose Futura Std. Now we can use the Find Font feature, and change all the instances of "Futura" to "Futura Std", but if we pass the file back to the designer, they have to then do the reverse. What is the right way for managing this?

    Read the article

  • Internet Connection not working - USB LAN connection - from particular modem

    - by Paul
    I am trying to fix Internet connection on a friends Dell inspiron 1720 with XP service pack 3. It has an integrated network card that stopped working, after powering down/up the modem still didnt work I brought it back to my place to try a few things ie check cable, update driver etc... still didnt work. So I bought a USB LAN connector. It didnt work straight away but I went to configure the properties and changed the ConnectionType from AutoSense to 100 BaseT 10BaseT Full_Duplex, I basically just tried them all. From my place when connected to my desktop - 10 BaseT and 10BaseT Full_Duplex worked. From my place When connected to their laptop - 10 BaseT and 10BaseT Full_Duplex worked. Happy I went back to my friends house confident it would all work, and it didnt. Brought it back to mine and it did. While there, in Network Connections the connection is there recognized, enabled, 'working properly' it just says not connected. Also there is no led on the USB connector While at mine as above except there is an led on the USB connector and it says connected. Other difference I can think of is they have a cable modem, I'm plugged into the back of a Belkin wireless router - would this make a difference? Any other ideas what to try? (Would getting the model of the cable modem help anyone?) The USB connector is "DM9601 USB to Fast Ethernet"

    Read the article

  • Virtual box lost files

    - by Paul Lloyd
    I have been running Virtual Box on a Macbook Pro for a year or more, with an old version of Windows XP running on the virtual machine. Recently my Mac battery dies completely and I had left VB / Windows running. When I re-started the Mac, VB would not load. I reinstalled a recent download of VB from a DMG file I had stored. Now VB starts but when I try to start the virtual machine I get the following message, Failed to open a session for the virtual machine Windows XP. Failed to load VMMR0.r0 (VERR_SUPLIB_WRITE_NON_SYS_GROUP). Result Code: NS_ERROR_FAILURE (0x80004005) Component: Console Interface: IConsole {1968b7d3-e3bf-4ceb-99e0-cb7c913317bb} I have been backing up the MAC using Time Machine and have backups going back months. I really need to access the windows files as they have my Accounts and other business critical stuff. Any ideas please, or does anyone know where I can get some support for this combination of hardware / software, apologies I am a novice in this area. Thanks in advance.

    Read the article

  • Decrypt column in SQL 2008

    - by Paul
    I need to decrypt a column in a table that has previously been encrypted at application level. The algorithm is DES at 192 bits and block size = 64. I have the password but DecryptByPassPhrase doesn't seem to work.

    Read the article

  • Gnome not working well with NoMachine

    - by paul
    First of all, NoMachine is working for me. However, at some point with my Ubuntu/Gnome machine, I got an update for Gnome and now when running NoMachine I no longer get all of the theme elements. For example the panels are there but with the default or stock colors. This is true for the icons as well. Does anybody know a solution for this problem?

    Read the article

  • mod_rewrite REQUEST_FILENAME doesn't contain absolute path

    - by Paul Dixon
    I have a problem with a file test operation in a mod_rewrite RewriteCond entry which is testing whether %{REQUEST_FILENAME} exists. It seems that rather than %{REQUEST_FILENAME} being an absolute path, I'm getting a path which is rooted at the DocumentRoot instead. Configuration I have this inside a <VirtualHost> block in my apache 2.2.9 configuration: RewriteEngine on RewriteLog /tmp/rewrite.log RewriteLogLevel 5 #push virtually everything through our dispatcher script RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^/([^/]*)/?([^/]*) /dispatch.php?_c=$1&_m=$2 [qsa,L] Diagnostics attempted That rule is a common enough idiom for routing requests for non-existent files or directories through a script. Trouble is, it's firing even if a file does exist. If I remove the rule, I can request normal files just fine. But with the rule in place, these requests get directed to dispatch.php Rewrite log trace Here's what I see in the rewrite.log init rewrite engine with requested uri /test.txt applying pattern '^/([^/]*)/?([^/]*)' to uri '/test.txt' RewriteCond: input='/test.txt' pattern='!-f' => matched RewriteCond: input='/test.txt' pattern='!-d' => matched rewrite '/test.txt' -> '/dispatch.php?_c=test.txt&_m=' split uri=/dispatch.php?_c=test.txt&_m= -> uri=/dispatch.php, args=_c=test.txt&_m= local path result: /dispatch.php prefixed with document_root to /path/to/my/public_html/dispatch.php go-ahead with /path/to/my/public_html/dispatch.php [OK] So, it looks to me like the REQUEST_FILENAME is being presented as a path from the document root, rather than the file system root, which is presumably why the file test operator fails. Any pointers for resolving this gratefully received...

    Read the article

< Previous Page | 60 61 62 63 64 65 66 67 68 69 70 71  | Next Page >