Search Results

Search found 21227 results on 850 pages for 'zombie process'.

Page 524/850 | < Previous Page | 520 521 522 523 524 525 526 527 528 529 530 531  | Next Page >

  • Black Screen after installing recommended Nvidia drivers. What to do?

    - by former_Windows_user
    New to Ubuntu. Problem description: Until recently I had Windows on my computer. My hard disk is divided into two partitions. On the first one (app. 10 GB) I had my Windows XP On the second one (app. 30 GB) I have some data I tried to install Ubuntu 12.04 on the first partition (the smaller one). Since I wanted to keep the data on my second partition, I chose the third install option. During the installation process I deleted the data on partition one, created a new partition with the same size, formatted it as ext4 and mounted / on it. The installation continued fine and at the end I restarted and took the CD out when it ejected automatically (it could have been also before the restart). Ubuntu started but I noticed that my computer was slow. Then a prompt appeared telling me that I did not have the optimal NVidia drivers and recommended to install a specific one. I clicked on the recommended driver, installation went apparently just fine and at the end I had to restart the system again. I did it, Ubuntu started, asked for my password, I typed it, pressed Enter, the screen turned black and remained like that (only the cursor was there and I could move it). I restarted and the same thing happened again. Has anyone had such a problem before and was able to solve it? With Windows I always installed drivers from CDs after installing Windows. Are the same CDs going to work for Ubuntu too or I should find special drivers? P.S. During the installation I was connected to the internet and I agreed on installing updates and the third party software. In the time before I installed that problematic but recommended NVidia driver I checked that there was between 6 and 7 GB free space on the first partition where I installed Ubuntu.

    Read the article

  • Is this too much to ask for a game programming and developing enthusiast? Am I doing this wrong?

    - by I_Question_Things_Deeply
    I have been a computer-fanatic for almost a decade now. I've always loved and wondered how computers work, even from the purest, lowest hardware level to the very smallest pixel on the screen, and all the software around that. That seems to be my problem though ... as I try to write code (I'm pretty fluent at C++) I always sit there enormous amounts of time in front of a text-editor wondering how every line, statement, datum, function, etc. will correspond to every Assembly and machine instruction performed to do absolutely everything necessary for the kernel to allocate memory to run my compiled program, and all of the other hardware being used as well. For example ... I would write cout << "Before memory changed" << endl; and run the debugger to get the Assembly for this, and then try and reverse disassemble the Assembly to machine code based on my ISA, and then research every .dll, library file, linked library, linking process, linker source code of the program, the make file, the kernel I'm using's steps of processing this compilation, the hardware's part aside from the processor (e.g. video card, sound card, chipset, cache latency, byte-sized registers, calling convention use, DDR3 RAM and disk drive, filesystem functioning and so many other things). Am I going about programming wrong? I mean I feel I should know everything that goes on underneath English syntax on a computer program. But the problem is that the more I research every little thing the less I actually accomplish at all. I can never finish anything because of this mentality, yet I feel compelled to know everything... what should I do?

    Read the article

  • Is there a difference in page fault rates between CPU bound and I/O bound processes?

    - by user198864
    I was thinking, should there be any difference in expectation of the page fault rate on CPU-bound vs I/O bound processes? At first I thought maybe we could, since CPU-bound processes would likely be using more memory accesses per time quantum, so I expect it would move from locality to locality faster. At the same time, the CPU-bound process is probably given a larger working set... but this doesn't affect the fault overhead as it hits a new locality IF this wasn't pre-paged in. Is there actually any real difference in the page fault rates or am I just musing about something nonexistent? And if there is, how would it impact a real-world OS like linux?

    Read the article

  • Is there a Linux mail server with an outgoing pickup directory?

    - by Paul D'Ambra
    On my Exchange server I can drop appropriately formatted text files in the "pickup" directory and Exchange will process them. I'd like to split this bulk mailing functionality onto another box to protect our business mail IP from the bumpy ride that our monthly newsletter gives us. I should note at this point that the mailing is opt-in with an opt out link included and only goes to people who pay to be a member of our organisation The ideal solution for me would be to add a linux box to use just for this purpose so we're not paying for Exchange licenses. So is there a linux equivalent of the Exchange pickup directory?

    Read the article

  • Is there a Linux mail server with an outgoing pickup directory?

    - by Paul D'Ambra
    On my Exchange server I can drop appropriately formatted text files in the "pickup" directory and Exchange will process them. I'd like to split this bulk mailing functionality onto another box to protect our business mail IP from the bumpy ride that our monthly newsletter gives us. I should note at this point that the mailing is opt-in with an opt out link included and only goes to people who pay to be a member of our organisation The ideal solution for me would be to add a linux box to use just for this purpose so we're not paying for Exchange licenses. So is there a linux equivalent of the Exchange pickup directory?

    Read the article

  • SQL Server 2008 R2 installation Error "Not all privileges or groups referenced are assigned to the caller"

    - by CodeSlinger
    Im getting an error half way through an SQL Server 2008 r2 installation. The error states Not all privileges or groups referenced are assigned to the caller. it asks to retry or cancel. Upon trying both, the error message returns and I must cntrl-alt-delete to end the process. I have checked all permissions associated with local account, network domain account and ran installation as administrator. I have searched on the web and other people are having this problem too but I cant find a solution. so I turn to the experts, anyone encounter this error??

    Read the article

  • Archive format & tool for large amounts of data (50gb+)

    - by marcusstarnes
    I only realised this afternoon that the ZIP format has a limit of what appears to be around 20gb. I am trying to automate an archive process (using Automate) to zip/rar/whatever a collection of folders/files on one of my disks. It always appeared to bomb out with an incomplete archive at about 20gb. So I tried using WinRAR and doing it manually as a ZIP file, but it told me of the limit. So, I was wondering, what is a recommended zip format (and tool for accomplishing the task) for archiving up a large amount of data (around 50gb)?

    Read the article

  • I get this error after upgade. please help

    - by user203404
    dpkg: dependency problems prevent configuration of initramfs-tools: initramfs-tools depends on initramfs-tools-bin (<< 0.99ubuntu13.2.1~); however: Version of initramfs-tools-bin on system is 0.103ubuntu0.2. klibc-utils (2.0.1-1ubuntu2) breaks initramfs-tools (<< 0.103) and is installed. Version of initramfs-tools to be configured is 0.99ubuntu13.2. dpkg: error processing initramfs-tools (--configure): dependency problems - leaving unconfigured dpkg: dependency problems prevent configuration of plymouth: plymouth depends on initramfs-tools; however: Package initramfs-tools is not configured yet. dpkg: error processing plymouth (--configure): dependency problems - leaving unconfigured dpkg: dependency problems prevent configuration of mountall: mountall depends on plymouth; however: Package plymouth is not configured yet. dpkg: error processing mountall (--configure): dependency problems - leaving unconfigured No apport report written because MaxReports is reached already No apport report written because MaxReports is reached already No apport report written because MaxReports is reached already dpkg: dependency problems prevent configuration of initscripts: initscripts depends on mountall (>= 2.28); however: Package mountall is not configured yet. dpkg: error processing initscripts (--configure): dependency problems - leaving unconfigured dpkg: dependency problems prevent configuration of upstart: upstart depends on initscripts; however: Package initscripts is not configured yet. upstart depends on mountall; however: Package mountall is not configured yet. dpkg: error processing upstart (--configure): dependency problems - leaving unconfigured dpkg: dependency problems prevent configuration of passwd: passwd depends on upstart-job; however: Package upstart-job is not installed. Package upstart which provides upstart-job is not configured yet. dpkg: error processing passwd (--configure): dependency problems - leaving unconfigured No apport report written because MaxReports is reached already No apport report written because MaxReports is reached already No apport report written because MaxReports is reached already Errors were encountered while processing: initramfs-tools plymouth mountall initscripts upstart passwd E: Sub-process /usr/bin/dpkg returned an error code (1)

    Read the article

  • Saving backup files automatically in (g)Vim after saving a file.

    - by Somebody still uses you MS-DOS
    I had a problem with my gVim. I lost some important modifications after I plugged on my machine after a hibernating process. To avoid this kind of problem, I would like to know if it's possible to add something in my .vimrc (or a plugin) that automatically backups all saving made to my files. Disk space is not an issue, I can delete these files after. I'm already using set backup set backupdir=~/.backup/vim set directory=~/.swap/vim This creates a myfile.extension~ in my .backup/vim. ...but I would like this configuration to add ~ to first save, ~0 to second, ~1 to third, ~2 to fourth, and so on - something that keeps copies from all modifications I made to a file. Is this possible? Do you know if there's a plugin for this?

    Read the article

  • What is the procedure to replace a failing hard drive in a RAID array?

    - by slayton
    3 years ago a co-worker setup a software RAID-6 array on Ubuntu 9.04 and I'm getting messages from the OS that the drive has bad sectors and should be replaced. I'd like to remove this drive and replace it with a new drive, however, I have never done this before and I'm terrified that in the process of fixing the array I'm going to end up ruining it. I know the device ID of the array and I know the device IDs of the individual drives in the array. Additionally I physically have the bad drive. What are the steps to replace the bad drive with a new drive and get the array running again?

    Read the article

  • SQL Server 2005 transactional replication break before a configured number of retries

    - by ti2
    We have a SQL Server 2000 Standard database with some tables being replicated (continuous transactional replication) to dozens of SQL Server 2005 Express and MSDE computers. The step 2 of the replication agent job (Run agent) is configured by default to retry every 1 minute for 10 times if some problem ocurr. Because the client machines get shut down at night (they are POS machines), we changed the number of retries to 5760 (4 days), so replication would not be broken at night and would not need to be restarted manually. But the problem is that every other day we have at least one machine with broken replication, with this error: The process could not connect to Subscriber 'POS986'. NOTE: The step was retried the requested number of times (5760) without succeeding. The step failed. It seems that SQL Server is not respecting the number of retries or the interval between retries as we configured. PS: I have restarted the replication job after changing the number of retries from 10 to 5760.

    Read the article

  • How do you coordinate with co-workers to give a balanced interview?

    - by goldierox
    My company has been conducting a lot of interviews lately for candidates with various experience levels, ranging from interns to senior candidates. We put our candidates through five 45 minute interview sessions where we try to ask a range of questions. One person always asks the same questions that test logic and communication. The rest typically split time between a whiteboard coding question and a discussion of previous projects, technologies the interviewee has worked with, and what he/she is looking for a job. Generally, we know the range of questions that other people on the loop will ask. Sometimes we switch things up and end up having redundancies. Today, 3 interviewers asked tree-related questions. Other times, we've all honed in on the same project on a resume and have had the interviewee talk about it with everyone. I think a smooth interview process would help us learn more about the candidate while giving the impression to the candidate that we have our act together as a team. How do you coordinate with others in the interview loop to give a balanced interview?

    Read the article

  • Unix / linux permissions setup for shared hosting with Apache

    - by weiyin
    I'm in the process of setting up a server from a clean CentOS 5 install. What is the best permission structure (users, groups, unix permissions) for running a single instance of apache for multiple users? Ideally, it should satisfy these requirements: Each user's websites are stored in a subdirectory of their home directory. Users can edit files and permissions. Apache can read the websites of all users. No user can read the website files of other users. Bonus question: how to add PHP and/or Perl and/or Ruby to Apache without allowing any users to access any other user's files?

    Read the article

  • Mac OS 10.7 Missing Java Completely

    - by Stuartsoft
    I'm in dire straights. I was trying to replace my java JDK and downgrade to 1.6 and somehow managed to completely screw up all the previous versions in the process. Bottom line, my Mac has no JDK installed at all. I've tried reinstalling java 1.7 from Oracle, I've tried using Pacifist to manually extract the files from the 1.6 Apple Java... nothing. When I open terminal and use java -version all I get is -bash: java: command not found My real goal is just to get back to java 1.7, but even after running the installer, java is still inaccessible to terminal and other applications.

    Read the article

  • "Access denied" to C:\Documents and Settings after removing malware?

    - by Rising Star
    My Windows 7 PC became infected with the so-called "Malware Protection designed to protect" trojan while I was at work the other day. I managed to kill the process so that the malware is no longer running. The removal instructions specify to delete the following file: c:\documents and settings\all users\application data\defender.exe However, when I click to c:\documents and settings, it says "Access denied". Prior to this malware infection, I've never had any trouble accessing "Documents and Settings" or "Application Data." I read that in Windows 7, c:\documents and settings is a psudonym for c:\users, but I still cannot find the file defender.exe. Suggestions?

    Read the article

  • Inbox lock for exclusive access [duplicate]

    - by user212051
    This question already has an answer here: Dovecot pop3: Disconnected for inactivity 2 answers -I found server logged into mailbox on my smtp server -This server released connection for inactivity after 10 minutes. -in the 10 minutes between logged in & disconnected for inactivity, 3 attempts to send message from 3 different clients to this mailbox failed due to unable to lock for exclusive access: Resource temporarily unavailable -after disconnection the 3 messages reached mailbox good. I tried to simulate the process and lock test mailbox but I couldn't, I was aiming to understand who can lock ? who has exclusive access ? and why only client server can lock ? and how to solve this ?

    Read the article

  • enable tcp_syncookies even after reboot

    - by Tim
    I'm running Scientific Linux 6.1 and would like to set net.ipv4.tcp_syncookies=1. I've set that in /etc/sysctl.conf and, if I do a sysctl -p then sysctl -q net.ipv4.tcp_syncookies it shows it's properly set. Sadly, if I reboot the machine, and sysctl -q again, it goes back to 0. I've tried to grep around and see if something else is resetting it to 0 during the boot process but haven't turned up anything. I've googled and everything points to sysctl.conf. The only thing I can think of is maybe networking isn't up by the time that file gets read but, honestly, I'm a developer and well beyond my natural skills here:) I'm tempted to just set it directly in /etc/init.d/network but then that feels hackish and so, I thought better of it and I'm here in search of the "right" way to do it. Any pointers?

    Read the article

  • Utilising a Magento server cluster to drive hot reindexing

    - by WOBenji
    We've asked a similar question in the past, basically we have a very large Magento store with 500000 products which are currently reindexed once a day, during the night. We'd like to speed this process up significantly, we're at about 4-5 hours now. The solution was suggested for us to do something like this on a server cluster and replicate the database changes after they've been done on a machine that isn't being bothered with serving customers. But what is the mechanism for that? How do we replicate those changes across to the live site from the server cluster? Can someone point me in the right direction here?

    Read the article

  • Programmatically closing a dialog box - win32 [migrated]

    - by user1697579
    I'm working on an application to detect a pop-up dialog and then automatically dismiss it. I'm writing this as a C++/Win32 app. The dialog box is generated by IE 7 and I can detect the window, but several methods to get the OK button to "click" have failed. Doing searches for other people's solutions, sending these messages to the button handle seems to have worked in a lot of situations: PostMessage( handle, WM_LBUTTONDOWN, 0, 0 ); PostMessage( handle, WM_LBUTTONUP, 0, 0 ); PostMessage( handle, BM_SETSTATE, 1, 0 ); It has no effect on the button state in my tests though. I can send tab characters to the main window and see that the OK button gets focus, but then sending return characters does nothing. To learn more about this I used Spy++ to get information about the window hierarchy and what messages are delievered when I manually click the OK button. Looking at the message log and reading about WM_MOUSEACTIVATE seamed to offer a solution. The log info shows that 0002166C was the button window. So in my code I tried this: GetClassNameA( handle, str, str_size ); if( strcmp( str, "Internet Explorer_Server" ) != 0 ) return TRUE; // Not the window we're interested in. // Send a message to activate the button window and have it process a mouse click. PostMessage( handle, WM_MOUSEACTIVATE, (WPARAM) dialog_handle, MAKELPARAM( HTCLIENT, WM_LBUTTONDOWN ); Based on the window hierarchy and message log, I think the window with the class name "Internet Explorer_Server" is the button. Maybe I'm wrong, because it does seem like an odd class name for a button... Below is a link to the window hierarchy image, message log when I manually click the OK button. Last is the code that's executed on a 1 second timer ticket, looking for the window. Any insight and help is appreciated! Image of the window hierarchy, source, window messages, and test dialog source are available here: https://sites.google.com/site/matthewmillersmiscellanea/Home/

    Read the article

  • Slow network file copy on Windows 7

    - by Jason V.
    I wrote a program that uses xcopy to transfer files (usually between 1KB and 2MB) over our intranet. Usually, I am copying files from my host machine (Windows 7 x64) to a VMWare virtual machine running Windows Server 2008 (the VM is running on my host machine, if that matters). On Windows XP, the file transfers usually only require a few seconds to complete. But on my Windows 7 machine, the transfer of the first file (1.5 MB) takes around 1.5 minutes to complete. This is true whether I use xcopy, robocopy, or programmatically using File.Copy(). I noticed that if I use File.Copy, the first transfer is very slow and subsequent transfers are much faster. Any clue how I can speed up the process? Is there a setting in Windows 7 (or server 2008) that I could try?

    Read the article

  • Why would an ext3 filsystem be rolled back on a Debian VM running in VirtualBox after loss of power to the host

    - by Sevas
    A Debian Virtual machine runs as a Guest VirtualBox VM. It's filesystem is EXT3. The host system loses power and after booting up the host system and guest VM, I find that the VM's filesystem has been rolled back to a previous state, losing changes made to the filesystem some time before losing power. The operations that were rolled back had been fully completed before the loss of power (files fully copied, file handles closed, etc.), but it's possible and even likely that other write operations were occuring on the VM at the point of the crash. So I am trying to figure out if it's the filesystem recovery process that rolls back filesystem operations after encountering corruption post power loss, or is it possibly related to VirtualBox and the way it ignores flush requests for performance gains by default (discussed here) Are there any other factors that would result in the filesystem being rolled back after losing power?

    Read the article

  • Is it possible to extend the Active Directory schema in a Windows 2003 DC (NOT R2) to support DFSR?

    - by JohannesH
    We're in the process of installing a brand new Windows Server 2008 Web cluster and we would like to synchronize some files between the servers. The problem is that the DC in the domain is an old Windows Server 2003 Standard (NOT R2) which apparently doesn't contain some extension to the AD schema. Is it possible to upgrade the schema without upgrading the DC servers to R2? When I try to create a Replication Group on the 2008 Server I get the following message: --------------------------- Error --------------------------- srv.XXXXXX.XX: The Active Directory Domain Services schema on domain controller activedc07.srv.XXXXXX.XX cannot be read. This error might be caused by a schema that has not been extended, or was extended improperly. See Help and Support Center for information about extending the Active Directory Domain Services schema. Schema version 30 is not supported. --------------------------- OK ---------------------------

    Read the article

  • need help setting up a VPN for remote computer connection

    - by Chowdan
    I am on a low budget right now. I am currently in the process of starting a computer company. I am in need of a VPN network so I can run Dameware tools for working on customers/partners computers remotely. I will be working with Windows and some Apple and linux machines. I have desktop with an AMD Phenom II 965BE(currently running stable at 3.8Ghz) processor with 8 GB of ram and a radeon hd 6870(i know graphics aren't too useful) and about 1.5TB of HDD space. I am attempting to create a network out of my office based all on one machine that would also be secure for me to remotely connect to my partners computers so when they have issues I would be able to connect and do the diagnosing and repairs remotely. What types of servers besides a VPN server would i need to create this? I have access to all Microsoft products so I can run Windows Server 2012, Windows Server 2008 R2, or any other Microsoft Software. thanks for the help all

    Read the article

  • Find RARs with duplicate content

    - by Scott McClenning
    I need a utility to find RAR files that contain duplicate data (i.e. files within the RAR that hash the same, but could have different names). I can open the RARs and see the CRCs are the same, but I was hoping for a more automated process that would work in bulk (hundreds of files). Hashing the overall RAR won't help because the file contained within could have different names, or the archive could be compressed at different levels. If needed, a utility that would extract the contents of the RARs and then compare would work, but is not preferred. I would prefer a free utility for Windows, but a pay utility or a utility for Linux would be acceptable.

    Read the article

  • Has anyone seen an HTTP 500 error when HTTPS traffic going through Pound Proxy forwards to an HTTP page?

    - by scientastic
    We have Varnish as our load balancer and reverse proxy cache for normal HTTP traffic. For HTTPS traffic, we use Pound proxy to unwrap the SSL and forward to Varnish, which then forwards to the back-end servers. This is used for our "checkout" process to encrypt credit card info in transition. However, on the last stage of checkout, users are always getting an HTTP 500 (Internal Server) error. It doesn't seem to be due to our back-end app server, by all tests I've tried. Does anyone know anything about how that transition works-- the transition back from HTTPS to HTTP and the interaction between Pound and Varnish-- and why it might cause 500 errors?

    Read the article

< Previous Page | 520 521 522 523 524 525 526 527 528 529 530 531  | Next Page >