Search Results

Search found 3225 results on 129 pages for 'david a gibson'.

Page 47/129 | < Previous Page | 43 44 45 46 47 48 49 50 51 52 53 54  | Next Page >

  • Completely unintuitive Apache/PHP memory-freeing behavior

    - by David
    Okay, this one's weird. I have a Turnkey Linux server with a gig of dedicated RAM. It's running WP3.2 with a boatload of plug-ins. It's a new site, so it has very limited traffic (other than search engines, maybe 20 hits a week). Now, for a few weeks, every few days, it would max out on main RAM, start eating up virtual RAM, and then crash. It's had this behavior for a while and I've been trying to figure out which element was causing the crash. Nine days ago, I pointed my external server monitor to this server. I wrote a 5-line HTML file (not PHP and not WP) that the server monitor accesses every minute, to see if the server is up. So, now, nine days later, the server has been rock solid, up all the time, no memory leak at all. I changed NOTHING on the server itself to see this behavior change. Have you EVER seen anything like this? All the server monitor is doing is retrieving a single, super-simple HTML file and all the memory leak problems have gone away. Weird, eh?

    Read the article

  • Debian 6: setting up FTP just for website editing

    - by David Oliver
    I have a VPS using Debian 6.0. Currently, SSH is set to not accept password logins, and only key-based ones. A person who needs to work on one particular website (a vhost) wishes to use FTP. He doesn't need/want SSH. How can I set up FTP access for him, enabling him to have write permissions for all files in the relevant directory, and only the relevant directory? The directory is /srv/www/domainname.com/public_html Currently, all directories and files in that directory belong to www-data:www-data and are 644/755. I've installed vsftpd and have been reading some guides, but they all seem to deal with allowing multiple users to have their own user-named directories which isn't what I'm after. I can't seem to work out how to simply define one FTP user with a password that has access to one directory of my choosing. This is my first experience of setting up an FTP server. Thanks. Edit: have also found this - maybe I should be using ProFTPd, or can vsftpd also do what I want?

    Read the article

  • How to upgrade OS on Mac Mini with external USB Drive?

    - by David
    We have a G4 Mac Mini, circa 2005, running Mac OS X 10.4 Tiger, and we want to upgrade to 10.5 Leopard. We have a Leopard install disk, but the optical drive in this mini is broken. So we transferred the install disk image to a USB HDD, but now we can't figure out how to boot off it. From what I've read in Mac forums, some PPC Macs, including some G4's, have been able to boot from USB, even though it sounds like this wasn't officially supported, and it may well depend on the specific model of USB drive and Mac. My Mac says CPU is "PowerPC G4 (1.2)" and Boot ROM is "4.8.9f4". I was hoping I might just find somebody here who had that same Mac Mini and find out if they could make it work. I'd especially like to know any specifics about the USB drive they found success with. Any insights at all would be appreciated. Thanks.

    Read the article

  • Moving data to lower sectors on an SSD without defragging

    - by David Freitag
    I bought an Intel SSD drive a while back and now I want to dual-boot with it. But for some reason there are sectors near the end of the drive filled and I can’t seem to find a way to remove the data so that I can safely shrink the partition. I know I have sectors near the end full because I am using Defraggler to analyze my drive (not to defrag it). I can see what files need to be moved/deleted but short of actually deleting some drivers and/or other necessary files, I am completely stuck. This the diskmap: I am only able to shave off that last 1.72GB of space from the drive which isn’t even enough for the most minimal Linux install.

    Read the article

  • In AD how to get groups the workstation is a member of from the workstation itself?

    - by David
    If I'm at the domain controller (in the Active Directory) to find out what groups the "XPSP3-A" workstation is a member of, I do: dsquery computer "CN=XPSP3-A,CN=Computers,DC=pvk,DC=local" -name XPSP3-A | dsget computer -memberof and receive the following: "CN=Sec Group 001,OU=OU1,DC=pvk,DC=local" "CN=Domain Computers,CN=Users,DC=pvk,DC=local" But how to do the same from the "XPSP3-A" workstation?

    Read the article

  • How do I set or remove a Windows SMB password?

    - by David Is Not Here
    I recently purchased a computer with the intention of using it as network-attached storage. It presently has no Windows password and only one account. Can I create two different shares (share two different folders) in different ways -- make one publicly accessible, and make another only visible to people with a certain username/password combination? Creating multiple users isn't a problem, I just can't figure out where to start within Windows.

    Read the article

  • How do I find the source of soft page faults?

    - by David Robison
    I have Windows 7 x64 computer that according to Performance Monitor has 70,000 page faults / second when idling. That's seems like a lot to me (every other computer I check has basically 0 page faults / second when idling). If I use Resource Monitor or Process Explorer to check hard faults, I see that they are basically 0. So all the page faults are soft. Normally, soft page faults are not a problem, but I suspect they might be causing issues for this computer given there are so many. I would like to identify what programs are causing the soft faults. Are there any tools that exist the display the number of soft page faults for each process?

    Read the article

  • Is this distributed database server idea feasible?

    - by David
    I often use SQLite for creating simple programs in companies. The database is placed on a file server. This works fine as long as there are not more than about 50 users working towards the database concurrently (though depending on whether it is reads or writes). Once there are more than this, they will notice a slowdown if there are a lot of concurrent writing on the server as lots of time is spent on locks, and there is nothing like a cache as there is no database server. The advantage of not needing a database server is that the time to set up something like a company Wiki or similar can be reduced from several months to just days. It often takes several months because some IT-department needs to order the server and it needs to conform with the company policies and security rules and it needs to be placed on the outsourced server hosting facility, which screws up and places it in the wrong localtion etc. etc. Therefore, I thought of an idea to create a distributed database server. The process would be as follows: A user on a company computer edits something on a Wiki page (which uses this database as its backend), to do this he reads a file on the local harddisk stating the ip-address of the last desktop computer to be a database server. He then tries to contact this computer directly via TCP/IP. If it does not answer, then he will read a file on the file server stating the ip-address of the last desktop computer to be a database server. If this server does not answer either, his own desktop computer will become the database server and register its ip-address in the same file. The SQL update statement can then be executed, and other desktop computers can connect to his directly. The point with this architecture is that, the higher load, the better it will function, as each desktop computer will always know the ip-address of the database server. Also, using this setup, I believe that a database placed on a fileserver could serve hundreds of desktop computers instead of the current 50 or so. I also do not believe that the load on the single desktop computer, which has become database server will ever be noticable, as there will be no hard disk operations on this desktop, only on the file server. Is this idea feasible? Does it already exist? What kind of database could support such an architecture?

    Read the article

  • Samba share will not connect (was working yesterday)

    - by David Gard
    I have a CentOS websver with a Samba share set up (\\webserver\websites). I was connected to this share just yesterday without issue, but today my Windows 8 PC will not connect to it. I've also tried making a connection from Windows 7 and Windows XP, all without success. I initially tried restarting my computer, but that did not work. I then tried restarting the Samba service on the webserver (service smb restart), and when that failed I restarted the webserver. All of that was to no avail, and I still cannot connect to the share. The webserver is contactable from my PC (and the others I tried), as the websites it hosts work fine and I'm able to Putty to the server. When connected to the webserver, I can see that Samba is running by using service smb status - service smb status smbd (pid 4685) is running... nmbd (pid 4688) is running... Can anyone please help me to get this share working? Here is my full Samba config (/etc/samba/smb.conf) - [global] workgroup = MYGROUP server string = Samba Server %v log file = /var/log/samba/log.%m max log size = 50 security = user encrypt passwords = yes socket options = TCP_NODELAY SO_RCVBUF=8192 SO_SNDBUF=8192 local master = no [websites] comment = Websites browseable = yes writable = yes path=/var/www/html/ valid users = dgard

    Read the article

  • Redirect to new domain and preserve username

    - by David Brown
    I recently switched to a new domain for a version control server I run. The server is usually accessed with a username included in the url such as https://[email protected]/some/stuff. I want to redirect requests to the old domain to the new domain and preserve everything else in the url (including the username). So the former url would be redirected to https://[email protected]/some/stuff. Currently I have the following rewrite condition and rule: RewriteCond %{HTTP_HOST} sub\.olddomain\.com RewriteRule (.*) https://sub.newdomain.net$1 [R=301,L] This works except it drops the userinfo part of the URL. Is there a way I can preserve the user info?

    Read the article

  • Disable page cache on a specific page

    - by David Brunelle
    Hi, Not sure if I really am on the right forum, but if not, just tell me. I have a page that is coded in ASP (not .net) which is used to send email. We are currently having a problem in which the page seem to be sent twice sometime. Upon checking, we found out that those who have this problem are coming from big organisation, so it was suggested that their server might cache the file for some reason. I would like to know, is there a way in HTML (or ASP ) to prevent that from happening ? Or is it in IIS that we must set this up ? Thanks,

    Read the article

  • Available filesystems for Linux that are case-insensitive?

    - by David
    I have a client whose web application was written entirely in a windows environment and served from windows. Unfortunately there's way to many cases of get "file/At/Somelocation.php" where the file is actually something horrible like "File/at/SomeLocation.PHP". I really don't want to be forced to work in Windows but it will take weeks if not longer to fix all the casing issues. Am I SOL here?

    Read the article

  • Restricting access to a subdirectory on linux

    - by David
    I'm looking for a way to make a directory accessible only to its parent directories. That is, suppose you have two directories, A and B, at the same level in the file hierarchy. Now suppose that you have a directory A' which is a subdirectory of A. I'd like to enforce that A is able to access the contents of A' but B is not. My problem is that I'd like to use a library (directory A) which builds on top of a legacy version of another library (directory A'). At the same time, I want to be able to use the newest version of this legacy library (directory B). I want to make sure that people aren't somehow using library A and linking against new library B by enforcing that library A must use library A'. I could just link A against library B, but then I'm risking compatibility.

    Read the article

  • Nginx 502 Bad Gateway: It just won't stop

    - by David
    I have the same problem that most people seem to have with Nginx: 502 bad gateway errors. They are intermittent but typically happen more than once per session, which means my users are probably running into it nearly every time they use the app. I've tried adjusting fastcgi_buffers and fastcgi_buffer_size (in both directions) to no avail. I've tried various other things with the configuration file but nothing seems to work. Here's my config (note that I've stripped away most of the things I've tried, since they didn't work and I didn't want to bloat the file with a bunch of un-related directives): server { root /usr/share/nginx/www/; index index.php; # Make site accessible from http://localhost/ server_name localhost; # Pass PHP scripts to PHP-FPM location ~ \.php { include /etc/nginx/fastcgi_params; fastcgi_pass 127.0.0.1:9000; } # Lock the site location / { auth_basic "Administrator Login"; auth_basic_user_file /usr/share/nginx/.htpasswd; } # Hide the password file location ~ /\. { deny all; } client_max_body_size 8M; } I'm running a small Rackspace cloud server, which should be plenty for handling an app with a small user base...

    Read the article

  • How to make PDF pages with U3D images print properly?

    - by David Thornley
    I'm creating some PDF files that have three-dimensional images (U3D) in them. When I bring them up in Acrobat, everything is fine until I try to print the file. If I've viewed a page, it prints fine, showing the U3D image as I last saw it. If I haven't viewed a page yet, the printout is blank. (I can demonstrate this right from the Print dialog, by previewing pages.) The only reference I've seen to printing U3D is in the PDF Standard, which says, basically, that this should work. It recommends "PV" or "PO" for the A key in the 3D activation dictionary, and I'm using libharu which uses "PV".

    Read the article

  • Secondary Monitor Flickers Off

    - by David Robison
    I have a Windows XP computer with two monitors attached. The secondary monitor occasionally flickers off and then back on. It kind of looks like it is going into power saving mode, then immediately exiting. I have tried swapping the two monitors, and this consistently happens on the whichever one is connected as the secondary one. The video card I am using is a "Gigabyte GeForce 9600GT 512MB Silent".

    Read the article

  • Internet Explorer wont open docx files, saves them as zip

    - by David Gard
    I have several docx documents on an Intranet for my work, but IE8 refuses to open them, instaed only saving them as a zip (filename_docx.zip). This seems to be only an IE8 problem (surprise, surprise!) as both FF and Chrome open the documents just fine. Unfortunately as this is work based, I cannot simply drop IE as I would, in favour of a decent browser. Does anybody know how to fix this issue in IE? Thanks.

    Read the article

  • DVD drive not recognized

    - by David Oneill
    I'm in the midst of building a computer (first time builder) I got everything plugged in, and hit the power button. After the excitement of the first boot coming up on my screen, I was looking through the BIOS settings. However, joy soon turned sour, as I looked at the list of installed SATA devices. My DVD drive wasn't on the list :( So, what are common things I should check/try?

    Read the article

  • How to determine non-movable files in Windows 7?

    - by David
    Is there a way to determine which unmovable files are preventing Shrink Volume from releasing the full potential free space? Background: I have a 90 GB partition with Windows 7 on it, and 60 GB free space. I want to shrink it down to about 40 GB, and use the reclaimed 50 GB for a separate data partition. The Shrink Volume tool in Disk Management is only willing to give me 8 GB back. My understanding is that this is because of immovable files. I've followed the instructions found here, which involved disabling hibernation, pagefile, system restore, kernal dump, making sure all related files were deleted, and defrag'ing. I have successfully followed those same instructions before on this same drive, and partitoned the original 150 GB space into 90 GB and 60 GB, but I'm not so lucky this time.

    Read the article

  • Enabling Compiz Viewport Switcher key bindings

    - by David Moles
    I'm running compiz 0.8.2 with compizconfig on Scientific Linux 6.2 with Gnome 2.28.2. In the compizconfig "General Options" I have "Desktop Size" set as follows: Horizontal Virtual Size: 6 Vertical Virtual Size: 1 Number of Desktops: 1 This gets me the layout I want, i.e. 6 workspaces in a horizontal layout. Ctrl-alt-cursor-keys work fine for switching between them. However, I can't figure out how to get key bindings for specific workspaces. I've tried enabling "Viewport Switcher" in compizconfig, and tried various combinations both in "Number-based viewport switching" and "Go to specific viewport", to no apparent effect. My first thought was that something else was eating the specific key bindings I chose, but I think I've tried every combination of shift, control, alt and super (i.e., the Windows key) by now. I tried setting 6 desktops under "General Options" instead of one desktop with horizontal virtual size 6, but that doesn't seem to make a difference either. What am I missing?

    Read the article

  • What differences are there between an official Ubuntu AMI image and a base install from an ISO?

    - by David Winter
    When creating a new instance on AWS using an official Ubuntu 12.04 server AMI, what differences are there compared to if I was to do a standard server install on a computer of my own? For example, the default user is 'ubuntu'. An SSH public key is added to that users authorized_keys file. Sudo is passwordless for that user. PasswordAuthentication is disabled for SSH. etc etc. Configurations have been changed from their defaults, and I'd like to know if there is a list, or somewhere I could find out the modifications made.

    Read the article

  • Split a big folder (and its contents) into several small new folders ?

    - by David
    Hello, By luck do you know a software on Windows (Vista) which can easily split a big folder (and its contents) into several small new folders ? notes: the new small folders should be independent the "file splitter" that I have already tested don't do alas the "folder splitter" job ! ;( Example : I have a main big folder C:\big\ (size: about 3Go) The result I need : C:\big01\ (size: about 200MB) C:\big02\ (size: about 200MB) etc... Thanks in advance ;)

    Read the article

  • Are these three brand new sticks of RAM really dead?

    - by David Brown
    I'm working on a Dell Dimension 4700 desktop for a friend. It came with 512MB of DDR2 RAM (two sticks of 256MB). One morning, it started blue screening on startup with no helpful error messages. It refused to boot into any form of Windows installation, including Safe Mode, original recovery disk, and my custom Windows PE disk. It did boot into the Ultimate Boot CD, so I ran memtest86, which reported errors everywhere. I removed one stick of RAM and the system booted up just fine. I moved the remaining stick into each slot and the system continued to operate normally, so I came to the conclusion that the stick that I removed was dead. I ordered an exact replacement, along with 2 more sticks of 256MB DDR2 (again, exactly the same as the original), bringing the total system memory to 1GB. Upon installing the three brand new sticks, the system blue screened again, this time stating that win32k.sys attempted to write to read-only memory. I inserted my custom Windows PE disk in order to get a better look at the memory dump with BlueScreenView, but it refused to boot and produced another blue screen, but without an error message. I removed each new stick one-by-one, restarting each time. It continued to blue screen until I was left with only the original stick. I then tried inserting the new sticks in various different orders, but this only produced more blue screens. I reinserted all three sticks (along with the original) and ran memtest86 again, which reported errors all over the place. So, now I'm right back where I started. I don't think it could be the slots themselves, because I can plug the original stick into any slot and it works just fine. System setup reports each stick correctly and shows the total as 1GB, however. It just seems strange to me that all three brand new sticks of RAM could be dead on arrival. Is there something I missed? Or should I just go ahead and RMA them?

    Read the article

< Previous Page | 43 44 45 46 47 48 49 50 51 52 53 54  | Next Page >