Search Results

Search found 696 results on 28 pages for 'matthew'.

Page 14/28 | < Previous Page | 10 11 12 13 14 15 16 17 18 19 20 21  | Next Page >

  • Delay init from starting a service for a period of time?

    - by Matthew
    I am trying to get a rudimentary NFS server up and running. Right now the server is configured as an NFS server due to a workaround for a vendor issue not supporting direct attached clustered storage, which we are trying to get them to resolve. The vendor software is Splunk. The splunk feature we are using requires files be located on shared storage (which for us is /mnt/nfs until they support a real clustered filesystem). Currently the server has a GFS2 filesystem mounted at bootup (it is the only server with the filesystem actively mounted so there should be no problems with locking). We went with GFS2 so switching over to a clustered filesystem is easy should the vendor begin supporting it. NFS is configured to mount that filesystem at /mnt/nfs, which the splunk installation than sees. Splunk is configured to find it's configuration files in /mnt/nfs. However, I am running into a problem where the splunk daemon starts before nfs is finished loading, and because it sees nothing at /mnt/nfs it starts creating files there, and then when the files disappear (nfs finishes mounting the share), splunk craps out. Splunk is set to run at runlevel 3, S90. NFS is set at runlevels 2-5, S60. Is there any way to delay the startup of the splunk process further?

    Read the article

  • Testing performance from around the world - how do I get a linux shell easily in multiple countries?

    - by Matthew O'Riordan
    We are building a socket based service where latency is paramount, and as such we have servers distributed into 7 data centres around the world. However, whilst we know we're bringing the servers closer to the clients, it's very difficult to know how effective this is, and importantly, what difference this makes compared to our competitors. As such, we want to run simple scripts that test latency and throughput for both our service and our competitors, which is easy enough using Amazon, however Amazon only have 7 data centres. We would like to know for example how we perform in locations all over the world such as South Africa, Australia, China, Peru etc. Does anyone know of any service where we could piggy back off their global infrastructure and run some scripts to test this performance? The obvious contenders are people like Monitis, but I don't think they would allow us to run custom scripts, only standard protocol monitors. Thanks for your help. Matt

    Read the article

  • Array on servers which receive several hundred GB of data a day

    - by Matthew
    This is hopefully a simple question. Right now we are deploying servers which will serve as data warehouses. I know with raid 5 the best practice is 6 disks per raid 5. However, our plan is to use RAID 10 (both for performance and safety). We have a total of 14 disks (16 actually, but two are being used for OS). Keeping in mind that performance is very much an issue, which is better - doing several raid 1's? Do one large raid 10? One large raid 10 had been our original plan, but I want to see if anyone has any opinions I haven't thought of. Please note: This system was designed for using Raid 1+0, so losing half of the raw storage capacity is not an issue. Sorry i hadn't mentioned that initially. The concern is more whether or not we want to use one large Raid 1+0 containing all 14 disks, or several smaller raid 1+0's and then stripe across them using LVM. I know the best practice for higher raid levels is to never use more than 6 disks in an array.

    Read the article

  • PC powers up but there is no display

    - by Matthew
    I built a computer with standard components I bought from Newegg about three years ago. It ran great for 2 years and has sat powered down for the last year. I tried to power it up today and the display was blank. It powers up, lights come on, drives start spinning but there is nothing on the display. I verified that the monitor and video adapter work. I also tried the video adapter in a different slot on the mother board with no luck. What's the next thing I should try? Is the mother board shot? Thanks.

    Read the article

  • International Calls For Free

    Are you sick of paying high fees for making international calls? There are much better ways out there and doing your research on the best product for you will save you a ton of money. First of course... [Author: Matthew Bailey - Computers and Internet - April 21, 2010]

    Read the article

  • How do I remove the ServerSignature added by mod_fcgid?

    - by matthew
    I'm running Mod_Security and I'm using the SecServerSignature to customize the Server header that Apache returns. This part works fine, however I'm also running mod_fcgid which appends "mod_fcgid/2.3.5" to the header. Is there any way I can turn this off? Setting ServerSignature off doesn't do anything. I was able to get it to go away by changing the ServerTokens but that removed the customization I had added.

    Read the article

  • A web app provider has asked for specific browser config

    - by Matthew
    They have asks to turn off caching on our browsers. I was aghast that they would ask such a thing. I said to them; To avoid caching it is best practice to use; <meta http-equiv="pragma" content="no-cache" /> <meta http-equiv="cache-control" content="no-cache" /> This should work across all browsers. Their reply was; We need to refresh javascript at runtime, this will not help us – any more ideas? I replied; Unsure what you mean by “refresh javascript at runtime”. If you are using ajax, browser caching can effect the XMLHttpRequest open method. Adding these meta tags to the source has fixed this for me in the past. Browser caching only caches resources, it should have no effect on site scripting. These meta tags will bypass browser caching. This is a reasonable request, isn't it?

    Read the article

  • Why won't my files push to my SFTP server?

    - by Matthew
    I'm having trouble pushing my branch to an SFTP server. I'm following the instructions here. When I push the branch, everything seems to complete successfully. I get the message "Created new branch.", and if I do "bzr push" again, it says "No new revisions to push." But when I ssh to the SFTP server to look at the directory I put my branch in, only the .bzr directory is there. None of my files are there. Does anyone have any idea why this might be?

    Read the article

  • Terminal Server 2003 Gaining Time when Windows 7 Client Connects

    - by Matthew
    A Windows 2003 Terminal Server keeps time perfectly until a Windows 7 Home client connects. Then it gains time at a rate of several seconds per minute. The client connects through a firewall with only the RDP port open. The client runs the same apps on the terminal server that XP clients run with no issues. Using the Microsoft Terminal Server Client application copied to the W7 computer from an XPsp3 computer gives the same results. Current workaround is to sync time every 5 minutes. Any better ideas?

    Read the article

  • How do I connect to my home's primary wired network through an extra (wireless) router?

    - by Matthew Patrick Cashatt
    Thanks for looking! I have set up a desktop PC in my workshop. The Cat 5 cable connects from this PC to a wireless router which is connected to my home network. The Internet connection is working just fine. However, the "wired" network this is on shows up as a different wired network than the one that the PCs inside my house are connected to. This is a problem because I would like to connect this workshop PC to various shared resources like printers, HD Homerun (cable tv card), shared drives, etc. When I go to "Network and Sharing" and attempt to find the network that the PCs inside my home are connected to, I don't see it. Any help is appreciated. Thanks!

    Read the article

  • Dns - wildcard vs. cname subdomains

    - by Matthew
    Alright I have to admit I'm confused with how DNS works. I've always just added things until they worked, and now it's time to learn how they work. So one confusing thing to me is that there's sort of two places I can have records. I have an account with rackspace cloud servers. And then there's the place I registered the domain. But both allow me to edit DNS records. Should I do everything at both places or is one better than the other or am I missing the point? Subdomains confuse me too. I'd like to be able to just have a wildcard subdomain (I've done this in the past.) I just don't like the idea of adding a cname record or A record every time I need a new subdomain. Then I read this and it says: The exact rules for when a wild card will match are specified in RFC 1034, but the rules are neither intuitive nor clearly specified. This has resulted in incompatible implementations and unexpected results when they are used.

    Read the article

  • running autobench (httperf)

    - by Matthew
    So I ran apt-get install httperf on my system and I can now run httperf. But how can I run 'autobench'? I downloaded the file and unarchived it and if I go in it and run autobench it says -bash command not found I think it's a perl script but if I run perl autobench, it says: root@example:/tmp/autobench-2.1.2# perl autobench Autobench configuration file not found - installing new copy in /root/.autobench.conf cp: cannot stat `/etc/autobench.conf': No such file or directory Installation complete - please rerun autobench Even if I run it again it says the same thing.

    Read the article

  • What causes Mac OS X Permission errors?

    - by Matthew Savage
    This is out of interest rather then looking for a fix to a problem. What actually causes permissions on Mac OS X Systems to become messed up? Its an easily fixed problem (i.e. there's a quick and easy fix via Disk Utility) but its something I'd encountered a few times doing support in a Mac-reseller store without actually understanding the causes. I'd guess that part of it is due to some applications not playing nicely, but what else might be the source of this issue?

    Read the article

  • How can I totally flatten a PDF in Mac OS on the command line?

    - by Matthew Leingang
    I use Mac OS X Snow Leopard. I have a PDF with form fields, annotations, and stamps on it. I would like to freeze (or "flatten") that PDF so that the form fields can't be changed and the annotations/stamps are no longer editable. Since I actually have many of these PDFs, I want to do this automatically on the command line. Some things I've tried/considered, with their degree of success: Open in Preview and Print to File. This creates a totally flat PDF without changing the file size. The only way to automate seems to be to write a kludgy UI-based AppleScript, though, which I've been trying to avoid. Open in Acrobat Pro and use a JavaScript function to flatten. Again, not sure how to automate this on the command line. Use pdftk with the flatten option. But this only flattens form fields, not stamps and other annotations. Use cupsfilter which can create PDF from many file formats. Like pdftk this flattened only the form fields. Use cups-pdf to hook into the Mac's printserver and save a PDF file instead of print. I used the macports version. The resulting file is flat but huge. I tried this on an 8MB file; the flattened PDF was 358MB! Perhaps this can be combined with a ghostscript call as in Ubuntu Tip:Howto reduce PDF file size from command line. Any other suggestions would be appreciated.

    Read the article

  • How to move an existing Win7 setup to a RAID-1 array?

    - by Matthew Scharley
    Currently I have a the following setup: Drive A/B: Identical HDD's Drive C: 2TB external drive with plenty of space currently. Hardware RAID controller Currently I have Drive A which is my boot drive (Win7), and Drive C which is a data drive, and Drive B which I only recently received and is blank. What is the best way of moving off of Drive A, setting up the hardware RAID and then migrating my data back onto the RAID array? I'm proficient with Linux, but I'm not sure if I can get away with simply using dd here. There is currently enough free space on Drive C to take 5-6 copies of a disk image of Drive A, so space isn't an issue.

    Read the article

  • At what point does the performance gap between GPU & CPU become so great that the CPU is holding back a system?

    - by Matthew Galloway
    I know that generally speaking for gaming performance the GPU is the primary factor which holds back performance, with everything else such as RAM/motherboard/PSU/CPU being secondary in importance to the graphics card. But at some point the other components ARE going to be significant in holding back the whole system! For instance nobody would be silly enough to play modern games with 512MB RAM and the very latest graphics cards (such as an HD7970) as I bet the performance increase over such a system with only 512MB but a mid range card would be non-existent! Thus it would be a "waste" for such a person to buy any high end graphics card without resolving first the system's other problems. The same point applies to other components, such as if it only had a Pentium II a current high end graphics card would be wasted on it! So my core question is how do you determine at what point for your system is spending on extra GPU power be completely "wasted"? (also, a slightly more nuanced question is trying work out at what point might the extra graphics power not be "wasted" but would be "sub optimal" value for money, when the expenditure should then be split around graphics card and other components. As obviously a gamer shouldn't always just spend on upgrading the graphics card! But needs to balance it out)

    Read the article

  • Cannot install windows. Compaq Presario CQ62

    - by Matthew
    I bought a used Compaq Presario CQ62 for cheap, and went to install windows on it. I formatted the partition and went to install when I got this error.... Windows cannot install required files. The file may be corrupt or missing. Make sure all files required for installation are available and restart the installation. Error code: 0x80070017 I have used this disk before with no problems, but internet searching suggested I burn one at 2x speed because that helps for some reason... I'm burning one now, but my question is, why would I get this error, OTHER than the disc being bad? I'm pretty certain this one isn't as I have used it before... (ok so the slowly burned cd (using imgburn) didn't work either so it's DEFINITELY not the disc) Thanks in advanced for any answers Also I took one stick of ram out because internet searching also suggested that, but it didn't make a difference. Also I ran memory and hard drive checks and they passed fine. Also I reset the motherboard options to default What could it be!? Help I'm completely stumped...

    Read the article

  • Can Grub be configured to remember the last OS you booted into?

    - by Matthew
    I typically have 2-3 OSes in my boot menu: Ubuntu 10.04 Windows 7 [Sometimes, a third option, such as Ubuntu 10.04 Netbook] If I am in an OS other than the default choice (Windows 7 in this case), and I reboot, Grub boots into Ubuntu 10.04. This shouldn't happen--when I reboot in Windows 7, I want to stay in Windows 7. Is there any way to configure Grub to remember my last choice, and boot into that? Alternatively, it would be even better if Grub only did this when I rebooted (not when I first turn the computer on). I realize this may not be possible, so I'd be happy with simple last-choice-remembering.

    Read the article

  • NFS on top of GFS2 - does it work?

    - by Matthew
    We're currently using a NoSQL derivative called Splunk to receive our data. The software supports something called "search head pooling" in which the job-dispatching engine is housed on several servers which share a common storage point. Originally our intention was to use a clustered filesystem like GFS2 because of low latency, stability, and ease of setup. We set up GFS2, and it's working with no issues. However when trying to run the software, it's trying to create lock files, and a bunch of other things that their support team can't quite explain. Ultimate feedback from them was that they only support NFS. Our network administration team heavily frowns on NFS (lack of stability, file lock issues, etc). So, I was thinking about the possibility of setting up NFS on each server in the cluster to act as a wedge layer between the GFS2 filesystem and the software. Basically configure each server to export the GFS2 filesystem's mountpoint via NFS, and then tell each server to connect to that NFS share. That way we aren't introducing any single-points-of-failure should a dedicated NFS server go down, but the vendor gets their "required" NFS share. I'm just brainstorming ways around, so please tear this apart :)

    Read the article

  • Installing Solaris 10 on sunT5220 - ZFS/UFS raid 10?

    - by Matthew
    I am in a bit of a time crunch, and need to get two T5220's built. We were very happy to see two boxes in our aged inventory which had 8 HDD's each, but didn't think to check if they were running hardware RAID or not. Turns out that they aren't. When we install, we are given the option to use UFS or ZFS, but when we select a place to install we're only given the option of installing on one single disk. Is it possible to create a software raid 10 across all of the disks and install the OS on that? Sorry if any lingo is wrong, I'm not really a Sun guy and our guru is out of town right now. Any help would be really appreciated! Note: Most of the guides I've found on google entail installing the OS on a single disk, and then creating a separate RAID 10 on other disks. We would actually like the OS to reside on the RAID 10. Hope that clarifies things.

    Read the article

  • How to move windows that open up offscreen?

    - by Matthew Scharley
    I have a dual monitor setup, and I recently played around with the positioning settings, and some of my single window applications do the favour of preserving where they were last closed, and opening in the same position later. Unfortuanately, that now places them out of the viewable area of my screens! Is there some way to force a particular window into the viewable area? If it matters at all, this is on Windows XP 32b.

    Read the article

  • WSUS moved how to reset the database and check the folders?

    - by Matthew Zielonka.co.uk
    I have a bit of a problem with a WSUS box, I backed up the WsusContent folder (about 180 gig) to a second machine, wiped the first machine then realized I don't have the database folder! I can not find any articles / guidance on this, I have re installed WSUS however it does not find the other files which where previsously downloaded (I guess the database being the reason). If I do a reset it will force it to download the ones it has again I guess and does not check what is still to be downloaded? Sigh... love / hate WSUS at the moment with a passion :) Thanks for any help.

    Read the article

< Previous Page | 10 11 12 13 14 15 16 17 18 19 20 21  | Next Page >