Search Results

Search found 419 results on 17 pages for 'keith nicholas'.

Page 5/17 | < Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >

  • ML 350 additional SATA RAID controller (mirror only)

    - by Nicholas
    I have a Proliant ML350 G8 with two SAS raid arrays currently set up - thereby maxing the default P420i raid controller. I need to set up a large video dump space in addition to this existing set-up (for non backed up, non-critical, temporary storage). I had planned to just add a 2TB SATA disk and plug it into the motherboard. However I it occurred to me the motherboard might have built-in mirror RAID support? Therefore I could use two SATA disks and have some semblance of redundancy. Is this possible? Or would I need to get a cheap raid card? Any recommendations?

    Read the article

  • Is it possible to make a persistent USB serial port regardless ofthe existence of the USB device?

    - by Keith Nicholas
    USB serial ports, especially devices which emulate serial ports, don't quite behave the same as old serial ports, which causes a few problems with some software. old serial ports, always existed, they never come and go out of existence. With USB devices that emulate serial ports, they come and go depending on whether they are powered / reset etc. Is there a way under windows to make the USB serial port permanently exist regardless of the presence of the device? (not just come back as the same name as it was before).

    Read the article

  • How do I configure nVidia drivers on a Portable Ubuntu setup?

    - by Nicholas Flynt
    I've been pulling my hair out over this one for a couple of days now, google is no help. I've created a wonderful (until this issue) portable copy of Ubuntu linux that will boot on mostly anything by using a USB enclosure for my laptop's 80GB SATA drive. So far so good, it boots and runs on everything, and on non-nVidia card setups was even detecting the drivers, or letting me install the required drivers for hardware acceleration and compiz. Because you know, the wobble windows are the most awesome thing ever. Anyway, my desktop machine had an nVidia card, so I'm thinking, sure, I'll just install the nVidia drivers like before and everything will work happily. Not so-- now the desktop and any other nVidia cards work great, but it seems to have completely disabled any other graphics cards. When the kernel module detects that an nVidia card isn't present, it shoots up this nasty little dialog box giving me the option to boot into "low graphics" mode, which doesn't even allow me to use the correct screen resolution, much less see the installed graphics card and try to configure a driver for it. Is there any way to configure Ubuntu (with the dreaded nVidia kernel module) so that it can use nVidia's drivers when an nVidia card is present, and default to the normal (not low-graphics) setup in other cases, so that it has a fair chance of using what's actually present? I'm not afraid to much with config files, I just don't know the underlying system well enough to feel comfortable diving in without a push in the right direction. Thanks guys!

    Read the article

  • Using AT on Ubuntu to Background Downloads (w/ Queue)

    - by Nicholas Yost
    I am writing a PHP script, but I want to use the AT command in Ubuntu to fetch a remote file via WGET. I'm basically looking to background the process, so PHP can finish fairly quickly. I cannot find any questions on here about how to use both, but I basically want to do the following pseudo-code: <?php exec('at now -q queuename wget http://path.to/remote/file.ext'); ?> Additionally, I'd like to queue this between providers. I'd like to have each path.to have its own queue, so I only download one file from each provider at a time. Meaning: <?php exec('at now -q remote wget http://path.to/remote/file.ext /local/path'); exec('at now -q vendorone wget http://vendor.one/remote/file.ext /local/path'); exec('at now -q vendortwo wget http://vendor.two/remote/file.ext /local/path'); exec('at now -q vendorone wget http://vendor.one/remote/file.ext /local/path'); ?> This should download the files from path.to, vendor.one, vendor.two immediately, and when the first file is finished downloading from vendor.one, it starts the second file. Does that make sense? I can't find anything like this anywhere on the web, much less on SO/SF. If we can use the crontab to run a one-off wget command, thats fine too.

    Read the article

  • Drupal - Site Building, Menus

    - by Nicholas O'Neil
    Hi, I'm working with Drupal for the first time and trying to figure out how to change the Menu item names through the Administer Site Building configuration menus. I have logged into the Drupal admin site and navigated to Administer/ Site Building/ Menus, select the menu item and click Edit, change the Title from , 'About' to 'About Us' however when loading the site page it still says 'About' Please pardon my lack of understanding on drupal, i am somewhat familiar with joomla another CMS however I just need some pointers in the right direction as to how to edit the links and rename them within drupal. the links are not images as far as i can tell while viewing the source and properties on the links. Thank you!

    Read the article

  • Disk Response Time in Windows 7 Resource Monitor?

    - by Keith Nicholas
    In the resource monitor I'am looking at the disk response time. There are a lot of processes where the response time is thousands of milliseconds consistently, I'm pretty sure this is the source of my computer slowing down. I'm not sure what normal response times are though? I'm running win 7 64bit ultimate. This is running on a new computer, i5 with a terabyte drive, 4gigs of ram, etc, disk is still pretty much empty, so it should all be pretty snappy. And if it is going really slow, how do I track down whats causing it? I've turned off things like real time virus protection as experiments to see if there is something weird there, but makes no real difference (other than it doesn't contribute to the problem by accessing the disk)

    Read the article

  • Orphan IBM JVM process

    - by Nicholas Key
    Hi people, I have this issue about orphan IBM JVM process being created in the process tree: For example: C:\Program Files\IBM\WebSphere\AppServer\bin>wsadmin -lang jython -f "C:\Hello.py" Hello.py has the simple implementation: import time i = 0 while (1): i = i + 1 print "Hello World " + str(i) time.sleep(3.0) My machine has such JVM information: C:\Program Files\WebSphere\java\bin>java -verbose:sizes -version -Xmca32K RAM class segment increment -Xmco128K ROM class segment increment -Xmns0K initial new space size -Xmnx0K maximum new space size -Xms4M initial memory size -Xmos4M initial old space size -Xmox1624995K maximum old space size -Xmx1624995K memory maximum -Xmr16K remembered set size -Xlp4K large page size available large page sizes: 4K 4M -Xmso256K operating system thread stack size -Xiss2K java thread stack initial size -Xssi16K java thread stack increment -Xss256K java thread stack maximum size java version "1.6.0" Java(TM) SE Runtime Environment (build pwi3260sr6ifix-20091015_01(SR6+152211+155930+156106)) IBM J9 VM (build 2.4, JRE 1.6.0 IBM J9 2.4 Windows Server 2003 x86-32 jvmwi3260sr6-20091001_43491 (JIT enabled, AOT enabled) J9VM - 20091001_043491 JIT - r9_20090902_1330ifx1 GC - 20090817_AA) JCL - 20091006_01 While the program is running, I tried to kill it and subsequently I found an orphan IBM JVM process in the process tree. Is there a way to fix this issue? Why is there an orphan process in the first place? Is there something wrong with my code? I really don't believe that my simplistic code is wrongly implemented. Any suggestions?

    Read the article

  • How to disable mod_security2 rule (false positive) for one domain on centos 5

    - by nicholas.alipaz
    Hi I have mod_security enabled on a centos5 server and one of the rules is keeping a user from posting some text on a form. The text is legitimate but it has the words 'create' and an html <table> tag later in it so it is causing a false positive. The error I am receiving is below: [Sun Apr 25 20:36:53 2010] [error] [client 76.171.171.xxx] ModSecurity: Access denied with code 500 (phase 2). Pattern match "((alter|create|drop)[[:space:]]+(column|database|procedure|table)|delete[[:space:]]+from|update.+set.+=)" at ARGS:body. [file "/usr/local/apache/conf/modsec2.user.conf"] [line "352"] [id "300015"] [rev "1"] [msg "Generic SQL injection protection"] [severity "CRITICAL"] [hostname "www.mysite.com"] [uri "/node/181/edit"] [unique_id "@TaVDEWnlusAABQv9@oAAAAD"] and here is /usr/local/apache/conf/modsec2.user.conf (line 352) #Generic SQL sigs SecRule ARGS "((alter|create|drop)[[:space:]]+(column|database|procedure|table)|delete[[:space:]]+from|update.+set.+=)" "id:1,rev:1,severity:2,msg:'Generic SQL injection protection'" The questions I have are: What should I do to "whitelist" or allow this rule to get through? What file do I create and where? How should I alter this rule? Can I set it to only be allowed for the one domain, since it is the only one having the issue on this dedicated server or is there a better way to exclude table tags perhaps? Thanks guys

    Read the article

  • Better performance with memcached cluster or local memcaches?

    - by Nicholas Tolley Cottrell
    I have a small cluster of servers balancing a Java web app. Currently I have 3 memcached servers caching data and all web apps shares all 3 memcached instances. I often get strange slowdowns and timeouts to some of the memcacheds and I wondering if there is a good way of analyzing the performance. I am wondering whether my iptables rules (or some other system limitation) are blocking/slowing connections. I am considering reconfiguring the web apps so that they only query the memcached process on their own localhost.

    Read the article

  • Can I save an Apache environment variable value with SetEnv?

    - by Nicholas Tolley Cottrell
    I am running Apache 2.2 with Tomcat 6 and have several layers of URL rewriting going on in both Apache with RewriteRule and in Tomcat. I want to pass through the original REQUEST_URI that Apache sees so that I can log it properly for "page not found" errors etc. In httpd.conf I have a line: SetEnv ORIG_URL %{REQUEST_URI} and in the mod_jk.conf, I have: JkEnvVar ORIG_URL Which i thought should make the value available via request.getAttribute("ORIG_URL") in Servlets. However, all that I see is "%{REQUEST_URI}", so I assume that SetEnv doesn't interpret the %{...} syntax. What is the right way to get the URL the user requested in Tomcat?

    Read the article

  • What is the ideal way to set up multiple FTP enabled web accounts on Fedora?

    - by Nicholas Flynt
    I'm setting up a test server for use as a web development platform, and I'd like to mimic as closely as I can a typical shared hosting setup. That is, I'd like my server to have multple user FTP accounts, each of which links to a directory containing the webroot of the site, and I'd like apache to be able to easily see and manupulate these files. I'll admit: I'm not as familiar with Fedora as I'd like, I run Ubuntu on my home box and SElinux is giving me some grief. My initial plan was to have each user FTP into their home directory, and put the web directory there as well, but SElinux throws a hissy fit when apache tries to access anything outside of its web directory, so that plan was a no go. Would it be wise to continue this route, and perhaps mount web directories in user home folders so that FTP could still be used to access them, even though apache saw them in var/www like it expects? Would it make more sense to set up custom FTP accounts and use a single FTP user on the server box? What's the general course of action on something like this? I'm using vsftpd right now to host web directories, which is why I'm liking the home directory approach (it's simple and secure) but of course there's bound to be a better way to go about it. Thanks. (I'll leave other things, like restricted DB access and such, to another post. I'm interested right now with just getting FTP and apache to play nice in a multi-user environment.) PS: For the record, an issue I ran into when doing all of this was that if apache isn't running as the same user as the FTP account is saving as, there are permissions errors when FTP creates files, requiring the remote user to chmod the files to fix it. A logical fix would be to run apache in a special group, put all web users in this group, and have FTP access default to giving this group read/write access to everything like apache would expect, but I never could figure out how to accomplish this. Bonus points and cake if you know a solution.

    Read the article

  • rkhunter warns of inode change by no file modification date changes

    - by Nicholas Tolley Cottrell
    I have several systems running Centos 6 with rkhunter installed. I have a daily cron running rkhunter and reporting back via email. I very often get reports like: ---------------------- Start Rootkit Hunter Scan ---------------------- Warning: The file properties have changed: File: /sbin/fsck Current inode: 6029384 Stored inode: 6029326 Warning: The file properties have changed: File: /sbin/ip Current inode: 6029506 Stored inode: 6029343 Warning: The file properties have changed: File: /sbin/nologin Current inode: 6029443 Stored inode: 6029531 Warning: The file properties have changed: File: /bin/dmesg Current inode: 13369362 Stored inode: 13369366 From what I understand, rkhunter will usually report a changed hash and/or modification date on the scanned files to, so this leads me to think that there is no real change. My question: is there some other activity on the machine that could make the inode change (running ext4) or is this really yum making regular (~ once a week) changes to these files as part of normal security updates?

    Read the article

  • Inexpensive degaussers or HDD shredders?

    - by Nicholas Knight
    I do a lot of work for a small cash-strapped business that has a lot of active hard drives, most are consumer-grade SATA of about five years of age, and predictably they are dying at an increasing rate, and a lot of the time they can't even be detected, let alone complete a zero-out cycle. Right now those drives are just being stored, but that can't continue forever. We've got a couple bad LTO tapes it'd be nice to deal with, too. There are very real security and legal issues that make dropping them off with someone who claims they'll be properly destroyed a gamble. I've looked around at degaussers and HDD shredders, and the ones that don't look like they come from some guy in his basement all seem to be $3000+, which is hard to swallow right now. Is there anything out there in the $500-1500 range that you would recommend? (Speed isn't a big issue, if it takes several minutes or even hours per drive, that's completely OK, we've only got 10 or so thus far.)

    Read the article

  • How to keep Varnish cached populate after backend down for an extended period?

    - by Nicholas Tolley Cottrell
    We have Varnish 3.0.2 running on Amazon's Linux and it works great. We have a ttl of 48 hours for most content pages and much longer for images, PDFs etc. This weekend we've taken the backend down for some maintenance, so I upped the ttl to 5 days earlier in the week. I had assumed that anything in cache would continue to be served for up to 5 days, but much to our disappointment we checked varnishstat this morning and the cache was almost completely empty and varnish was serving "page not found" messages. I know that this is not what Varnish is designed to do, but why would it reset its cache when the backend is down? And how can I prevent it for next time?

    Read the article

  • How to write a ProxyPass rule to go from HTTPS to HTTP in IIRF

    - by Keith Nicholas
    I have a server which is running a web app that self serves HTTP. I'm wanting to use IIS6 (on the same server) to provide a HTTPS layer to this web app. From what I can tell doing a reverse proxy will allow me to do this. IIRF seems like the tool to do this job. There are no domain names involved.... its all ip numbers. So I think I want :- https:<ipnumber>:5001 to send all its requests to the same server but on a different port and use HTTP ( not exposed to the net ) http:<ipnumber>:5000 but not sure how to go about it with IIRF, I'm not entirely sure how to write the rules? I think I need to make a virtual web app on 5001 using HTTPS? then add a rules file.

    Read the article

  • Network Server Cabinets - How does the ventilation work?

    - by Nicholas
    Recently I've been looking at a range of network server cabinets with a mesh door at the rear. At the top of the cabinet are ceiling mounted fans with dust filters attached. These fans suck air out of the cabinet - presumably to remove the hot air that has risen to the top. My question is why do these fans have dust filters? What point do they serve? If the air flow is pulled in by these fans through the mesh door, then surely they are not there to protect the hardware inside from dust? Also, there are other minor grilled gaps around the enclosure, how does this kind of ventilation design work?

    Read the article

  • Whats a good secure Windows FTP server?

    - by Keith Nicholas
    Whats a good FTP server? I have been running FileZilla, which seems okish. But I've noticed that a lot of people try to hack ftp servers and FileZilla only has very basic controls to prevent people from hacking. (so far no ones actually managed to get in... so thats good!) I was wondering if there were better options out there? Especially interested in recommendations from people who know they get targeted by hackers.

    Read the article

  • Create PDF with internal hyperlinks

    - by Nicholas
    I have an Excel workbook with an index page containing a series of images, that are each hyperlinked to different worksheets. Can anybody suggest how I can publish this to PDF so that the hyperlinks work in the PDF? When I publish or use PDF Factory, the hyperlinks are not preserved (and bookmarks aren't created for that matter either). I would prefer to do this with inbuilt functionality or free tools, though any suggestions welcome. Also, it doesn't necessarily have to be with Excel. Publishing from powerpoint would be almost as good, and failing that, any easy solution which allows me to create a PDF containing clickable images that takes one to internal bookmarks might also be a workaround.

    Read the article

  • Stopping an unstoppable service

    - by Nicholas
    I have antivirus service (Kaspersky) that occasionally becomes unresponsive to the normal stop/stop gui interface provided by said vendor. I would like to find a way to kill the service for a restart without rebooting, however all attempts I have tried result in failure with an 'Access is Denied' error. These include: Services Control Panel (grayed out stop button) Task Manager Killing Process Explorer Killing command line net and sc stopping runas with domain admin using net stop Some details include: Machine: Windows Vista Service Type: 10 WIN32_OWN_PROCESS Service State: 4 Running (NOT_STOPPABLE, NOT_PAUSABLE, ACCEPTS_SHUTDOWN)

    Read the article

  • Configuring Windows Server Backup Destination Drive Sets

    - by Nicholas
    Is it possible to set up the standard backup system in SBS 2011 (or Server 2008 R2) to use an internal drive as a destination as well as external drives? Before you say yes, from my tests and from what I've read on the web, backups with internal drives included as a destination always seems to prefer the internal drive over connected external ones. (Regardless of what drive might be marked as 'active'). So no data ever gets written to a plugged in external drive. In my opinion external drives should always have priority over internal drives or including them is pointless.

    Read the article

  • Prevent Excel Chart Data Labels overlapping

    - by Nicholas
    I have an Excel dashboard with line charts containing data labels. Specifically, we are only using the data labels at the rightmost end of the lines, and the labels consist of the Series name and final value. By changing a dropdown, the dashboard is automatically updated to give 19 different dashboards. The problem is that we can't work out any way of preventing the labels overlapping. Everything else on the dashboard can be made to automatically update nicely, except for this. Can anybody think of a way to do this? E.g. plugin or macro.

    Read the article

  • How should oracle vbox look like in terms of Memory, CPU and Performance? [duplicate]

    - by Nicholas DiPiazza
    This question already has an answer here: Can you help me with my capacity planning? 2 answers I've got a need for a ton of VMs to simulate some realistic load testing scenarios. I've got a bunch of different host machines that differ in ram, cpu's, etc. What should my resource manager look like? Is there a standard way to know what the CPU, Memory and Disk Utilization should be given your CPUs + Memory available + Disks available? For example, I have a box: MemTotal: 50 Gb CPUs: 8 CPUs are pretty much 100% all day long. Memory is at about 60%. Swap not getting hit. Little bewildered by why the VMs, while doing the exact same test script, are showing different virtual memory consumption. Huh.

    Read the article

  • What Filesystem should be used for a 4TB drive for both Windows and OSX compatibility? [closed]

    - by Nicholas Yost
    Note: I am aware of similar questions. The one's I seen here are for Windows, OSX, and Linux (which I do not need). I also can use Mountain Lion, which the other questions did not mention. I was going to use NTFS, but OSX Mountain Lion can only read that filesystem and not write to it for some reason. I want to use something native between OSX and Windows, as I don't want to risk losing the data over filesystem incompatibilities. I have USB 3.0 and want something that will allow files greater than 4GB. I do not mind installing a small set of drivers on the Windows machine(s), but I would strongly prefer to leave the Mac machine untouched. Thanks!

    Read the article

< Previous Page | 1 2 3 4 5 6 7 8 9 10 11 12  | Next Page >