Search Results

Search found 18761 results on 751 pages for 'lot'.

Page 127/751 | < Previous Page | 123 124 125 126 127 128 129 130 131 132 133 134  | Next Page >

  • How to use SSL for some pages and not for others?

    - by solid
    We are building an application where we want anything the following to be SSL: secure.ourapp.com - all pages must be SSL (can this be done with a Vhost in apache?) www.ourapp.com/signup - SSL www.ourapp.com/login - SSL www.ourapp.com/information - NOT SSL So my question is: can you put SSL on some pages and not on others? Or does it only work on a per-subdomain basis? thanks a lot for helping me out

    Read the article

  • how to redirect itunes to external hard drive

    - by oo
    i have all my music on my hard drive but its running out of space. i bought an external hard drive, copied all my music over i went into itunes advanced preferences and changed my music location but when i view the itunes music library XML file in notepad, i still see a lot of things pointing to my regular hard drive i didn't check "Keep music files organized" because i didn't want itunes trying to copy everything over from my regular hard drive (because i already did it) is there any other way to simply refresh the music list given a new root directory.

    Read the article

  • How to get a Unicode-supporting font for Windows 7 command-line?

    - by Tim
    I've pointed the command-line to the right codepage (chcp 65001), but there's a lot of Unicode characters that Consolas and Lucida Console can't show. Specifically, I want the printable IPA characters to show up. It's not important to fix multi-codepoint glyphs, although it would be nice. How can I get such a font and install it for the command-line? Below is an example of some characters that can't be rendered.

    Read the article

  • Removing the password from a PDF file

    - by Alister
    I have a couple of ebooks as PDFs with passwords, however my ebook reader (sony prs600) doesn't seem to support PDFs with passwords. What is the easiest of removing the password from a PDF (I know the password, which presumably helps a lot). It's a bit annoying buying a book and then only being able to read it in front of a computer.

    Read the article

  • NK2 files doesn't keep the email addresses in memory

    - by r0ca
    When I send an email to someone outside the firm, when I only type the first letters of its name (Contact), I get the auto-suggest of the "Already-sent" users. So now, since a few days, the emails are not kept in memory by Outlook (NK2 file). I see that that file is only 2kb and on my old machine, it's almost 200kb (So a lot more email addresses kept in memory) Should I just rebuilt the Outlook profile or the whole Windows Profile? A simple Outlook reinstall or to build a new PC?

    Read the article

  • Windows XP slow directory move

    - by maaartinus
    When I move a directory containing 900 MB in 4k files to another directory in the same filesystem, it takes nearly 1 minute and I hear the disk working. It's NTFS on Windows XP, the disk is quite fast (ST3100015 28AS) and works fine according to CrystalMark. I switched the antivirus off, and there's nothing else running (there's a lot of processes, but none doing any work). WTF is it doing instead of changing two directory entries?

    Read the article

  • What is the most efficient way to scan in thousands of pictures? [closed]

    - by leora
    My parents have a number of really large albums and the pictures are starting to face in a lot of cases so we thought it would be a good idea to scan in all the pictures and move them to online albums. The issue is that the task is daunting given that there are thousands of pictures. Are there any services or ideas on how to scan in albums of pictures that won't take up hundreds of man hours for me?

    Read the article

  • Exporting members of all DLs in an OU

    - by Bo Shubinsky
    I'm trying to export all the members of all the DLs within an OU (either to a single file that's categorized or individual files). I tried to use: csvde -f "C:\Documents and Settings\root\Desktop\AD Export\DL Export\DL.txt" -r "OU=DLs,OU=Personnel,DC=csi,DC=org" -l "cn,mail" but that only works for individual DLs and there are a lot to input each time. Any help on getting this done in the most efficient pattern would be helpful.

    Read the article

  • Security: Managing network shares remotely on Ubuntu?

    - by Industrial
    Hi everyone, I am about to setup a home network server running Ubuntu Server and I'm currently a bit worried about how to handle network shares and permissions in a good way. After working a bit lately with Netgears ReadyNAS's units, I have become really spoiled with how easy it was to set up network shares and giving a specific user different levels of network access to a specific share (forbidden access, read, read/write). How would I accomplish the same with my Ubuntu server through SSH? Thanks a lot

    Read the article

  • strange processes on Windows

    - by lego 69
    hello, can somebody explain, how can I delete unnecessary processes on windwos, for example I deleted icq, but I still have on Windows Task Manager some icq processes, thanks in advance (I have windows xp), every time when I start my windows I have a lot of unnecessary processes

    Read the article

  • How to tune system settings for mongoDB on Linux?

    - by jsh
    Trying to squeeze a lot out of one question here -- please bear with me. Although the MongoDB man pages make several useful recommendations about system settings like ulimit (http://docs.mongodb.org/manual/reference/ulimit/), and other production factors (http://docs.mongodb.org/manual/administration/production-notes/) they seem mysteriously silent on things like virtual memory and swap settings. The closest we get to a hint is that "...the operating system’s virtual memory subsystem manages MongoDB’s memory..." (http://docs.mongodb.org/manual/faq/fundamentals/#does-mongodb-require-a-lot-of-ram). Running the same job - high writes and high reads on about 10,000,000 records in a single collection -- on my 4-processor, 4GB RAM macbook and an 8-core ubuntu box with 64GB RAM I saw dramatically WORSE read performance on the linux box with factory settings, and could hear the disk constantly spinning, indicating high I/O and presumably swapping. Yes, other things were happening on the box, but there was plenty of free RAM, disk space, etc.; furthermore, I did not see evidence that Mongo was expanding to take advantage of all that free RAM as it is touted to do. Linux box default settings were as follows: vm.swappiness =60 vm.dirty_background_ratio = 10 vm.dirty_ratio = 20 vm.dirty_expire_centisecs =3000 vm.dirty_writeback_centisecs=500 I hazarded some guesses looking at docs and blogs for other types of databases (Oracle, MYSQL, etc.), experimented, and adjusted as below. vm.swappiness=10 vm.dirty_background_ratio=5 vm.dirty_ratio=5 vm.dirty_writeback_centisecs=250 vm.dirty_expire_centisecs=500 I saw some immediate apparent improvements in read time. However, when I ran my test jobs again, read performance continued to be painfully sluggish during heavy writes. Then, I REBUILT the collection from an available data source - and suddenly I can read at 1ms or less per record WHILE doing the write job! So the question is really two-fold: 1) What are appropriate VM settings for MongoDB on Linux? 2) (bonus) Does Mongo do some checking or optimization with the OS while data is being built? In other words, if I have built a large data set with suboptimal VM or I/O settings, does Mongo make assumptions during the memory-mapping process that will fail to take advantage of optimizations down the road? Obviously I don't fully grok memory mapping under the hood (I was hoping I wouldn't have to). Any help appreciated...thanks! -j

    Read the article

  • Shopping for a very compact and powerful workstation [closed]

    - by qdot
    I'm on the market for something small, size of an Mac Mini, could be heavier, that packs a lot of compute power - goal is to have a fast, known development machine at customer's site, but still easier to fly with than desktop-replacement laptops. If Mac Mini came with a quad-core i7, it would be my likely choice. Any other manufacturers to look at? Or perhaps a Mini can be upgraded beyond official specs?

    Read the article

  • Permit any user to mount any CIFS share

    - by A.K
    Essentially, I want the Ubuntu pre-10.10 behaviour back. The setuid method (see notes) does not work anymore. I search a lot on the Internet, but I haven't found a satisfying solution. I have read a solution that involves editing the sudoers file (ALL ALL=NOPASSWD:/sbin/mount.cifs). But then, the users would also be able to specify a directory as a mount-point they normally would not have access to, right? This is not what I want.

    Read the article

  • Capacity Planner errors

    - by Gabrie
    Hi Performing a discovery at customer site and although systems do get discovered, I get a lot (100000+) of these errors in the log: 02/01/2010 10:45:48:1:2200: Module = vcpDiscover Function = GetObfuscatedName Source = vcpDiscover Error = Method '~' of object '~' failed(-2147467259:0:5003251) 02/01/2010 10:45:48:1:2200: Module = vcpDiscover Function = GetObfuscatedName Source = vcpDiscover Error = Method '~' of object '~' failed(-2147467259:0:5003251) Any tips? Using latest version, just downloaded it. Gabrie

    Read the article

  • Oracle automatic patch download

    - by watain
    I need to update/patch the firmware of a lot of Sun/Oracle systems quite frequently. To do that I first need to download the latest firmware snapshot by hand from Oracle's saggy and unsophisticated "My Oracle Support" (list of all firmware snapshots can be found at http://www.oracle.com/technetwork/systems/patches/firmware/release-history-jsp-138416.html). Now I wondered whether there is a known way to automate the download of all the latest patches, maybe using a script. Kind regards

    Read the article

  • Bringing TechNet licensed server into production

    - by David Heggie
    I am currently trialling an install of Windows Server 2008R2 + SQL Server 2008 R2 using license keys from my company's TechNet subscription. If the trial is successful, I'd like to bring the server into production as is (there's a lot of config needed for it that I don't want to repeat) so my question really is is it possible to change the Windows and SQL server licenses from TechNet ones to "proper" volume licenses and legally use the server in production? Or do I have to reinstall everything with volume licences?

    Read the article

  • MySQL server installation problems (windows)

    - by waitinforatrain
    Hi guys, I'm trying to install some CMS software (Wiccle). I was using XAMPP's MySQL but was getting a lot of errors (the same configuration works on another machine) so thought I'd install MySQL Community Edition to see if the proplem was related to the MySQL server. When I install and run the MySQL Community Edition service, however, it only works with my XAMPP password, and contains the same tables as the XAMPP install. Is there a common local database file where the database and login info is stored? Any help appreciated

    Read the article

  • Messenger for voice calls

    - by Rogue
    I have been using Skype for a very long time, but lately I have been facing a lot of issues with the voice clarity. What messenger or VOIP clients that you know out of experience that are better than Skype for voice communication ?

    Read the article

  • Lots of dropped packages when tcpdumping on busy interface

    - by Frands Hansen
    My challenge I need to do tcpdumping of a lot of data - actually from 2 interfaces left in promiscuous mode that are able to see a lot of traffic. To sum it up Log all traffic in promiscuous mode from 2 interfaces Those interfaces are not assigned an IP address pcap files must be rotated per ~1G When 10 TB of files are stored, start truncating the oldest What I currently do Right now I use tcpdump like this: tcpdump -n -C 1000 -z /data/compress.sh -i any -w /data/livedump/capture.pcap $FILTER The $FILTER contains src/dst filters so that I can use -i any. The reason for this is, that I have two interfaces and I would like to run the dump in a single thread rather than two. compress.sh takes care of assigning tar to another CPU core, compress the data, give it a reasonable filename and move it to an archive location. I cannot specify two interfaces, thus I have chosen to use filters and dump from any interface. Right now, I do not do any housekeeping, but I plan on monitoring disk and when I have 100G left I will start wiping the oldest files - this should be fine. And now; my problem I see dropped packets. This is from a dump that has been running for a few hours and collected roughly 250 gigs of pcap files: 430083369 packets captured 430115470 packets received by filter 32057 packets dropped by kernel <-- This is my concern How can I avoid so many packets being dropped? These things I did already try or look at Changed the value of /proc/sys/net/core/rmem_max and /proc/sys/net/core/rmem_default which did indeed help - actually it took care of just around half of the dropped packets. I have also looked at gulp - the problem with gulp is, that it does not support multiple interfaces in one process and it gets angry if the interface does not have an IP address. Unfortunately, that is a deal breaker in my case. Next problem is, that when the traffic flows though a pipe, I cannot get the automatic rotation going. Getting one huge 10 TB file is not very efficient and I don't have a machine with 10TB+ RAM that I can run wireshark on, so that's out. Do you have any suggestions? Maybe even a better way of doing my traffic dump altogether.

    Read the article

  • Where is the statusbar in Firefox 4? How do I get it back?

    - by lovinglinux
    Since version 4.0b7pre, there is no more statusbar in Firefox, which has been replaced by the new add-ons bar. The main problem is that a lot of users like me are missing some valuable information that was displayed in the statusbar on previous versions. For instance, when hovering a link the url is now displayed in the address bar and you can't see the entire address. Other information like which servers are being contacted when you load a page are no longer displayed.

    Read the article

< Previous Page | 123 124 125 126 127 128 129 130 131 132 133 134  | Next Page >