Search Results

Search found 14961 results on 599 pages for 'tab complete'.

Page 332/599 | < Previous Page | 328 329 330 331 332 333 334 335 336 337 338 339  | Next Page >

  • Songbird too damn slow, at least on my mac

    - by Cawas
    I've read on few places that Songbird is no good with more than some thousands of library items because it starts getting quite slow. Well, in my case (which is a clean install) I've imported 17k items (which I know is not that much) and instead of becoming just too slow it frequently gets to not responding for several minutes until getting back to its senses again. That's for whichever random operation such as deleting 1 item from library. I've also read on few more places that gives very little hope on fixing this issue, but I wonder... Is there any way to tweak it and make it work as fast as expected? Am I missing something or is this just a complete and utterly useless piece of software for libraries with more than 10 thousand thingies?

    Read the article

  • How can I automatically convert all source code files in a folder (recursively) to a single PDF with syntax highlighting?

    - by Bentley4
    I would like to convert source code of a few projects to one printable file to save on a usb and print out easily later. How can I do that? Edit First off I want to clarify that I only want to print the non-hidden files and directories(so no contents of .git e.g.). To get a list of all non-hidden files in non-hidden directories in the current directory you can run the find . -type f ! -regex ".*/\..*" ! -name ".*" command as seen as the answer in this thread. As suggested in that same thread I tried making a pdf file of the files by using the command find . -type f ! -regex ".*/\..*" ! -name ".*" ! -empty -print0 | xargs -0 a2ps -1 --delegate no -P pdf but unfortunately the resulting pdf file is a complete mess.

    Read the article

  • running autobench (httperf)

    - by Matthew
    So I ran apt-get install httperf on my system and I can now run httperf. But how can I run 'autobench'? I downloaded the file and unarchived it and if I go in it and run autobench it says -bash command not found I think it's a perl script but if I run perl autobench, it says: root@example:/tmp/autobench-2.1.2# perl autobench Autobench configuration file not found - installing new copy in /root/.autobench.conf cp: cannot stat `/etc/autobench.conf': No such file or directory Installation complete - please rerun autobench Even if I run it again it says the same thing.

    Read the article

  • APC serving old code intermittently running with Lighttpd and PHP Fast CGI

    - by APZ
    I recently started facing this problem that APC shows old code when we upload a html template file to fix/ change something on our websites. We run APC with Stat=0 and want to keep it that way because we seldomly make changes to templates. Every time we upload a template we make sure to flush APC cache and we execute this script(shown only some part of the script here) to clear the cache: apc_clear_cache(); apc_clear_cache('user'); apc_clear_cache('system'); apc_clear_cache(opcode); We use lightpd and PHP Fast CGI and fast cgi has "max-procs" = 2, "PHP_FCGI_CHILDREN" = "5", Even after flushing APC once upload is complete it serves the old template intermittently. Any help would be appreciated.

    Read the article

  • Verifying Time Machine backups

    - by jtimberman
    I'm preparing my system for a Snow Leopard upgrade, and I prepare for the worst case scenario: full reinstall and restore. I would like to verify that my Time Machine backups are valid, and will restore correctly. My Time Machine backups go to a Linux server running Netatalk, and the backups complete successfully. How do I do a test restore to an alternative location, or otherwise verify my data without overwriting any existing files? Do I need to save anything in particular externally to make sure I can access the backups if I have to reinstall from scratch?

    Read the article

  • Rails 2 and Ngnix: https pages can't load css or js (but will load graphics)

    - by Max Williams
    ADMISSION: i've posted this same question on stackoverflow, before realising it's probabaly better suited to superuser, but it kind of depends on the answer: If it turns out to be a problem in my nginx config, it's definitely superuser. If it turns out to be a problem in my Rails config (or code) then it's arguably stackoverflow. I'm adding some https pages to my rails site. In order to test it locally, i'm running my site under one mongrel_rails instance (on 3000) and nginx. I've managed to get my nginx config to the point where i can actually go to the https pages, and they load. Except, the javascript and css files all fail to load: looking in the Network tab in chrome web tools, i can see that it is trying to load them via an https url. Eg, one of the non-working file urls is https://cmw-local.co.uk/stylesheets/cmw-logged-out.css?1383759216 I have these set up (or at least think i do) in my nginx config to redirect to the http versions of the static files. This seems to be working for graphics, but not for css and js files. If i click on this in the Network tab, it takes me to the above url, which redirects to the http version. So, the redirect seems to be working in some sense, but not when they're loaded by an https page. Like i say, i thought i had this covered in the second try_files directive in my config below, but maybe not. Can anyone see what i'm doing wrong? thanks, Max Here's my nginx config - sorry it's a bit lengthy! I think the error is likely to be in the first (ssl) server block: server { listen 443 ssl; keepalive_timeout 70; ssl_certificate /home/max/work/charanga/elearn_container/elearn/config/nginx/certs/max-local-server.crt; ssl_certificate_key /home/max/work/charanga/elearn_container/elearn/config/nginx/certs/max-local-server.key; ssl_session_cache shared:SSL:10m; ssl_session_timeout 10m; ssl_protocols SSLv3 TLSv1; ssl_ciphers RC4:HIGH:!aNULL:!MD5; ssl_prefer_server_ciphers on; server_name elearning.dev cmw-dev.co.uk cmw-dev.com cmw-nginx.co.uk cmw-local.co.uk; root /home/max/work/charanga/elearn_container/elearn; # ensure that we serve css, js, other statics when requested # as SSL, but if the files don't exist (i.e. any non /basket controller) # then redirect to the non-https version location / { try_files $uri @non-ssl-redirect; } # securely serve everything under /basket (/basket/checkout etc) # we need general too, because of the email/username checking location ~ ^/(basket|general|cmw/account/check_username_availability) { # make sure cached copies are revalidated once they're stale add_header Cache-Control "public, must-revalidate, proxy-revalidate"; # this serves Rails static files that exist without running # other rewrite tests try_files $uri @rails-ssl; expires 1h; } location @non-ssl-redirect { return 301 http://$host$request_uri; } location @rails-ssl { proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header Host $http_host; proxy_redirect off; proxy_read_timeout 180; proxy_next_upstream off; proxy_pass http://127.0.0.1:3000; expires 0d; } } #upstream elrs { # server 127.0.0.1:3000; #} server { listen 80; server_name elearning.dev cmw-dev.co.uk cmw-dev.com cmw-nginx.co.uk cmw-local.co.uk; root /home/max/work/charanga/elearn_container/elearn; access_log /home/max/work/charanga/elearn_container/elearn/log/access.log; error_log /home/max/work/charanga/elearn_container/elearn/log/error.log debug; client_max_body_size 50M; index index.html index.htm; # gzip html, css & javascript, but don't gzip javascript for pre-SP2 MSIE6 (i.e. those *without* SV1 in their user-agent string) gzip on; gzip_http_version 1.1; gzip_vary on; gzip_comp_level 6; gzip_proxied any; gzip_types text/plain text/css application/json application/x-javascript text/xml application/xml application/xml+rss text/javascript; #text/html # make sure gzip does not lose large gzipped js or css files # see http://blog.leetsoft.com/2007/7/25/nginx-gzip-ssl gzip_buffers 16 8k; # Disable gzip for certain browsers. #gzip_disable "MSIE [1-6].(?!.*SV1)"; gzip_disable "MSIE [1-6]"; # blank gif like it's 1995 location = /images/blank.gif { empty_gif; } # don't serve files beginning with dots location ~ /\. { access_log off; log_not_found off; deny all; } # we don't care if these are missing location = /robots.txt { log_not_found off; } location = /favicon.ico { log_not_found off; } location ~ affiliate.xml { log_not_found off; } location ~ copyright.xml { log_not_found off; } # convert urls with multiple slashes to a single / if ($request ~ /+ ) { rewrite ^(/)+(.*) /$2 break; } # X-Accel-Redirect # Don't tie up mongrels with serving the lesson zips or exes, let Nginx do it instead location /zips { internal; root /var/www/apps/e_learning_resource/shared/assets; } location /tmp { internal; root /; } location /mnt{ root /; } # resource library thumbnails should be served as usual location ~ ^/resource_library/.*/*thumbnail.jpg$ { if (!-f $request_filename) { rewrite ^(.*)$ /images/no-thumb.png break; } expires 1m; } # don't make Rails generate the dynamic routes to the dcr and swf, we'll do it here location ~ "lesson viewer.dcr" { rewrite ^(.*)$ "/assets/players/lesson viewer.dcr" break; } # we need this rule so we don't serve the older lessonviewer when the rule below is matched location = /assets/players/virgin_lesson_viewer/_cha5513/lessonViewer.swf { rewrite ^(.*)$ /assets/players/virgin_lesson_viewer/_cha5513/lessonViewer.swf break; } location ~ v6lessonViewer.swf { rewrite ^(.*)$ /assets/players/v6lessonViewer.swf break; } location ~ lessonViewer.swf { rewrite ^(.*)$ /assets/players/lessonViewer.swf break; } location ~ lgn111.dat { empty_gif; } # try to get autocomplete school names from memcache first, then # fallback to rails when we can't location /schools/autocomplete { set $memcached_key $uri?q=$arg_q; memcached_pass 127.0.0.1:11211; default_type text/html; error_page 404 =200 @rails; # 404 not really! Hand off to rails } location / { # make sure cached copies are revalidated once they're stale add_header Cache-Control "public, must-revalidate, proxy-revalidate"; # this serves Rails static files that exist without running other rewrite tests try_files $uri @rails; expires 1h; } location @rails { proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header Host $http_host; proxy_redirect off; proxy_read_timeout 180; proxy_next_upstream off; proxy_pass http://127.0.0.1:3000; expires 0d; } }

    Read the article

  • Export 1 year of CVS to another repo?

    - by John Dibling
    We have a CVS repo with many years of history. It has become huge and unwieldly, so we would like to split this singe repo in to two repos: The main repo would have 1 year's worth of history, up to and including present day. This is where all dev work would take place. An archive repo would have the complete history, up to the point where the main repo would take over. This would be read-only, and only used to look at historical changes. Given that we are starting with one huge, monolithic CVS repo, is it possible to split it up in this way? How can this be accomplished?

    Read the article

  • Windows 8 Secureboot: How do you bootup from another device such as cd/dvd drive or usb flash etc.?

    - by Victor T.
    On PC's and laptops running an older OS, this is just a simple matter of going into the BIOS and setting the boot sequence and putting the boot cd/dvd in the drive. In many cases you can even just hit one of the F* function keys to bring up the boot sequence menu on-the-fly during POST. The main problem I'm running into is that other devices besides the primary HD is disabled when SecureBoot is enabled. So far the only way I've gotten it to work is to disable SecureBoot and enable something called legacy mode. Needless to say this make it difficult to boot things like OS recovery tools, PartitionHD backups, Linux LiveCD's and a bunch of others. Is there another procedure for doing this since it seems to mess up the Windows 8 install? By mess up I mean that after re-enabling SecureBoot and starting Windows 8 normally, the OS goes into a 'preping repair phase' for some reason that takes forever to complete before getting into a usable desktop.

    Read the article

  • Do I need to conver the older Access Database, and, if so, how?

    - by octopusgrabbus
    I have an Access 2003 database. When I click on a pivot table, I get this message MS Access There isn't enough memory to complete the Automation object operation on the worksheet object. There is a lot of discussion concerning this message. Here is one link. http://community.spiceworks.com/topic/113228-access-2003-file-pivot-table-issue-when-opening-in-access-2010 But this particular link's explanation doesn't really go into fixing the problem in general, like fixing the pivot tables and getting things all nicely back together in the original Access database. That's why I am also interested in converting the database to 2010 format if that is possible. Are there instructions -- I cannot currently find them and would very much appreciate a link -- on dealing with this problem in a nice stepwise fashion?

    Read the article

  • How to make Windows display application crashes in the console?

    - by Name is carl
    I'm developping a C application on Windows 7 64 bits, and each time it crashes I get the following dialog: I would prefer something more "old school" with the complete error dumped in the console. Something like > HelloWorld The application has crashed in somedll.dll for some Horrible C reason. I have an error code 0x4445506 And a kinda stacktrace And walls of hexa 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 0x4445506 Is there a way to get that? Should I run my application with a command line debugger ? Note: should I post this on Stackoverflow instead?

    Read the article

  • How can I rewrite a URL and pass on the original URL as a parameter?

    - by Bobby Jack
    I'm building a site that needs to include a 'check' procedure, to do several initiation tasks for a user's session. Examples include checking whether they're accepting cookies, determining if their IP address grants them specific privileges, etc. Once the check is complete, I need to redirect the user back to the page they originally requested. The plan is to use RewriteCond and map all URLs to an 'initiator' if the user doesn't have a specific cookie set. Let's say I want to rewrite all URLs (ultimately, with some conditions, of course) to: /foo?original_url=... Where the ... is the original URL requested, URL-encoded. The closest I've got is this: RewriteRule ^(.*)$ http://localhost/php/cookie.php$1 [R=301] I can then inspect the original URL, captured in the backreference, via PATH_INFO. However, this is pretty messy - I would much prefer to pass that value as a URL parameter

    Read the article

  • top process state column under FreeBSD

    - by Eric DANNIELOU
    When running top interactively, I can see various word in the state column : nanslp, biord, select, uwait, lockf, pause, kqread, piperd, sbwait ... Some like nanslp or kqread are self explanatory, others are not. Tried man pages : STATE is the current state (one of "START", "RUN" (shown as "CPUn" on SMP systems), "SLEEP", "STOP", "ZOMB", "WAIT", "LOCK" or the event on which the process waits), C is the processor number on which the process is executing (visible only on SMP systems) Tried search engines : stack overflow mailing lists archives Where may I get a complete list of possible process state under FreeBSD 9, and their meanings?

    Read the article

  • Windows 8 start screen will only find apps matching full words

    - by Ian Oakes
    Since doing a fresh install of Windows 8 x64 RTM I'm having trouble starting apps from the start screen. In previous version I could just start typing the apps name and I would be presented with a list of matches that decreased in number as I continued typing. In the RTM version I have to type complete words to find a match. i.e. if I want to run powershell I have to type in powershell before I get any hits. Like wise if I want to start Visual Studio I have to type visual before I get a match. If I continue and start typing Studio the match goes away and doesn't return until I finish typing studio. This problem does not affect Setting or Files which show all possible matches. Any ideas what is wrong?

    Read the article

  • Use a media player in Linux just to play files from an iPod device (no sync, no manage, just play)?

    - by Somebody still uses you MS-DOS
    I have an ipod classic 160gb, that I sync with my machine at home. I use Linux at work, and want to just plug my ipod and just listen to the tracks, with all the playlists and such. I don't want to sync nothing, I just want to listen to the tracks as if I was using the ipod itself. Why? Because this way I can use the usb port. So, I don't want to manage my ipod in Linux, I just want to listen to the tracks on it in Linux, like it was a local library but it's instead in my ipod. (I've tried gtkpod, it works to show my files, but I can't play, shuffle, etc. It would be interesting to have a complete audio software to handle everything like it was a local library)

    Read the article

  • Shuttle FB51 mobo does not boot with external USB drive attached [closed]

    - by user127236
    I am repurposing an old Alienware desktop as a home media server. The PC is based on the Shuttle FB51 motherboard. The BIOS is a Phoenix Version 6.00 PG, release date 12/16/2002. I have loaded Ubuntu 12.04 LTS on the internal hard drive. I am using a Western Digital WD Elements 1.5 TB USB 2.0 Desktop External Hard Drive for media storage. When the external drive is plugged in and the PC is powered on, it freezes very early in the BIOS self-test, even before it begins the memory test. If I unplug the drive, the self-test proceeds without further problems. I can plug the USB drive back in when the self-test is complete, and Ubuntu will boot and find the external drive normally. I've tried several changes to the BIOS setup without finding a cure for the boot issue. Any assistance gratefully accepted. JGB

    Read the article

  • Entire filesystem restore from rdiff-backup snapshot

    - by atmosx
    I'm trying to make a complete system restore from an rdiff-backup. The cli for backing was: rdiff-backup --exclude-special-files --exclude /tmp --exclude /mnt --exclude /proc --exclude /sys / /mnt/backup/ebox/ I created a new partition mounted the partition at /mnt/gentoo and did: rdiff-backup -r /mnt/vol2 /mnt/gentoo However when I try to chroot to this system (following gentoo's manual, which means mounting /dev/ and /proc) I get the following error: chroot: failed to run command/bin/bash': No such file or directory` All this takes place on a Parallels (virtual machine) Debian installation. Any ideas on how to proceed in order to fully restore the system? Best Regards ps. /mnt/gentoo/bin/bash works fine if I execute it. All files and permissions are in place rdiff-backup seems to work just fine. However the system cannot neither boot (exits with kernel panic - cannot find init) or be chrooted.

    Read the article

  • USB port causing device problems after new motherboard upgrade

    - by enbuyukfener
    I'm looking into a PC with an issue with one of the front USB ports (perhaps both). It was working before a motherboard upgrade but since then, a USB drive was inserted, overheated and does not appear to be working. Then a phone (with USB charging) was plugged in, and the phone OS suspended. Removing and re-inserting the battery led to the phone (and hence battery) working, however the battery no longer charges (including with a wall charger). It seems too much to be a coincidence and am wondering what the issue may be? Ideas so far are short circuiting, or over-current to the USB ports. Note: Did not occur to me, so details are not 100% accurate or complete. Feel free to ask for missing info that I may have forgotten though.

    Read the article

  • Windows 8 account and MSN

    - by win8user
    I've installed windows 8 and during setup it wanted my email. I provided an alternative email (let's call it email_1) for the installation to complete. Then I've installed windows live (msn only), for which I use another email (email_2). Now, every time I start windows and open msn, I'm signed in with email_1. I selected "forget me" for email_1 address in the msn login window, but I'm somehow always signed in with email_1. Is there any way to completely remove email_1 from the msn?

    Read the article

  • Date based sum in Excel / Google Docs spreadsheets

    - by alumb
    I have a bunch of rows with a date and a dollar amount (expenses). I want to produce a list of the days of the month and what the balance of the expenses is. So, for example the 5th entry in the list would be 8/5/2008 and the sum of all the expenses that occurred on or before 8/5/2008. Approximately this is =sumif(D4:D30-A5,">0",E4:E30) but of course that doesn't work (where the source data is dates in D4:D30 and the expenses are in E4:E30). Notes source data can't be sorted for various reasons. must work in google spreadsheets, which is a fairly complete subset of excel's functions.

    Read the article

  • How to run a process and completely detach it of its parent shell

    - by Bicou
    I'm running a program on a linux server that will take days to complete. I'm launching it from my workstation from an SSH terminal, as this program is command-line only. I want to be able to do all of these : launch that program, redirect standard outputs to files, exit my SSH session without making this terminate the process. I thought about $ ./MyProg.csh -params -foo -bar </dev/null 1>~/out.log 2>~/err.log & However, the process is terminated the moment I close my SSH session. My workstation is running Windows XP, and I cannot guarantee its uptime over several days, which is required for the processing of my data on the Linux server. As you may have noted, my program requires to be launched from CSH. Is it possible to do this ? Thanks.

    Read the article

  • Routing between two subnets. (Need Solution)

    - by rehanplus
    Need help according to scenario given: Client end PCs: 400 + Network : Server 1 (Linux) : 192.168.2.0/24 (For Application, Internet not working) GW: 192.168.2.1 Clients: 192.168.2.1 - 254 Server 2 (Linux) : 192.168.3.0/24 (For Internet users) GW: 192.168.3.1 Clients: 192.168.3.2 - 254 Server 2 is connected to DSL Broadband. Server 1 and Server 2 both are on same physical network i.e. Same switches. Current issue: i have to deploy a file and print server but this server will be accessed by both (192.168.2.x and 192.168.3.x) one same workgroup. as both subnets are on same switched network. Limitations: Currently there is no hardware routers and firewall. Need to complete this task with Linux / Windows / AD. Tested / Worked so far: Configured one PC with two NIC's With the IPS: NIC 1 : 192.168.3.2 GW: 192.168.3.1 Subnet: 255.255.255.0 NIC 2 : 192.168.2.2 GW: Empty Subnet: 255.255.255.0 Kindly provide any solution what should i do to get sharing enable on both Subnets. Thank you All

    Read the article

  • rsync to ONLY keep files in destination that have been removed from source

    - by David Corley
    We use rsync to copy filesystem contents from one machine to another as a backup. We first run MACHINE-X-MACHINE-Y rsync for a straight backup with the --delete and --delete-excluded switches We also run an internal Rsync between the MACHINE-Y destination, and another folder on MACHINE-Y with either of the delete flags. This maintains a non-destructive copy in the event someone inadvertently deletes a file on MACHINE-X. However, it also has the overhead of being a complete copy of what has already been synchronized. Ideally I want to be able to run the non-destructive rsync in such a way that the destination ONLY receives the deleted files and so avoids unnecessary duplication . Is there any way to do this?

    Read the article

  • How can I search for Gmail conversations that have more than one e-mail? [closed]

    - by Matt Refghi
    I previously used a service that sent me Slashdot articles as individual e-mails. All I had to do was give them the RSS url, and they handled the rest. On the Gmail side, I simply made a filter, labeling these e-mails as "News". Once I stopped using the service, I realized that I had more than 20,000 such e-mails stockpiled. While I could delete all "News" e-mails, there are some that I've forwarded to friends, sometimes receiving multiple replies. These conversations are valuable - I would like to keep them. As far as I can tell, there's no way to search for conversations that have more than one e-mail in the thread. Up until now, I've been deleting them page by page, carefully de-selecting the conversations that have more than one e-mail; however, it will take far too much time to complete. Is there an easier way that I'm missing? Thanks.

    Read the article

  • Windows 7: enabling navigation of subfolders in pinned Start Menu folders

    - by AspNyc
    I'm just about to move from Windows XP to Windows 7, and I'm struggling with some of the interface changes. In XP, I was able to throw a folder intoC:\Documents and Settings\username\Start Menuand have it appear on the Start Menu, complete with the ability to navigate through subfolders. I've figured out how to pin a folder onto the Start Menu in Windows 7, which required a registry hack. However, I am unable to view the subfolders of the pinned folder without opening a new Windows Explorer window. Is there any way to replicate the old XP behavior I'm used to? I'd like to be only a single click away from these handful of application links and folders, since I use them all the time throughout the day.

    Read the article

  • Can I use the snipping tool to take a screenshot of the windows 8 start screen or modern apps?

    - by Journeyman Geek
    One of the tools I've found invaluable in answering questions on SU is the snipping tool. I may on occasion need to take screenshots of part of the start screen or 'modern' apps. I may not want to take a complete screenshot, and while I can use PrtSc and switch back into desktop to paste it, this is clunky if I need to document a multi-step process. Can I use the snipping tool on modern apps or the start screen? If not is there a configurable way to save a series of screenshots to a fixed folder, say when I press a combination of keys, so I can work, screenshot, then crop and annotate the folder of images?

    Read the article

< Previous Page | 328 329 330 331 332 333 334 335 336 337 338 339  | Next Page >