Search Results

Search found 33182 results on 1328 pages for 'linux port'.

Page 507/1328 | < Previous Page | 503 504 505 506 507 508 509 510 511 512 513 514  | Next Page >

  • Why doesn't my cron.d per minute job run?

    - by Travis Griggs
    I have thrown a bunch of darts trying to get a python script of mine to execute every minute. So I thought I'd simplify it to just do the "simplest thing that could could possibly work" once per minute (I'm running debian/testing). I created a single line file in /etc/cron.d/perminute: * * * * * /bin/touch /home/me/ding_dong It's owned by root, and executable (not sure if either of those matter). And then I did: sudo service cron reload And then sit back and start running ls -ltr again and again in my home directory (/home/me). But my ding_dong file never shows up. I know if I do a sudo /bin/touch /home/me/ding_dong, it shows up right away. Obviously missing something stupid here.

    Read the article

  • nginx load balance with IIS backend servers waiting Host header

    - by Elgreco08
    i have a ubuntu 10.04 with nginx /0.8.54 running as a load balance proxy named: www.local.com I have two IIS backend servers which responds on Host header request web1.local.com web2.local.com Problem: When i hit my nginx balancer on www.local.com my backend servers respond with the default server blank webpage (IIS default page) since they are waiting for a right host header (e.g. web1.local.com) my nginx.conf upstream backend { server web1.local.com:80; server web2.local.com:80; } server { listen 80; location / { proxy_pass http://backend; proxy_set_header X-Real-IP $remote_addr; proxy_set_header Host $proxy_host; } } any hint ?

    Read the article

  • I lost /dev/md2 on my server

    - by sten
    Hi, My 2 hard drives fried at the same moment apparently. My host company rebooted my server in rescue mode and I am trying to recover my data. They told me to mount /dev/sda2 to recover the data I need but, looking at a similar server that I have in pool, the data I'm looking for should be instead in /dev/md2. I can find /dev/md0 but not /dev/md2 (nor /dev/md1). I've looked on several places on the web and I could only find messages explaining how to create new partition. I just need to recover some data, not all of it and I'll be glad if anyone could help me to mount the /dev/md2 folder (or any other trick that would allow me to recover the data that was stored there). Thanks in advance, Sten

    Read the article

  • Question marks showing in ls of directory. IO errors too.

    - by jaymoo
    Has anyone seen this before? I've got a raid 5 mounted on my server and for whatever reason it started showing this: jason@box2:/mnt/raid1/cra$ ls -alh ls: cannot access e6eacc985fea729b2d5bc74078632738: Input/output error ls: cannot access 257ad35ee0b12a714530c30dccf9210f: Input/output error total 0 drwxr-xr-x 5 root root 123 2009-08-19 16:33 . drwxr-xr-x 3 root root 16 2009-08-14 17:15 .. ?????????? ? ? ? ? ? 257ad35ee0b12a714530c30dccf9210f drwxr-xr-x 3 root root 57 2009-08-19 16:58 9c89a78e93ae6738e01136db9153361b ?????????? ? ? ? ? ? e6eacc985fea729b2d5bc74078632738 The md5 strings are actual directory names and not part of the error. The question marks are odd, and any directory with a question mark throws an io error when you attempt to use/delete/etc it. I was unable to umount the drive due to "busy". Rebooting the server "fixed" it but it was throwing some raid errors on shutdown. I have configured two raid 5 arrays and both started doing this on random files. Both are using the following config: mkfs.xfs -l size=128m -d agcount=32 mount -t xfs -o noatime,logbufs=8 Nothing too fancy, but part of an optimized config for this box. We're not partitioning the drives and that was suggested as a possible issue. Could this be the culprit?

    Read the article

  • How can Yum tell me the version of depencencies?

    - by BanksySan
    I've installed a package with Yum (Varnish 3.0.2). However, a feature I need has been removed so I need to roll back. According to the documentation, the feature was removed in 3.0-Beta, so I wanted to revert to the latest version prior to that. (This is 2.1.5) This is installed now, but there were three RPMs to install for it (from http://users.linpro.no/ingvar/varnish/2.1.5/2.1.5-1/el5/x86_64/). After some faffing about with missing dependencies these got installed. However, the feature I need is still not there. I'm wondering if one of the dependencies is actually from a newer build. Do you know how I can see the versions of all the dependencies of a Yum installed package?

    Read the article

  • Client unable to reach Internet through OpenVPN

    - by Carroarmato0
    The clients can all connect through OpenVPN. OpenVPN serves the following pool: server 10.8.0.0 255.255.255.0 I've configured the server's iptable with the following rule: iptables -t nat -A POSTROUTING -s 10.8.0.0/24 -o eth0 -j MASQUERADE and echo 1 /proc/sys/net/ipv4/ip_forward This used to work back on the old vps I used. Now I've migrated to a vps which has ipv6 connectivity. Is it possible that Ipv6 has something to do with the fact that the clients can't reach the internet?

    Read the article

  • How can I use apt-get to resolve package dependencies when there are multiple versions in the repository?

    - by user1165144
    I've package a-package.deb which depends on b-package.deb in version 1.0. Everything works fine. But now a b-package in version 1.1 gets added to the repository. I'd suspect that apt-get installs the a-package and version 1.0 of the b-package. What really happens is, that a-package won't get installed: # apt-get install a-package Reading package lists... Done Building dependency tree Reading state information... Done Some packages could not be installed. This may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of Incoming. The following information may help to resolve the situation: The following packages have unmet dependencies: a-package : Depends: b-package (= 1.0) but 1.1 is to be installed E: Unable to correct problems, you have held broken packages. Is there a workaround to fix the behavior? Is there other software to use, that can handle the dependencies as defined?

    Read the article

  • how to know which display number for the variable DISPLAY to be exported when ssh to server

    - by insidepower
    When i ssh to server using -X, i always confuse about which display number i should export. It seems to me sometimes the display number has been used by something, so what i can do is only export DISPLAY=localhost:0 && xclock export DISPLAY=localhost:1 && xclock export DISPLAY=localhost:2 && xclock export DISPLAY=localhost:... until the clock appear. Then i will use that display number. Each time log in to the server, the display number which is able to tunnel the gui data correct would be different. I know many of such similar questions has been asked and answer. However I couldn't find answer to my question, anyone know about it? Thanks!

    Read the article

  • Vmware Workstation 10 connect remote server (Debian, Guest-Windows XP) Does not allow raw disk access nor shared folders

    - by Alex
    The setup: Ubuntu with local Vmware Workstation 10 (everything works locally) Connects(File- Connect to Server) Debian server with the same Vmware Workstation 10 (Windows XP Guest) Debian setup does not allow raw disk access nor shared folders (most options does not exist) No shared folder No physical disk option I use root user for this machine. Default install. I've tried to add shared folder from command line - it does not work. How to enable shared folders or raw disk access? I have created new Windows 8 64 bit template from scratch - I cannot use physical HDD either, and no SharedFolder option. I think this is something about security policy of remote server.

    Read the article

  • How do I get openssh to save my ssh identity permanently?

    - by Alliswell
    How do I change it to save my identity permanently? During my login session after I have entered: $ ssh-add Then identity is saved and I can connect via ssh without getting the dreaded: Enter passphrase for key Once I logout and login back-in I get the following when running: $ ssh-add -L The agent has no identities. $ ssh <hostname> Enter passphrase for key '/home/user/.ssh/id_rsa': Stackoverflow has the following solution, yet I do not understand the reference to in the config file: IdentityFile ~/.ssh/id_rsa_buhlServer Specifically I do not understand what I would put on the identity file. Would I put the above word for word? Or just add my file: IdentityFile ~/.ssh/id_rsa StackOverflow solution

    Read the article

  • I dont understand why [closed]

    - by gcc
    I dont understand why they vote down my question ,I couldont understand them -they do know nothing ,donot try to solve it-but they do vote down.If you do know nothing , dont fuck something or dont try to mess up

    Read the article

  • Where can I download mod_shaper for proftpd?

    - by lorus
    www.castaglia.org/proftpd/ says: mod_shaper is distributed with the proftpd source code but in the latest src from ftp://ftp1.at.proftpd.org/ProFTPD/distrib/source/proftpd-1.3.2e.tar.gz there is no mod_shaper.c inside ... so where can I get this mod?

    Read the article

  • Do control groups improve system performances?

    - by qdii
    According to this website, enabling cgroups in the kernel can boost performances by sharing resources in a better way. In particular, the conclusion states that:  Nevertheless, with a little trial and error, cgroups can help you improve the efficiency of your systems’ resource usage and avoid downtime due to overusage of a single service. Kernel seeds, however, recommend to deactivate them altogether. They say: Consider these [kernel] settings poison. They remain nothing but system slow-downs. They are all off by default [in the proposed kernel config file]. Who should I trust?

    Read the article

  • Backing Up User Data when data is not in use. Should I be concerned?

    - by jberryman
    This may be a dumb question. I would like to use duplicity to make backups to Amazon S3 of directories, each of which contains a different user's data. Each directory could be written to at any time. So I have two questions: Should I be concerned that a scheduled backup of a directory might occur in the middle of data being written to files in the directory, resulting in a corrupted backup? And if that is a valid concern, how would I go about temporarily delaying an operation while IO was happening, to try to minimize that effect. Thanks for the advice

    Read the article

  • How to securely control access to a backend key server?

    - by andy
    I need to securely encrypt data in my database so that if the database is dumped, hackers are unable to decrypt the data. I'm planning on creating a simple key server on a different machine, and allowing the DB server access to it (restricted by IP address on the key server to permit the DB server). The key server would contain the key required to encrypt/decrypt data. However, if a hacker were able to get a shell on the DB server, they could request the key from the key server and therefore decrypt the data in the database. How could I prevent this (assuming all firewalls are in place, DB is not connected directly to the internet, etc)? i.e. is there some method I could use that could secure a request from the DB server to the key server so that even if a hacker had a shell on the DB server they'd be unable to make those same requests? Signed requests from the DB server could make issuing these requests less trivial - I suppose that'd help increase the amount of time it'd take to compromise the key server, something a hacker probably wouldn't have much of. As far as I can see, if someone can get a shell on the DB server everything's lost anyway. This could be mitigated by using one key per data item in the DB so at least there's not a single "master" key, but multiple keys that the hacker would need to access. What would be a secure method of ensuring requests from the DB server to the key server were authentic and could be trusted?

    Read the article

  • SSH not working after installing SVN server on Debian

    - by sLIDe
    Today I had to install SVN server on my Debian server. I used this tutorial (only I didn't do anything to connect to SVN through file://, http:// or https://, only svn:// and svn+ssh://). After I installed SVN server and configured it after that tutorial, I tried to connect to it. I could connect to it using svn:// protocol, but when I tried to connect using svn+ssh:// protocol, my servers SSH stopped responding. Even after I stopped SVN server and restarted SSH server I can't connect to it.

    Read the article

  • mdadm raid1 fails to resync

    - by JuanD
    Hello, I'm trying to solve this problem I'm having with an mdadm raid1. I have an ubuntu 9.04 server running on a software 2-drive raid1 with mdadm. Yesterday, one of the drives failed, and so I replaced it with a brand new drive of the same size. I removed the faulty drive, copied the partition from the remaining good drive to the new drive and then added it to the raid. It re-synced and the system worked fine, until the drive that hadn't failed, was also labeled failed. Now I had the raid running solely on the new drive. So I purchased another drive and repeated the procedure above. So now I had 2 brand new drives and the raid was syncing. However, after a few minutes I checked /proc/mdstat and the raid was no longer syncing. mdadm --detail /dev/md1 shows: (sdb is the first new drive, and sdc is the second new drive) root@dola:/home/jjaramillo# mdadm --detail /dev/md1 /dev/md1: Version : 00.90 Creation Time : Sat Dec 20 00:42:05 2008 Raid Level : raid1 Array Size : 974711680 (929.56 GiB 998.10 GB) Used Dev Size : 974711680 (929.56 GiB 998.10 GB) Raid Devices : 2 Total Devices : 2 Preferred Minor : 1 Persistence : Superblock is persistent Update Time : Wed Jun 2 10:09:35 2010 State : clean, degraded Active Devices : 1 Working Devices : 2 Failed Devices : 0 Spare Devices : 1 UUID : bba497c6:5029ba0b:bfa4f887:c0dc8f3d Events : 0.5395594 Number Major Minor RaidDevice State 2 8 35 0 spare rebuilding /dev/sdc3 1 8 19 1 active sync /dev/sdb3 I've tried removing and re-adding the drive a few times, but the same thing happens. The raid fails to resync. I've looked at /var/log/messages, and found the following: Jun 2 07:57:36 dola kernel: [35708.917337] sd 5:0:0:0: [sdb] Unhandled sense code Jun 2 07:57:36 dola kernel: [35708.917339] sd 5:0:0:0: [sdb] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE Jun 2 07:57:36 dola kernel: [35708.917342] sd 5:0:0:0: [sdb] Sense Key : Medium Error [current] [descriptor] Jun 2 07:57:36 dola kernel: [35708.917346] Descriptor sense data with sense descriptors (in hex): Jun 2 07:57:36 dola kernel: [35708.917348] 72 03 11 04 00 00 00 0c 00 0a 80 00 00 00 00 00 Jun 2 07:57:36 dola kernel: [35708.917357] 00 43 9e 47 Jun 2 07:57:36 dola kernel: [35708.917360] sd 5:0:0:0: [sdb] Add. Sense: Unrecovered read error - auto reallocate failed So it looks like there's some kind of error on sdb (the first new drive). My question is, what would be the best approach to get the raid up and running again? I've thought about dd'ing the /dev/md1 to a blank hard drive, then re-doing the raid from scratch and loading the data back, but there could be an easier solution.. Any help would be appreciated.

    Read the article

  • How to reduce Fedora's disk size in VMware player.

    - by user428862
    I'm new to Fedora 14, vmware player. After getting Fedora up and running in VMware player. The disk size was 2.7 GB. After three hours of working with it, the disk size has bloated to 4.3 GB. I havent added software to account for the near doubling in size. How do I reduce the size back to 2.7GB range or lower. Im new to Fedora and superuser controls. Im removing more software than adding software. Is this a VMWARE problem or Fedora problem?

    Read the article

  • Setting thresholds via smartd.conf

    - by JPerkSter
    Hello! We currently use smartd to monitor SMART health on our disks. I would like to set the 'thresholds' that smartd uses to report. For example: 197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 9 198 Offline_Uncorrectable 0x0010 100 100 000 Old_age Offline - 9 I would like to be able to set a threshold for these two attributes to only report if they are above 10. I haven't been able to find a way to set this via smartd.conf, and I'm also not looking to actually shutdown the daemon. Has anyone ever tried this, or know how I might be able to accomplish this better then throwing a script into cron.hourly?

    Read the article

  • Le noyau Linux 3.3 disponible : intégration du code d'Android, améliorations réseaux, Btrfs et support d'une nouvelle architecture

    Le noyau Linux 3.3 disponible : intégration du code d'Android améliorations réseaux, Btrfs et support d'une nouvelle architecture Linus Torvalds vient d'annoncer la disponibilité de la version 3.3 du noyau Linux. Au menu des nouveautés, on notera essentiellement la réintégration des portions de code du noyau d'Android . Pour rappel, en 2009, les pilotes d'Android avaient été exclus du noyau parce qu'ils n'étaient pas suffisamment maintenus. L'intégration d'Android permettra aux développeurs d'utiliser le noyau Linux pour faire fonctionner un système Android, développer un pilote pour les deux et réduira les couts de maintenance des correctifs indépendants d'une...

    Read the article

  • How do I reset my PYTHONPATH?

    - by Underyx
    Somehow my PYTHONPATH environment variable got emptied, and now I can't import some downloaded modules. I've tried running the following command to reinstall all Python packages, but it didn't set the variable back to its original value: dpkg-query -W -f '${package}\n' | grep python | xargs -I % sudo apt-get install % --reinstall How do I reinstall the packages so that this gets sorted out? The (Vagrant) box is running Ubuntu 12.04.

    Read the article

< Previous Page | 503 504 505 506 507 508 509 510 511 512 513 514  | Next Page >