Search Results

Search found 12017 results on 481 pages for 'no root'.

Page 308/481 | < Previous Page | 304 305 306 307 308 309 310 311 312 313 314 315  | Next Page >

  • Remove folder structure from archive, ignore folder while archiving and fix error

    - by Michael
    I am trying to make a script to backup each of my plesk hosts to individual files, I am having two problems: I would like to remove the folder structure from archive, the tar is 3 folders deep I am getting this error: tar: Removing leading `/' from member names I need my archive to ignore folders named "catch" because I don't need them in my archive. The code: FILES=/var/www/vhosts/* FNAME="" for f in $FILES do FNAME=`basename $f` tar cfv "/root/backup/ftp/$FNAME.tar" $f done Sample output: tar: Removing leading `/' from member names /var/www/vhosts/mydomain.com/ /var/www/vhosts/mydomain.com/conf /var/www/vhosts/mydomain.com/etc/ /var/www/vhosts/mydomain.com/etc/group /var/www/vhosts/mydomain.com/etc/termcap /var/www/vhosts/mydomain.com/etc/passwd /var/www/vhosts/mydomain.com/usr/

    Read the article

  • Mysql showing some high spikes

    - by user111196
    We one mysql server suddenly the access to it becomes slow of out sudden. So I read some place they said maybe due to var size is it? I am not too sure any idea how to check the root cause of it. The cpu is like nearly 150%. Any indication on it. I have tried this so far. du -sh * 4.0K account 67M cache 4.0K cvs 16K db 8.0K empty 4.0K games 4.0K gdm 148G lib 4.0K local 16K lock 624M log 0 mail 4.0K nis 4.0K opt 4.0K preserve 400K run 298M spool 4.0K tmp 359M www 12K yp

    Read the article

  • Running suspicious X programs in GNU/Linux

    - by Vi
    What the most harmful thing can malware program started as separate limited user account do if it has access to the X server? Network and filesystem things are already considered by chroot and netfilter. It obviously can lock the screen and I will need to switch to other vt and kill it manually. Can it for example disrupt other GUI programs on the same X server (access to root terminal in nearby window)? I know that it is safer to run it in separate X server, for example, in Xtightvnc or even some virtual machine, but how dangerous is to just run it like other programs?

    Read the article

  • What's the best practice for taking MySQL dump, encrypting it and then pushing to s3?

    - by HalogenCreative
    This current project requires that the DB be dumped, encrypted and pushed to s3. I'm wondering what might be some "best practices" for such a task. As of now I'm using a pretty straight ahead method but would like to have some better ideas where security is concerned. Here is the start of my script: mysqldump -u root --password="lepass" --all-databases --single-transaction > db.backup.sql tar -c db.backup.sql | openssl des3 -salt --passphrase foopass > db.backup.tarfile s3put backup/db.backup.tarfile db.backup.tarfile # Let's pull it down again and untar it for kicks s3get surgeryflow-backup/db/db.backup.tarfile db.backup.tarfile cat db.backup.tarfile | openssl des3 -d -salt --passphrase foopass |tar -xvj Obviously the problem is that this script everything an attacker would need to raise hell. Any thoughts, critiques and suggestions for this task will be appreciated.

    Read the article

  • ModRewrite Domain

    - by Mike Knoop
    I've done a little research into ModRewrite rules and conditions but have not been able to find a satisfactory set of rules/conds which achieve the effect I'm looking for. Essentially, I have a directory on domain A (http://www.domaina.com/dir/) which I would like to redirect to a different directory on domain B (http://www.domainb.com/diff_dir/). Note that I only want to apply the rewrite rule if the user is attempting to access /dir/ on domaina. If they are accessing a different directory or root I do not want to rewrite the URL. Thank you!

    Read the article

  • Delete protected files in ntfs?

    - by Balchev
    I want to delete an old Windows directory from my system drive (C), but I am unable to due NTFS permissions. I tried from Win 7 and Win 2003, but can't do it. I tried safe mode as well with same result. Is there any way to work around this (other then formating the drive)? Perhaps changing the owner or something? It errors at files like "oldwin/bfsvc.exe". Is there some "superuser" in windows similar to linux root account? Thanks

    Read the article

  • Vmware Workstation 10 connect remote server (Debian, Guest-Windows XP) Does not allow raw disk access nor shared folders

    - by Alex
    The setup: Ubuntu with local Vmware Workstation 10 (everything works locally) Connects(File- Connect to Server) Debian server with the same Vmware Workstation 10 (Windows XP Guest) Debian setup does not allow raw disk access nor shared folders (most options does not exist) No shared folder No physical disk option I use root user for this machine. Default install. I've tried to add shared folder from command line - it does not work. How to enable shared folders or raw disk access? I have created new Windows 8 64 bit template from scratch - I cannot use physical HDD either, and no SharedFolder option. I think this is something about security policy of remote server.

    Read the article

  • Use server git installation in GitHub for Windows

    - by Lg102
    We are using Git as the version control for our website development. I work from a laptop, which is connected to the internal network via a WiFi connection. I've mapped the server drives as network drives in Windows. Commands such as git status take significantly longer for me than they do for my co-workers on wired connections. When connecting to the server using SSH and running commands on the git installation there, performance is even better. Is there a way to configure GitHub for Windows to use the server-installed git (with my credentials)? Note: While our production servers has a user configuration with proper permissions, the development server has only one root user.

    Read the article

  • Application specific environment on the same server in Nginx/Passenger

    - by dexter
    I have two Rails applications (say app1 and app2) deployed using Nginx/Passenger. The server definition in nginx.conf looks like this: server { rails_env demo; client_max_body_size 50M; listen 80; server_name localhost; root /data/apps; passenger_enabled on; passenger_base_uri /app1; passenger_base_uri /app2; } You can see that both are configured to use demo as the RAILS_ENV. How should I change my configuration to run both the apps in different environments. Let's assume app2 is suppose to run with RAILS_ENV=qa and app1 with RAILS_ENV=demo

    Read the article

  • installing php from source

    - by samsunggalaxyss
    Hi, i have a centos5.5 server on which i want to host Magento. However i don't want to use a control panel but to do everything myself. The base repo on centos provides php 5.1.6 which is too old for the application to use. If i download and compile from source the newest php, will i also have to install the newest apache and mysql? I wan't all the pieces to work well together. I am new at linux but i know enough to install these things from latest source. But say i install the new apache 2.2.17 which has document root at /usr/local/apache2/htdocs/ how do i tell php where that is, if you understand what i mean? Thanks for your help.

    Read the article

  • How to remove file permissions from an external hard drive?

    - by user2540416
    My macbook recently died and I am currently trying to figure out how to copy my data. What I did was, I took out the hard drive, put it in an enclosure and plugged it in to my other laptop that runs linux. The problem is, I cannot copy files from the hard drive due to file permissions. I tried to access the hard drive as root. But I still cannot copy files. How do I remove file permissions from the harddrive?

    Read the article

  • Ubuntu - executable file - variable assignment throwing error on script run

    - by newcoder
    I am trying to run a small script - test - on ubuntu box. It is as follows: var1 = bash var2 = /home/test/directory ... ... <some more variable assignments and then program operations here> ... ... Now every time I run it, then it throws errors: root@localhost#/opt/test /opt/test: line 1: var1: command not found /opt/test: line 3: var2: command not found ... ... more similar errors ... Can someone help me understand what is wrong in this script? Many thanks.

    Read the article

  • Overload Protection

    - by Tyron
    Is there a simple way how I could redirect a visitor (via .htaccess or PHP script) to a static page when the server is overloaded from too many requests? It doesn't have to be a protection against huge amounts of requests at once or protect against DoS Attacks. I think our server would be protected enough if we could prevent the standard website to be shown and instead show a single file "overloaded.html". Also how could I get a measure for a server being overloaded on a typical managed server (= non root access to a Linux server) environment?

    Read the article

  • what may cause unexpected log outs?

    - by lalebarde
    Sometimes, my Linux PC unexpectedly closes my session (log out) while I am working, and puts me back the login window. If I have terminal 1 logged with root and 2 logged with another user (without X), these ones are not logged out. For example, for the last three hours, it has happened five times. Other times, it leads to a reboot. So I have investigated: the UPS, but the behaviour is unchanged without it). the temperatures, but everything is fine (cpu, disks, video card). I have started to investigate also: the memory (partially tested - I have to run a full testyt next night). EDIT: totally tested and passed. My configuration is: Linux 2.6.37-tuxonice x86_64 Intel(R) Core(TM) i7 CPU 860 @ 2.80GHz. Any idea please?

    Read the article

  • Is there good FAT driver for FUSE? (Lightweight, not mountlo)

    - by Vi.
    FUSE filesystem list show some FuseFat and FatFuse. Both are old, FatFuse is read-only , FuseFat is non-buildable and probably depends on glib. Now I'm using mountlo for the task (mounting USB drives in generic way without root access or suid things (except of fusermount itself)), but it looks too big for such task. Using FUSE to mount external storage devices is good both for security and for flexibility reason: the kernel sees only block reads and writes while actual code that deals with filesystem details runs with user privileges, allowing user to use custom filesystems and preventing from kernel filesystem exploits. Is there good vfat FUSE driver?

    Read the article

  • Ubuntu USB flash boot drive gets spontaneous "Unhandled sense code" error and causes drive to switch to Write protected

    - by Steve
    What happens is that the system runs fine for several days or even a week and then suddenly the root file-system / goes read-only. Looking at the syslog it shows that there was an 'Unhandled sense code'. This is under Ubuntu 10.04 but I saw the same thing with Ubuntu 9 with different flash media. /dev/sdg1 on / type ext4 (rw,errors=remount-ro) Jun 26 08:50:04 host1 kernel: [926247.565090] sd 5:0:0:0: [sda] Unhandled sense code Jun 26 08:50:04 host1 kernel: [926247.565094] sd 5:0:0:0: [sda] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE Jun 26 08:50:04 host1 kernel: [926247.565098] sd 5:0:0:0: [sda] Sense Key : Data Protect [current] Jun 26 08:50:04 host1 kernel: [926247.565103] sd 5:0:0:0: [sda] Add. Sense: Write protected Jun 26 08:50:04 host1 kernel: [926247.565108] sd 5:0:0:0: [sda] CDB: Write(10): 2a 00 00 46 29 18 00 00 08 00 Jun 26 08:50:04 host1 kernel: [926247.565117] end_request: I/O error, dev sda, sector 4598040 Jun 26 08:50:04 host1 kernel: [926247.569788] Buffer I/O error on device sda1, logical block 574499 Jun 26 08:50:04 host1 kernel: [926247.574677] lost page write due to I/O error on sda1

    Read the article

  • Error : [0.8879153] kernel panic -not syncing VFS unable to mount fs unknow block (8.3)

    - by user43069
    i installed ubuntu using wubit inside the windows and started working on it then i got this error afer updating [0.8879153] kernel panic -not syncing VFS unable to mount fs unknow block (8.3) and i can't user rescue mode and it's give me another error no filesystem could mount root ..... i looked at grub folder and didn' find any file on disks/boot/grub/ so i tryed to user super grub to fix it but it didn' work and it keep giving me. boot/grub/stage1 not found i didn't edit anything from grub folder. any idea plz .

    Read the article

  • Why doesn't my cron.d per minute job run?

    - by Travis Griggs
    I have thrown a bunch of darts trying to get a python script of mine to execute every minute. So I thought I'd simplify it to just do the "simplest thing that could could possibly work" once per minute (I'm running debian/testing). I created a single line file in /etc/cron.d/perminute: * * * * * /bin/touch /home/me/ding_dong It's owned by root, and executable (not sure if either of those matter). And then I did: sudo service cron reload And then sit back and start running ls -ltr again and again in my home directory (/home/me). But my ding_dong file never shows up. I know if I do a sudo /bin/touch /home/me/ding_dong, it shows up right away. Obviously missing something stupid here.

    Read the article

  • How can I disallow a user's scripts from accessing anything above their user folder?

    - by Jaxo
    This is probably an extremely simple question to answer for anybody who knows what they're doing, but I can't find any answers myself. I'm trying to set up a subdirectory for my good friend to test his PHP scripts on my (Apache) hosting plan. I don't want to let him access anything else on my server, however, for obvious reasons. His FTP login already leads him to the proper directory, which does not allow navigating any higher than it's root (mydomain.com/friend/). I would like the same behavior to be applied to any scripts, so he cannot simply <?php print_r(glob("../*")); ?> and view all my files. I'm thinking this can be done with an .htaccess file setting the DocumentRoot somewhere, but I can't have the file available for modification inside the user directory. Is this possible without majorly rewiring the web server? I've tried Googling all sorts of things to describe my problem, but without the proper terminology, all I get is "shared hosting" websites and people trying to sell me security packages.

    Read the article

  • nginx: handling 404 with error_page

    - by ytw
    Originally, I have something like this in the nginx.conf file. location ^~ /test_api { types { application/json json; } root /usr/local/www/data; rewrite "/test_api/(.*)" /api_response/test_api_$1.json break; error_page 404 /api_response/unknown_request.json; } When a requested resource is not found locally, unknown_request.json (default response) is returned correctly. Then I had to change the rewrite to point to a remote server as follows: rewrite "/test_api/(.*)" $scheme://www.somedomain.com/test_api_$1 break; It doesn't return unknown_request.json (default response) anymore even though the remote server returns a 404. Is there a way to continue to return unknown_request.json to the client when the remote server returns a 404 assuming the remote server can't be changed to return unknown_request.json? Thanks very much.

    Read the article

  • Missed something? Cant upload files to server (permissions)

    - by Camran
    I can upload files as "root" to the Ubuntu server. Then I created a user (me). Next I added the user to the group www-data. Then assigned rwx permissions to www-data. Next, when I try to upload, delete or modify files VIA FILEZILLA, I cant. But via the terminal, I can change files using sudo command. What should I do to be able to upload files without getting the "permission denied" in filezilla? If you need more input let me know. Thanks

    Read the article

  • VMWare Input/Output error on ubuntu host Win7 guest

    - by CShipley
    I am running ubuntu 10.10 as my host with Windows 7 as the guest. About a couple weeks ago I started getting error pop-ups from vmware reporting inability to read or write to a particular file. On futher investigation, I've noticed that I cannot copy that particular file from the linux command line even as root. I get the same input/output error. The SMART utility tells me I have 69 bad sectors on the disk (it is 1 TB) So my question is 1) How do I tell if that HD is going bad? 2) How can I move the VM files to a different part of the disk if I get read/write errors when I try to copy?

    Read the article

  • What is the name for a DNS record starting with @? [closed]

    - by dunxd
    Possible Duplicate: What's the meaning of '@' in a DNS zone file? I know that DNS records starting with * are called Wildcard records. What is the name for DNS record starting with @ (the at symbol). This is a record for the root domain (e.g. just example.com, not www.example.com) I want to find out more, but searching for "@ record dns" in Google doesn't return any useful results. What is the correct terminology for this type of record, and where might I find it described in more detail? RFC 1035 describes the use of @ in a DNS record, but doesn't go as far as giving it a name.

    Read the article

  • sync two huge filesystems

    - by guettli
    I need to sync two huge file systems. Both sides run linux with full root access. My preferred solution: I can read the list of changed files and directories and sync only the changed files. Here are some solutions and why they don't fit: rsync: Needs to check recursively all files. There are some million files and only little changes. The check takes too long. unison: the same: needs to check all files. inotify: I need a handler for every directory and there too many. Inotify was not build for "watch all files" scenarios. DRDB: Both sides should run independent.

    Read the article

  • Using windows 7 and fedora

    - by vedant1811
    I need to partition my hard disk for windows and fedora (root, swap, users). I thought of creating 3 (primary drives). 2 small ones (~5GB) each for win7 and fedora and a large (~700GB) for common files (pictures, vidoes, documents, etc.) one. Please Tell me which file systems to use in each case and the set-up of primary and extended partitions. Also I want to know, where and on which file system should the Linux Grub (my choice of OS chooser) be installed. I have just a bought a new Asus K53S and using the Fedora Installer Partitioner (Anaconda). Your help is greatly appreciated.

    Read the article

< Previous Page | 304 305 306 307 308 309 310 311 312 313 314 315  | Next Page >