Search Results

Search found 3730 results on 150 pages for 'bash'.

Page 80/150 | < Previous Page | 76 77 78 79 80 81 82 83 84 85 86 87  | Next Page >

  • Shell Script- each unique user

    - by Dinis Monteiro
    Hi guys I need "for each unique user, report which group they are a member of and when they last logged in" so i have: #!/bin/sh echo "Your initial login:" who | cut -d' ' -f1 | sort | uniq echo "Now is logged:" whoami echo "Group ID:" id -G $whoami case $1 in "-l") last -Fn 10 | tr -s " " ;; *) last -Fn 10 | tr -s " " | egrep -v '(^reboot)|(^$)|(^wtmp a)|(^ftp)' | cut -d" " -f1,5,7 | sort -uM | uniq -c esac My question is: how i can show the each unique user? the script above only show the more recent user logged in the system, but i need all unique users. anyone can help? thanks

    Read the article

  • Need help with shell script

    - by via-point
    I am a total newbie to Shell Scripting so please bear with me. I need to create a shell script called script1 that will calculate and then display letter grade of ABC2345. Read in the following grades from keyboard: Assignments 40% Test1 15% Test2 15% Final exam 30% Calculate and display the number grade using the weight of each factor above Convert the number grade to letter grade using the table below: Number Grade Letter Grade 90 - 100 A+ 85 - 89 A 80 - 84 77 - 79 B+ 73 - 76 B 70 - 72 B- 67 - 69 C+ 63 - 66 C 60 - 62 C- 57 - 59 D+ 53 - 56 D 50 - 52 D- 0 - 49 F Any help would be appreciated :) Thank you!

    Read the article

  • how to read rows?

    - by lego69
    I'm trying to read first row from the file > source ./rank file using this script set line = ($<) but when I enter echo $line I receive nothing, how can I change it? thanks in advance

    Read the article

  • remsh rsh error redirect problem

    - by soField
    using following command on hp-ux remsh opera -l myuser crontab -l /opt1/exp_opera_crontab 2/opt/a.log and when i echo $? i get 0 because its executing crontab -l on remote machine but i dont have opt1 directory so export wont be copied to my local machine in /opt1/exp_opera_crontab i dont get any error about this when i run this remsh or rsh command is there any way to identify both of remote and local machine related errors and redirecting them into my local machine ?

    Read the article

  • Give the mount point of a path

    - by Charles Stewart
    The following, very non-robust shell code will give the mount point of $path: (for i in $(df|cut -c 63-99); do case $path in $i*) echo $i;; esac; done) | tail -n 1 Is there a better way to do this? Postscript This script is really awful, but has the redeeming quality that it Works On My Systems. Note that several mount points may be prefixes of $path. Examples On a Linux system: cas@txtproof:~$ path=/sys/block/hda1 cas@txtproof:~$ for i in $(df -a|cut -c 57-99); do case $path in $i*) echo $i;; esac; done| tail -1 /sys On a Mac osx system cas local$ path=/dev/fd/0 cas local$ for i in $(df -a|cut -c 63-99); do case $path in $i*) echo $i;; esac; done| tail -1 /dev Note the need to vary cut's parameters, because of the way df's output differs: indeed, awk is better. Answer It looks like munging tabular output is the only way within the shell, but df /dev/fd/impossible | tail -1 | awk '{ print $NF}' is a big improvement on what I had. Note two differences in semantics: firstly, df $path insists that $path names an existing file, the script I had above doesn't care; secondly, there are no worries about dereferncing symlinks. It's not difficult to write Python code to do the job.

    Read the article

  • How to pass a variable to an awk print parameter...

    - by Jamie
    I'm trying extract the nth + 1 and nth + 3 columns from a file. This is what tried, which is a useful pseudo code: for i in {1..100} ; do awk -F "," " { printf \"%3d, %12.3f, %12.3f\\n\", \$1, \$($i+1), \$($i+3) } " All_Runs.csv > Run-$i.csv which, obviously doesn't work (but it seemed reasonable to hope). How can I do this?

    Read the article

  • How can I quickly sum all numbers in a file?

    - by Mark Roberts
    I have a file which contains several thousand numbers, each on it's own line: 34 42 11 6 2 99 ... I'm looking to write a script which will print the sum of all numbers in the file. I've got a solution, but it's not very efficient. (It takes several minutes to run.) I'm looking for a more efficient solution. Any suggestions?

    Read the article

  • How to order a ls output by suffix?

    - by Luca Borrione
    Having a ls output like GGGG_3.0.3_98/ GGGG_3.0.3_d_100/ GGGG_3.0.3_d_101/ GGGG_3.0.3_d_99/ GGGG_3.0.4_104/ GGGG_3.0.4_105/ GGGG_3.0.4_106/ GGGG_3.0_87/ GGGG_3.0_89/ GGGG_3.0_90/ GGGG_3.0_91/ GGGG_3.0_92/ GGGG_3.0_93/ SSS_2.2.3_01/ SSS_2.2.3_02/ SSS_2.2.3_03/ TTT_2.8.3_29/ how to get the elements ordered by suffix? Also, is there any quick command I can use to know that 106 is the last suffix in this example? Sorry: it wasn't clear that "the suffix" in the given example is everything following the final underscore.

    Read the article

  • How to generate changelog: git log since last Hudson build?

    - by takeshin
    I'm using Phing to do post build tasks in Hudson. I want to generate changelog containing all commits since last successful Hudson build. But looks like neither Hudson nor Git plugin for Hudson does not provide %last_build_time% variable. This would be satisfying solution, (but how to get the time?): git log --pretty="%s" --since="%last_build_time%" The only way I see for now is extracting it from the job xml file, but I do not know if it is possible with Phing. How do you generate your change logs?

    Read the article

  • Add zip files from one archive to another using command line

    - by Curious2learn
    I have two zip archives. Say, set1 has 10 csv files created using Mac OS X 10.5.8 compress option, and set2 has 4 csv files similarly created. I want to take the 4 files from zipped archive set2 and add them to list of files in archive set1. Is there a way I can do that? I tried the following in Terminal: zip set1.zip set2.zip This adds the whole archive set2.zip to set1.zip, i.e., in set1.zip now I have: file1.csv, file2.csv,..., file10.csv, set2.zip What I instead want is: file1.csv, file2.csv,..., file10.csv, file11.csv, ..., file14.csv where, set2.zip is the archive containing file11.csv, ..., file14.csv. Thanks.

    Read the article

  • How do I can linux flock command to prevent another root process deleting a file?

    - by Danmaxis
    Hello there, I would like to prevent one of my root process from deleting a certaing file. So I came across the flock command, it seems to fit my need, but I didnt get its sintax. If I only indicate a shared lock, it doesnt work: flock -s "./file.xml" If I add a timeout parameter, it still doesnt work flock -s -w5 "./file.xml" It seems that way, it fits in flock [-sxun][-w #] fd# way. (What is this fd# parameter?) So, I tried the flock [-sxon][-w #] file [-c] command Using flock -s -w5 "./file.xml" -c "tail -3 ./file.xml" and it worked, tail command at ./file.xml was executed. But I would like to know, does the lock end after the command or does it last 5 seconds after the end of the command execution? My main question is, how can I prevent another root process deleting a file in linux?

    Read the article

  • What is this project/compilation source structure?

    - by Werner
    Hi, I am working on some project my boss sent to me. After unzipping, this is the structure of the files: executable_file config.mk doc include lib Makefile sources I have only worked with "usual" Makefile projects. But I do not undersatnd the meaning of config.mk here. Do you know which kind of "build" structure (what is the right name for this concept?) is it, and where can i find more information about this? I need to modify it, so it compiles in another hardware, so I do not know where I sohuld work on. Thanks

    Read the article

  • Linux script to kill process listening on a particular port

    - by Evgeny
    I have a process that listens on a TCP port (?0003). From time to time it crashes - badly. It stops working, but continues hogging the port for some time, so I can't even restart it. I'm looking to automate this. What I do right now is: netstat -ntlp |grep -P "\*\:\d0003" To see what the PID is and then: kill -9 <pid> Does anyone have a script (or EXE for that matter) that would link the two steps together, ie. parse the PID from the first command and pass it to the second?

    Read the article

  • Socket left in TIME_WAIT after file transfer via netcat

    - by com
    Using Copying by NetCat I am trying to copy files throught network by NetCat. From console it work pretty well. First I run listening netcat on the destination machine and after I run sending on source machine. The problem is it's doen't work from script from the source machine: ssh -f user@$desthost 'nc -l 1234 | tar xvf - /dev/null &' #listening on destination host tar cv /tmp/file | nc $desthost 1234 #sending to destination host I saw that after running port 1234 is still was open and status of the socket was TIME_WAIT. If you know what's the problem, please, help me out. And by the way, after copying how can I validate that the content is identical? Thanks! Addendum: I found one very strange thing, the same implementation with screen on destination work works, but not stable, sometimes it doesn't copy a file. ssh user@$desthost screen -dm -S test 'nc -l 1234 | tar xvf - ' #listening on destination host Maybe there is an issue with timeout?

    Read the article

  • what does < < mean in the shell

    - by stib
    when looping recursively through folders with files containing spaces the shell script I use is of this form, copied from the internet: while IFS= read -r -d $'\0' file; do dosomethingwith "$file" # do something with each file done < <(find /bar -name *foo* -print0) I think I understand he IFS bit, but I don't understand what the < < characters mean. Obviously there's some sort of piping going on here.. It's very hard to google "< <", you see. TIA -stib

    Read the article

  • join 3 files by first Column with awk ?

    - by noinflection
    i have three similar files, they are all like this: File A ID1 Value1a ID2 Value2a . . . IDN Value2n and i want an output like this Output ID1 Value1a Value1b Value1c ID2 Value2a Value2b Value2c ..... IDN ValueNa ValueNb ValueNc Looking to the first line, i want value1A to be the value of id1 in fileA, value1B the value of id1 in fileB, and so on which each field and each line. I thougth it like a sql join. I've tried several things but none of them where even close.

    Read the article

  • List and Remove directories in an archive

    - by Tim
    I wonder how to list the content and remove some directories in an archive file? For example, I have an archive file data.tar. I would like to list its content without extracting it. I also would like to remove some directories inside it matching "*/count1000" Thanks and regards!

    Read the article

  • Is there a way to find a specific file and then change into the directory containing it in one go?

    - by bergyman
    I'm looking for a way to find what I know will be a unique file, and then change into the directory containing that file. Something along the lines of: find . -name 'Subscription.java' | xargs cd Or: find . -name 'Subscription.java' -exec cd {} \; I know this won't work because it's both trying to cd supplying the entire absolute path, which contains the file, and also because xargs can't do any built in shell commands...but you get the point of what I want to accomplish.

    Read the article

  • Counting unique values in a column with a shell script

    - by Lilly Tooner
    Hello. I have a tab delimited file with 5 columns and need to retrieve a count of just the number of unique lines from column 2. I would normally do this with Perl/Python but I am forced to use the shell for this one. I have successfully in the past used *nix uniq function piped to wc but it looks like I am going to have to use awk in here. Any advice would be greatly appreciated. (I have asked a similar question previously about column checks using awk but this is a little different and I wanted to separate it so if someone in the future has this question this will be here) Many many thanks! Lilly

    Read the article

  • Using /dev/tcp instead of wget

    - by User1
    Why does this work: exec 3</dev/tcp/www.google.com/80 echo -e "GET / HTTP/1.1\n\n"&3 cat <&3 And this fail: echo -e "GET / HTTP/1.1\n\n" /dev/tcp/www.google.com/80 cat </dev/tcp/www.google.com/80 Is there a way to do it in one-line w/o using wget, curl, or some other library?

    Read the article

< Previous Page | 76 77 78 79 80 81 82 83 84 85 86 87  | Next Page >