nl
stringlengths 13
387
| bash
stringlengths 1
532
|
|---|---|
Split "bigfile" into files of at most 1000 lines each with prefix "/lots/of/little/files/here"
|
split bigfile /lots/of/little/files/here
|
Split "complete.out" into files with at most "$lines_per_file" lines each
|
split --lines $lines_per_file complete.out
|
Split "data.tsv" into files of at most 100 MiB preserving lines and use a prefix of "data.tsv." and numeric suffixes
|
split -C 100m -d data.tsv data.tsv.
|
Split "data.tsv" into files of at most 5000000 lines each with prefix "_tmp"
|
split -l5000000 data.tsv '_tmp';
|
Split "database.sql" into files of at most 100000 lines each with prefix "database-"
|
split -l 100000 database.sql database-
|
Split "date.csv" into files with at most 100 lines each
|
split -l 100 date.csv
|
Split "file.tar.gz" into files of size 1024 MB with a prefix of "file.tar.gz.part-"
|
split -b 1024m "file.tar.gz" "file.tar.gz.part-"
|
Split "file.txt" excluding the first line into files of at most 4 lines each and with a prefix "split_"
|
tail -n +2 file.txt | split -l 4 - split_
|
Split "file.txt" excluding the first line into files with at most 20 lines each and a prefix "split_"
|
tail -n +2 file.txt | split -l 20 - split_
|
Split "file.txt" into files of at most 1 MiB in size with a numeric suffix, prefix "file", and additional suffix ".txt"
|
split -b 1M -d file.txt file --additional-suffix=.txt
|
Split "file.txt" into files of at most 20 lines each with a prefix "new"
|
split -l 20 file.txt new
|
Split "foo.txt" into files with 1 line each and use a suffix length of 5
|
split --suffix-length=5 --lines=1 foo.txt
|
Split "hugefile.txt" into files with 100000 lines each starting with "part." and using numeric suffixes
|
split -a4 -d -l100000 hugefile.txt part.
|
Split "infile" into 2 files of about equal size
|
split -n2 infile
|
Split "input.txt" into files of at most 10 bytes each with prefix "/tmp/split-file"
|
split -b 10 input.txt /tmp/split-file
|
Split "input.txt" into files of at most 10 bytes each with prefix "xxx/split-file"
|
split -b 10 input.txt xxx/split-file
|
Split "input.txt" into files with 1 line each and use a prefix "output." and a suffix length of 5
|
split --lines=1 --suffix-length=5 input.txt output.
|
Split "input_file" into files of at most 100 lines each with prefix "output_file"
|
split -l 100 input_file output_file
|
Split "list.txt" into files with at most 600 lines each
|
split -l 600 list.txt
|
Split "mybigfile.txt" into files of at most 200000 lines each
|
split -l 200000 mybigfile.txt
|
Split "randn20M.csv" into files of at most 5000000 lines each with prefix "_tmp"
|
split -l5000000 randn20M.csv '_tmp';
|
Split "system.log" into files of at most 10 MiB in size with a numeric suffix and prefix "system_split.log"
|
split -b 10M -d system.log system_split.log
|
Split "t.txt" into files with at most 30000000 lines each and use a prefix "t" and numeric suffixes of length 2
|
split --lines=30000000 --numeric-suffixes --suffix-length=2 t.txt t
|
Split "your_file" into files with at most 9 lines each
|
split -l9 your_file
|
split $SOURCE_FILE" into pieces per 100 lines
|
split -l 100 "$SOURCE_FILE"
|
Split a file "file.tar.gz" into pieces named as "file.tar.gz.part-NNN" with size 1024 MB where NNN is a numeric suffix
|
split -b 1024m "file.tar.gz" "file.tar.gz.part-"
|
Split a file "file.tar.gz" into pieces with size 1024 MB
|
split -b 1024m file.tar.gz
|
split a file "list.txt" into pieces per 600 lines
|
split -l 600 list.txt
|
Split a file ADDRESSS_FILE into pieces per 20 lines named with prefix "temp_file_ and numeric suffixes
|
split -l20 ADDRESSS_FILE temp_file_
|
Split all files in the directory tree "/dev/shm/split/" into files of at most 1000 lines each and use the filename as the prefix
|
find /dev/shm/split/ -type f -exec split -l 1000 {} {} \;
|
split compressed content of the directory /home into pieces per 4000 mb named as "/media/DRIVENAME/BACKUPNAME.tgz.NNN"
|
tar --one-file-system -czv /home | split -b 4000m - /media/DRIVENAME/BACKUPNAME.tgz
|
split compressed content of the file www into pieces per 1073741824 bytes named as "backup.tar.NNN"
|
tar czf - www|split -b 1073741824 - www-backup.tar.
|
Split the contents of all ".txt" excluding the first 1000 lines into files of at most 1000 lines each
|
cat *.txt | tail -n +1001 | split --lines=1000
|
split content of the file file.txt started from second line into pieces per 4 lines named as split_NNN
|
tail -n +2 file.txt | split -l 4 - split_
|
split content of the file inputfile except lines started with "^t:" into pieces per 200 lines
|
cat inputfile | grep "^t\:" | split -l 200
|
split file "${fspec} into pieces named as "xyzzyNNN" with numeric prefix from 1 to 6
|
split --number=l/6 ${fspec} xyzzy.
|
split the file "/path/to/large/file" into pieces per 50000 lines named as /path/to/output/file/prefixNNN
|
split --lines=50000 /path/to/large/file /path/to/output/file/prefix
|
split the file "file" into pieces per 2 lines
|
split -n2 infile
|
split file "your_file" into pieces per 9 lines
|
split -l9 your_file
|
split file "$file into pieces named with 5 character suffix
|
split -a 5 $file
|
split file /etc/gconf/schemas/gnome-terminal.schemas into pieces per 1000000 lines
|
split -n 1000000 /etc/gconf/schemas/gnome-terminal.schemas
|
split file /usr/bin/firefox into pieces per 1000 lines
|
split -n 1000 /usr/bin/firefox
|
split file /usr/bin/gcc into pieces per 100000 lines
|
split -n 100000 /usr/bin/gcc
|
split file data.csv into pieces per 100 lines
|
split -l 100 date.csv
|
split the file data.csv into pieces per 100 mb named as data.tsv.NNN with digital prefixes
|
split -C 100m -d data.tsv data.tsv.
|
split the file hugefile.txt into pieces per 100000 lines named as partNNN with digital prefixes
|
split -a4 -d -l100000 hugefile.txt part.
|
split file input.txt into pieces per 1 line named output.NNNNN
|
split --lines=1 --suffix-length=5 input.txt output.
|
split file t.txt into pieces per 30000000 lines named as "t.NN" with numeric suffix
|
split --lines=30000000 --numeric-suffixes --suffix-length=2 t.txt t
|
split listing of the current directory into pieces per 500 lines named "outputXYZNNN"
|
ls | split -l 500 - outputXYZ.
|
Split the output of "ls" into files of at most 500 lines each with prefix "outputXYZ."
|
ls | split -l 500 - outputXYZ.
|
Split the output of "tar [your params]" into files of at most 500 MiB in size and use prefix "output_prefix"
|
tar [your params] |split -b 500m - output_prefix
|
split the result of command "ping -c 25 google.com | tee " into pieces per 100000 bytes named as "/home/user/myLogFile.logNNN"
|
ping -c 25 google.com | tee >(split -d -b 100000 - /home/user/myLogFile.log)
|
split result of the command "tar [your params]" into pieces per 500 mb named as "output_prefixNNN"
|
tar [your params] |split -b 500m - output_prefix
|
Split the sorted and unique lines in files "emails_*.txt" into files with at most 200 lines each with numeric suffixes of length 4
|
sort --unique emails_*.txt | split --numeric-suffixes --lines=200 --suffix-length=4 --verbose
|
Split standard input into files of at most 1000 lines each
|
split
|
Split standard input into files of at most 3400000 lines each
|
split -l 3400000
|
Split standard input into files with at most 75 lines each
|
split --lines=75
|
SSH in server 'server' as user 'user' with X11 forwarding disabled
|
ssh -x user@server
|
SSH into "$NAME" as user "${USERNAME}" using key file "${KEYDIR}/${KEY}.pem", automatically add the host to list of known hosts and execute "${COMMANDS}"
|
ssh -o "StrictHostKeyChecking no" -i ${KEYDIR}/${KEY}.pem ${USERNAME}@$NAME "${COMMANDS}"
|
SSH into "localhost" with forced pseudo-terminal allocation, execute "$heredoc", and save the output to variable "REL_DIR"
|
REL_DIR="$(ssh -t localhost "$heredoc")"
|
SSH into "myhost.com" as user "myname" with a check every 60 seconds that the server is still alive
|
ssh -o ServerAliveInterval=60 myname@myhost.com
|
ssh into default vagrant host without running "vagrant ssh" by passing the configuration parameters vagrant uses for ssh
|
ssh vagrant@127.0.0.1 -p 2222 -o Compression=yes -o DSAAuthentication=yes -o LogLevel=FATAL -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -i ~/.vagrant.d/less_insecure_private_key -o ForwardAgent=yes
|
SSH into host "$1" using key file "/path/to/ssh/secret/key" and execute command "$2"
|
ssh -i /path/to/ssh/secret/key $1 $2
|
SSH into host "server" as user "user"
|
ssh user@server
|
ssh into localhost on port 4444
|
ssh -p 4444 localhost
|
SSH into server "app1" as the current user
|
ssh app1
|
SSH into server "server.com" as user "remote_user"
|
ssh remote_user@server.com
|
SSH login in 'middle.example.org' via port 2222 as user 'deviceuser' using a private key file './device_id.pem'
|
ssh -i ./device_id.pem deviceuser@middle.example.org:2222
|
SSH using parameters in $@ in master mode in the background without executing any commands and set the ControlPath to "$MASTERSOCK"
|
ssh -o ControlPath="$MASTERSOCK" -MNf "$@"
|
SSH with parameters specified in "$@" using key file "~/.ssh/gitkey_rsa"
|
ssh -i ~/.ssh/gitkey_rsa "$@"
|
Start 'top' to monitor all processes with the default settings.
|
top
|
start 2 sleep processes in the background
|
sleep 100 | sleep 200 &
|
start from current directory, skip the directory src/emacs and print it then skip all files and directories under it, and print the names of the other files found
|
find . -wholename './src/emacs' -prune , -print
|
start from current directory, skip the directory src/emacs and print it then skip all files and directories under it, and print the names of the other files found
|
find . -wholename './src/emacs' -prune -print -o -print
|
Store content of uncompressed file "$file.fastq" in variable "reads"
|
reads=$(zcat $file.fastq)
|
Store info about all mounted file systems, printing all sizes in powers of 1000
|
a=$( df -H )
|
Store N symbols of input into variable 'buffer'
|
read -N $BUFSIZE buffer
|
Stores system load average number in the 'proc_load_average' variable.
|
proc_load_average=$(w | head -1 | cut -d" " -f13 | cut -d"," -f1-2 | tr ',' '.')
|
Stores system load average number in the 'proc_load_average' variable.
|
proc_load_average=`w | head -1 | cut -d" " -f13 | cut -d"," -f1-2 | tr ',' '.'`
|
Strips two last sections from the path $pathname, and prints basename of the rest part.
|
echo $(basename $(dirname $(dirname $pathname)))
|
Strip all '\' and newlines from $output and save the result to variable 'output'
|
output=$(echo "$output" | tr -d '\' | tr -d '\n')
|
Suffix all files and folders in the current directory with "_SUF"
|
ls | xargs -I {} mv {} {}_SUF
|
switch to user username
|
su username
|
Synchronize "/home/user1/" to "wobgalaxy02:/home/user1/" including hidden files
|
rsync -av /home/user1/ wobgalaxy02:/home/user1/
|
Synchronize "/path/to/dir_a" with files in "/path/to/dir_b/" if the files are newer
|
rsync -rtuv /path/to/dir_b/* /path/to/dir_a
|
Synchronize "/path/to/dir_b" with files in "/path/to/dir_a/" if the files are newer
|
rsync -rtuv /path/to/dir_a/* /path/to/dir_b
|
Synchronize "xxx-files" to "different-stuff/xxx-files" recursively preserving metadata with a bandwidth limit of 2000 KiB/s
|
rsync -pogtEtvr --progress --bwlimit=2000 xxx-files different-stuff
|
Take a file path from standard input and remove it.
|
xargs -I '{}' rm '{}'
|
Take a file path from standard input and remove it.
|
xargs -i rm '{}'
|
Take first text field from file 'file.txt' as a domain name and get short A record for this one.
|
cut -d' ' -f1 file.txt | xargs dig +short
|
Takes folder path from string '/path/to/copy/file/to/is/very/deep/there' and created it with all parents.
|
mkdir -p `dirname /path/to/copy/file/to/is/very/deep/there` \
|
Take the last slash-separated section of variable "FILE" and copy it to variable "NAME".
|
NAME=`basename "$FILE"`
|
Takes path list from '.exportfiles.text' file, cuts off first two path segments and last one.
|
cut -d / -f 4- .exportfiles.text | xargs -n 1 dirname
|
Take the section of variable "FILE" between the last slash and the following dot, if any, and store that section in variable "NAME".
|
NAME=`basename "$FILE" | cut -d'.' -f-1`
|
tar all files in the current folder and ask for user confirmation before creating the tar ball
|
find . -ok tar rvf backup {} \;
|
Test if a file named 'file' in the current directory is more than 1 hour old
|
find file -chour +1 -exit 0 -o -exit 1
|
Test if a file named 'file' in the current directory is more than 1 hour old
|
find file -prune -cmin +60 -print | grep -q .
|
Time stamp every ping request to 8.8.8.8 in Unix epoch format
|
ping -D -n -O -i1 -W1 8.8.8.8
|
Traverse the filesystem just once, listing setuid files and directories into /root/suid.txt and large files into /root/big.txt.
|
find / \( -perm -4000 -fprintf /root/suid.txt '%#m %u %p\n' \) , \ \( -size +100M -fprintf /root/big.txt '%-10s %p\n' \)
|
Traverse the filesystem just once, listing setuid files and directories into /root/suid.txt and large files into /root/big.txt.
|
find / \( -perm -4000 -fprintf /root/suid.txt '%#m %u %p\n' \) , \( -size +100M -fprintf /root/big.txt '%-10s %p\n' \)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.