#!/usr/bin/perl -w =encoding utf8 =head1 NAME parallel_alternatives - Alternatives to GNU B =head1 DIFFERENCES BETWEEN GNU Parallel AND ALTERNATIVES There are a lot programs with some of the functionality of GNU B. GNU B strives to include the best of the functionality without sacrificing ease of use. =head2 SUMMARY TABLE The following features are in some of the comparable tools: Inputs I1. Arguments can be read from stdin I2. Arguments can be read from a file I3. Arguments can be read from multiple files I4. Arguments can be read from command line I5. Arguments can be read from a table I6. Arguments can be read from the same file using #! (shebang) I7. Line oriented input as default (Quoting of special chars not needed) Manipulation of input M1. Composed command M2. Multiple arguments can fill up an execution line M3. Arguments can be put anywhere in the execution line M4. Multiple arguments can be put anywhere in the execution line M5. Arguments can be replaced with context M6. Input can be treated as the complete command line Outputs O1. Grouping output so output from different jobs do not mix O2. Send stderr (standard error) to stderr (standard error) O3. Send stdout (standard output) to stdout (standard output) O4. Order of output can be same as order of input O5. Stdout only contains stdout (standard output) from the command O6. Stderr only contains stderr (standard error) from the command Execution E1. Running jobs in parallel E2. List running jobs E3. Finish running jobs, but do not start new jobs E4. Number of running jobs can depend on number of cpus E5. Finish running jobs, but do not start new jobs after first failure E6. Number of running jobs can be adjusted while running Remote execution R1. Jobs can be run on remote computers R2. Basefiles can be transferred R3. Argument files can be transferred R4. Result files can be transferred R5. Cleanup of transferred files R6. No config files needed R7. Do not run more than SSHD's MaxStartups can handle R8. Configurable SSH command R9. Retry if connection breaks occasionally Semaphore S1. Possibility to work as a mutex S2. Possibility to work as a counting semaphore Legend - = no x = not applicable ID = yes As every new version of the programs are not tested the table may be outdated. Please file a bug-report if you find errors (See REPORTING BUGS). parallel: I1 I2 I3 I4 I5 I6 I7 M1 M2 M3 M4 M5 M6 O1 O2 O3 O4 O5 O6 E1 E2 E3 E4 E5 E6 R1 R2 R3 R4 R5 R6 R7 R8 R9 S1 S2 xargs: I1 I2 - - - - - - M2 M3 - - - - O2 O3 - O5 O6 E1 - - - - - - - - - - x - - - - - find -exec: - - - x - x - - M2 M3 - - - - - O2 O3 O4 O5 O6 - - - - - - - - - - - - - - - - x x make -j: - - - - - - - - - - - - - O1 O2 O3 - x O6 E1 - - - E5 - - - - - - - - - - - - ppss: I1 I2 - - - - I7 M1 - M3 - - M6 O1 - - x - - E1 E2 ?E3 E4 - - R1 R2 R3 R4 - - ?R7 ? ? - - pexec: I1 I2 - I4 I5 - - M1 - M3 - - M6 O1 O2 O3 - O5 O6 E1 - - E4 - E6 R1 - - - - R6 - - - S1 - xjobs, prll, dxargs, mdm/middelman, xapply, paexec, ladon, jobflow, ClusterSSH: TODO - Please file a bug-report if you know what features they support (See REPORTING BUGS). =head2 DIFFERENCES BETWEEN xargs AND GNU Parallel B offers some of the same possibilities as GNU B. B deals badly with special characters (such as space, \, ' and "). To see the problem try this: touch important_file touch 'not important_file' ls not* | xargs rm mkdir -p "My brother's 12\" records" ls | xargs rmdir touch 'c:\windows\system32\clfs.sys' echo 'c:\windows\system32\clfs.sys' | xargs ls -l You can specify B<-0>, but many input generators are not optimized for using B as separator but are optimized for B as separator. E.g B, B, B, B, B, B, B, B (B<-0> and \0 instead of \n), B (requires using B<-0>), B (requires using B<-print0>), B (requires user to use B<-z> or B<-Z>), B (requires using B<-z>). GNU B's newline separation can be emulated with: B> B can run a given number of jobs in parallel, but has no support for running number-of-cpu-cores jobs in parallel. B has no support for grouping the output, therefore output may run together, e.g. the first half of a line is from one process and the last half of the line is from another process. The example B cannot be done reliably with B because of this. To see this in action try: parallel perl -e '\$a=\"1{}\"x10000000\;print\ \$a,\"\\n\"' '>' {} \ ::: a b c d e f ls -l a b c d e f parallel -kP4 -n1 grep 1 > out.par ::: a b c d e f echo a b c d e f | xargs -P4 -n1 grep 1 > out.xargs-unbuf echo a b c d e f | \ xargs -P4 -n1 grep --line-buffered 1 > out.xargs-linebuf echo a b c d e f | xargs -n1 grep 1 > out.xargs-serial ls -l out* md5sum out* B has no support for keeping the order of the output, therefore if running jobs in parallel using B the output of the second job cannot be postponed till the first job is done. B has no support for running jobs on remote computers. B has no support for context replace, so you will have to create the arguments. If you use a replace string in B (B<-I>) you can not force B to use more than one argument. Quoting in B works like B<-q> in GNU B. This means composed commands and redirection require using B. ls | parallel "wc {} >{}.wc" ls | parallel "echo {}; ls {}|wc" becomes (assuming you have 8 cores) ls | xargs -d "\n" -P8 -I {} bash -c "wc {} >{}.wc" ls | xargs -d "\n" -P8 -I {} bash -c "echo {}; ls {}|wc" =head2 DIFFERENCES BETWEEN find -exec AND GNU Parallel B offer some of the same possibilities as GNU B. B only works on files. So processing other input (such as hosts or URLs) will require creating these inputs as files. B has no support for running commands in parallel. =head2 DIFFERENCES BETWEEN make -j AND GNU Parallel B can run jobs in parallel, but requires a crafted Makefile to do this. That results in extra quoting to get filename containing newline to work correctly. B computes a dependency graph before running jobs. Jobs run by GNU B does not depend on eachother. (Very early versions of GNU B were coincidently implemented using B). =head2 DIFFERENCES BETWEEN ppss AND GNU Parallel B is also a tool for running jobs in parallel. The output of B is status information and thus not useful for using as input for another command. The output from the jobs are put into files. The argument replace string ($ITEM) cannot be changed. Arguments must be quoted - thus arguments containing special characters (space '"&!*) may cause problems. More than one argument is not supported. File names containing newlines are not processed correctly. When reading input from a file null cannot be used as a terminator. B needs to read the whole input file before starting any jobs. Output and status information is stored in ppss_dir and thus requires cleanup when completed. If the dir is not removed before running B again it may cause nothing to happen as B thinks the task is already done. GNU B will normally not need cleaning up if running locally and will only need cleaning up if stopped abnormally and running remote (B<--cleanup> may not complete if stopped abnormally). The example B would require extra postprocessing if written using B. For remote systems PPSS requires 3 steps: config, deploy, and start. GNU B only requires one step. =head3 EXAMPLES FROM ppss MANUAL Here are the examples from B's manual page with the equivalent using GNU B: B<1> ./ppss.sh standalone -d /path/to/files -c 'gzip ' B<1> find /path/to/files -type f | parallel gzip B<2> ./ppss.sh standalone -d /path/to/files -c 'cp "$ITEM" /destination/dir ' B<2> find /path/to/files -type f | parallel cp {} /destination/dir B<3> ./ppss.sh standalone -f list-of-urls.txt -c 'wget -q ' B<3> parallel -a list-of-urls.txt wget -q B<4> ./ppss.sh standalone -f list-of-urls.txt -c 'wget -q "$ITEM"' B<4> parallel -a list-of-urls.txt wget -q {} B<5> ./ppss config -C config.cfg -c 'encode.sh ' -d /source/dir -m 192.168.1.100 -u ppss -k ppss-key.key -S ./encode.sh -n nodes.txt -o /some/output/dir --upload --download ; ./ppss deploy -C config.cfg ; ./ppss start -C config B<5> # parallel does not use configs. If you want a different username put it in nodes.txt: user@hostname B<5> find source/dir -type f | parallel --sshloginfile nodes.txt --trc {.}.mp3 lame -a {} -o {.}.mp3 --preset standard --quiet B<6> ./ppss stop -C config.cfg B<6> killall -TERM parallel B<7> ./ppss pause -C config.cfg B<7> Press: CTRL-Z or killall -SIGTSTP parallel B<8> ./ppss continue -C config.cfg B<8> Enter: fg or killall -SIGCONT parallel B<9> ./ppss.sh status -C config.cfg B<9> killall -SIGUSR2 parallel =head2 DIFFERENCES BETWEEN pexec AND GNU Parallel B is also a tool for running jobs in parallel. =head3 EXAMPLES FROM pexec MANUAL Here are the examples from B's info page with the equivalent using GNU B: B<1> pexec -o sqrt-%s.dat -p "$(seq 10)" -e NUM -n 4 -c -- \ 'echo "scale=10000;sqrt($NUM)" | bc' B<1> seq 10 | parallel -j4 'echo "scale=10000;sqrt({})" | bc > sqrt-{}.dat' B<2> pexec -p "$(ls myfiles*.ext)" -i %s -o %s.sort -- sort B<2> ls myfiles*.ext | parallel sort {} ">{}.sort" B<3> pexec -f image.list -n auto -e B -u star.log -c -- \ 'fistar $B.fits -f 100 -F id,x,y,flux -o $B.star' B<3> parallel -a image.list \ 'fistar {}.fits -f 100 -F id,x,y,flux -o {}.star' 2>star.log B<4> pexec -r *.png -e IMG -c -o - -- \ 'convert $IMG ${IMG%.png}.jpeg ; "echo $IMG: done"' B<4> ls *.png | parallel 'convert {} {.}.jpeg; echo {}: done' B<5> pexec -r *.png -i %s -o %s.jpg -c 'pngtopnm | pnmtojpeg' B<5> ls *.png | parallel 'pngtopnm < {} | pnmtojpeg > {}.jpg' B<6> for p in *.png ; do echo ${p%.png} ; done | \ pexec -f - -i %s.png -o %s.jpg -c 'pngtopnm | pnmtojpeg' B<6> ls *.png | parallel 'pngtopnm < {} | pnmtojpeg > {.}.jpg' B<7> LIST=$(for p in *.png ; do echo ${p%.png} ; done) pexec -r $LIST -i %s.png -o %s.jpg -c 'pngtopnm | pnmtojpeg' B<7> ls *.png | parallel 'pngtopnm < {} | pnmtojpeg > {.}.jpg' B<8> pexec -n 8 -r *.jpg -y unix -e IMG -c \ 'pexec -j -m blockread -d $IMG | \ jpegtopnm | pnmscale 0.5 | pnmtojpeg | \ pexec -j -m blockwrite -s th_$IMG' B<8> Combining GNU B and GNU B. B<8> ls *jpg | parallel -j8 'sem --id blockread cat {} | jpegtopnm |' \ 'pnmscale 0.5 | pnmtojpeg | sem --id blockwrite cat > th_{}' B<8> If reading and writing is done to the same disk, this may be faster as only one process will be either reading or writing: B<8> ls *jpg | parallel -j8 'sem --id diskio cat {} | jpegtopnm |' \ 'pnmscale 0.5 | pnmtojpeg | sem --id diskio cat > th_{}' =head2 DIFFERENCES BETWEEN xjobs AND GNU Parallel B is also a tool for running jobs in parallel. It only supports running jobs on your local computer. B deals badly with special characters just like B. See the section B. Here are the examples from B's man page with the equivalent using GNU B: B<1> ls -1 *.zip | xjobs unzip B<1> ls *.zip | parallel unzip B<2> ls -1 *.zip | xjobs -n unzip B<2> ls *.zip | parallel unzip >/dev/null B<3> find . -name '*.bak' | xjobs gzip B<3> find . -name '*.bak' | parallel gzip B<4> ls -1 *.jar | sed 's/\(.*\)/\1 > \1.idx/' | xjobs jar tf B<4> ls *.jar | parallel jar tf {} '>' {}.idx B<5> xjobs -s script B<5> cat script | parallel B<6> mkfifo /var/run/my_named_pipe; xjobs -s /var/run/my_named_pipe & echo unzip 1.zip >> /var/run/my_named_pipe; echo tar cf /backup/myhome.tar /home/me >> /var/run/my_named_pipe B<6> mkfifo /var/run/my_named_pipe; cat /var/run/my_named_pipe | parallel & echo unzip 1.zip >> /var/run/my_named_pipe; echo tar cf /backup/myhome.tar /home/me >> /var/run/my_named_pipe =head2 DIFFERENCES BETWEEN prll AND GNU Parallel B is also a tool for running jobs in parallel. It does not support running jobs on remote computers. B encourages using BASH aliases and BASH functions instead of scripts. GNU B supports scripts directly, functions if they are exported using B, and aliases if using B. B generates a lot of status information on stderr (standard error) which makes it harder to use the stderr (standard error) output of the job directly as input for another program. Here is the example from B's man page with the equivalent using GNU B: prll -s 'mogrify -flip $1' *.jpg parallel mogrify -flip ::: *.jpg =head2 DIFFERENCES BETWEEN dxargs AND GNU Parallel B is also a tool for running jobs in parallel. B does not deal well with more simultaneous jobs than SSHD's MaxStartups. B is only built for remote run jobs, but does not support transferring of files. =head2 DIFFERENCES BETWEEN mdm/middleman AND GNU Parallel middleman(mdm) is also a tool for running jobs in parallel. Here are the shellscripts of http://mdm.berlios.de/usage.html ported to GNU B: seq 19 | parallel buffon -o - | sort -n > result cat files | parallel cmd find dir -execdir sem cmd {} \; =head2 DIFFERENCES BETWEEN xapply AND GNU Parallel B can run jobs in parallel on the local computer. Here are the examples from B's man page with the equivalent using GNU B: B<1> xapply '(cd %1 && make all)' */ B<1> parallel 'cd {} && make all' ::: */ B<2> xapply -f 'diff %1 ../version5/%1' manifest | more B<2> parallel diff {} ../version5/{} < manifest | more B<3> xapply -p/dev/null -f 'diff %1 %2' manifest1 checklist1 B<3> parallel --link diff {1} {2} :::: manifest1 checklist1 B<4> xapply 'indent' *.c B<4> parallel indent ::: *.c B<5> find ~ksb/bin -type f ! -perm -111 -print | xapply -f -v 'chmod a+x' - B<5> find ~ksb/bin -type f ! -perm -111 -print | parallel -v chmod a+x B<6> find */ -... | fmt 960 1024 | xapply -f -i /dev/tty 'vi' - B<6> sh <(find */ -... | parallel -s 1024 echo vi) B<6> find */ -... | parallel -s 1024 -Xuj1 vi B<7> find ... | xapply -f -5 -i /dev/tty 'vi' - - - - - B<7> sh <(find ... |parallel -n5 echo vi) B<7> find ... |parallel -n5 -uj1 vi B<8> xapply -fn "" /etc/passwd B<8> parallel -k echo < /etc/passwd B<9> tr ':' '\012' < /etc/passwd | xapply -7 -nf 'chown %1 %6' - - - - - - - B<9> tr ':' '\012' < /etc/passwd | parallel -N7 chown {1} {6} B<10> xapply '[ -d %1/RCS ] || echo %1' */ B<10> parallel '[ -d {}/RCS ] || echo {}' ::: */ B<11> xapply -f '[ -f %1 ] && echo %1' List | ... B<11> parallel '[ -f {} ] && echo {}' < List | ... =head2 DIFFERENCES BETWEEN AIX apply AND GNU Parallel B can build command lines based on a template and arguments - very much like GNU B. B does not run jobs in parallel. B does not use an argument separator (like B<:::>); instead the template must be the first argument. Here are the examples from https://www-01.ibm.com/support/knowledgecenter/ssw_aix_71/com.ibm.aix.cmds1/apply.htm 1. To obtain results similar to those of the B command, enter: apply echo * parallel echo ::: * 2. To compare the file named B to the file named B, and the file named B to the file named B, enter: apply -2 cmp a1 b1 a2 b2 parallel -N2 cmp ::: a1 b1 a2 b2 3. To run the B command five times, enter: apply -0 who 1 2 3 4 5 parallel -N0 who ::: 1 2 3 4 5 4. To link all files in the current directory to the directory B, enter: apply 'ln %1 /usr/joe' * parallel ln {} /usr/joe ::: * =head2 DIFFERENCES BETWEEN paexec AND GNU Parallel B can run jobs in parallel on both the local and remote computers. B requires commands to print a blank line as the last output. This means you will have to write a wrapper for most programs. B has a job dependency facility so a job can depend on another job to be executed successfully. Sort of a poor-man's B. Here are the examples from B's example catalog with the equivalent using GNU B: =over 1 =item 1_div_X_run: ../../paexec -s -l -c "`pwd`/1_div_X_cmd" -n +1 < sees it as a feature to have less features and in doing so it also handles corner cases incorrectly. A lot of GNU B's code is to handle corner cases correctly on every platform, so you will not get a nasty surprise if a user for example saves a file called: I B's example showing how to deal with special characters fails on special characters: echo "The Cure" > My\ brother\'s\ 12\"\ records ls | \ map 'echo -n `gzip < "%" | wc -c`; echo -n '*100/'; wc -c < "%"' | bc It works with GNU B: ls | \ parallel 'echo -n `gzip < {} | wc -c`; echo -n '*100/'; wc -c < {}' | bc And you can even get the file name prepended: ls | \ parallel --tag '(echo -n `gzip < {} | wc -c`'*100/'; wc -c < {}) | bc' B has no support for grouping. So this gives the wrong results without any warnings: parallel perl -e '\$a=\"1{}\"x10000000\;print\ \$a,\"\\n\"' '>' {} \ ::: a b c d e f ls -l a b c d e f parallel -kP4 -n1 grep 1 > out.par ::: a b c d e f map -p 4 'grep 1' a b c d e f > out.map-unbuf map -p 4 'grep --line-buffered 1' a b c d e f > out.map-linebuf map -p 1 'grep --line-buffered 1' a b c d e f > out.map-serial ls -l out* md5sum out* The documentation shows a workaround, but not only does that mix stdout (standard output) with stderr (standard error) it also fails completely for certain jobs (and may even be considered less readable): parallel echo -n {} ::: 1 2 3 map -p 4 'echo -n % 2>&1 | sed -e "s/^/$$:/"' 1 2 3 | sort | cut -f2- -d: Bs replacement strings (% %D %B %E) can be simulated in GNU B by putting this in B<~/.parallel/config>: --rpl '%' --rpl '%D $_=::shell_quote(::dirname($_));' --rpl '%B s:.*/::;s:\.[^/.]+$::;' --rpl '%E s:.*\.::' B cannot handle bundled options: B B does not have an argument separator on the command line, but uses the first argument as command. This makes quoting harder which again may affect readability. Compare: map -p 2 perl\\\ -ne\\\ \\\'/^\\\\S+\\\\s+\\\\S+\\\$/\\\ and\\\ print\\\ \\\$ARGV,\\\"\\\\n\\\"\\\' * parallel -q perl -ne '/^\S+\s+\S+$/ and print $ARGV,"\n"' ::: * B can do multiple arguments with context replace, but not without context replace: parallel --xargs echo 'BEGIN{'{}'}END' ::: 1 2 3 B does not set exit value according to whether one of the jobs failed: parallel false ::: 1 || echo Job failed map false 1 || echo Never run B requires Perl v5.10.0 making it harder to use on old systems. B has no way of using % in the command (GNU Parallel has -I to specify another replacement string than B<{}>). By design B is option incompatible with B, it does not have remote job execution, a structured way of saving results, multiple input sources, progress indicator, configurable record delimiter (only field delimiter), logging of jobs run with possibility to resume, keeping the output in the same order as input, --pipe processing, and dynamically timeouts. =head2 DIFFERENCES BETWEEN ladon AND GNU Parallel B can run multiple jobs on files in parallel. B only works on files and the only way to specify files is using a quoted glob string (such as \*.jpg). It is not possible to list the files manually. As replacement strings it uses FULLPATH DIRNAME BASENAME EXT RELDIR RELPATH These can be simulated using GNU B by putting this in B<~/.parallel/config>: --rpl 'FULLPATH $_=::shell_quote($_);chomp($_=qx{readlink -f $_});' --rpl 'DIRNAME $_=::shell_quote(::dirname($_));chomp($_=qx{readlink -f $_});' --rpl 'BASENAME s:.*/::;s:\.[^/.]+$::;' --rpl 'EXT s:.*\.::' --rpl 'RELDIR $_=::shell_quote($_);chomp(($_,$c)=qx{readlink -f $_;pwd});s:\Q$c/\E::;$_=::dirname($_);' --rpl 'RELPATH $_=::shell_quote($_);chomp(($_,$c)=qx{readlink -f $_;pwd});s:\Q$c/\E::;' B deals badly with filenames containing " and newline, and it fails for output larger than 200k: ladon '*' -- seq 36000 | wc =head3 EXAMPLES FROM ladon MANUAL It is assumed that the '--rpl's above are put in B<~/.parallel/config> and that it is run under a shell that supports '**' globbing (such as B): B<1> ladon "**/*.txt" -- echo RELPATH B<1> parallel echo RELPATH ::: **/*.txt B<2> ladon "~/Documents/**/*.pdf" -- shasum FULLPATH >hashes.txt B<2> parallel shasum FULLPATH ::: ~/Documents/**/*.pdf >hashes.txt B<3> ladon -m thumbs/RELDIR "**/*.jpg" -- convert FULLPATH -thumbnail 100x100^ -gravity center -extent 100x100 thumbs/RELPATH B<3> parallel mkdir -p thumbs/RELDIR\; convert FULLPATH -thumbnail 100x100^ -gravity center -extent 100x100 thumbs/RELPATH ::: **/*.jpg B<4> ladon "~/Music/*.wav" -- lame -V 2 FULLPATH DIRNAME/BASENAME.mp3 B<4> parallel lame -V 2 FULLPATH DIRNAME/BASENAME.mp3 ::: ~/Music/*.wav =head2 DIFFERENCES BETWEEN jobflow AND GNU Parallel B can run multiple jobs in parallel. Just like B output from B jobs running in parallel mix together by default. B can buffer into files (placed in /run/shm), but these are not cleaned up - not even if B dies unexpectently. If the total output is big (in the order of RAM+swap) it can cause the system to run out of memory. B gives no error if the command is unknown, and like B redirection requires wrapping with B. B makes it possible to set ressource limits on the running jobs. This can be emulated by GNU B using B's B: jobflow -limits=mem=100M,cpu=3,fsize=20M,nofiles=300 myjob parallel 'ulimit -v 102400 -t 3 -f 204800 -n 300 myjob' =head3 EXAMPLES FROM jobflow README B<1> cat things.list | jobflow -threads=8 -exec ./mytask {} B<1> cat things.list | parallel -j8 ./mytask {} B<2> seq 100 | jobflow -threads=100 -exec echo {} B<2> seq 100 | parallel -j100 echo {} B<3> cat urls.txt | jobflow -threads=32 -exec wget {} B<3> cat urls.txt | parallel -j32 wget {} B<4> find . -name '*.bmp' | jobflow -threads=8 -exec bmp2jpeg {.}.bmp {.}.jpg B<4> find . -name '*.bmp' | parallel -j8 bmp2jpeg {.}.bmp {.}.jpg =head2 DIFFERENCES BETWEEN gargs AND GNU Parallel B can run multiple jobs in parallel. It caches output in memory. This causes it to be extremely slow when the output is larger than the physical RAM, and can cause the system to run out of memory. See more details on this in B. Output to stderr (standard error) is changed if the command fails. Here are the two examples from B website. B<1> seq 12 -1 1 | gargs -p 4 -n 3 "sleep {0}; echo {1} {2}" B<1> seq 12 -1 1 | parallel -P 4 -n 3 "sleep {1}; echo {2} {3}" B<2> cat t.txt | gargs --sep "\s+" -p 2 "echo '{0}:{1}-{2}' full-line: \'{}\'" B<2> cat t.txt | parallel --colsep "\\s+" -P 2 "echo '{1}:{2}-{3}' full-line: \'{}\'" =head2 DIFFERENCES BETWEEN orgalorg AND GNU Parallel B can run the same job on multiple machines. This is related to B<--onall> and B<--nonall>. B supports entering the SSH password - provided it is the same for all servers. GNU B advocates using B instead, but it is possible to emulate B's behavior by setting SSHPASS and by using B<--ssh "sshpass ssh">. To make the emulation easier, make a simple alias: alias par_emul="parallel -j0 --ssh 'sshpass ssh' --nonall --tag --linebuffer" If you want to supply a password run: SSHPASS=`ssh-askpass` or set the password directly: SSHPASS=P4$$w0rd! If the above is set up you can then do: orgalorg -o frontend1 -o frontend2 -p -C uptime par_emul -S frontend1 -S frontend2 uptime orgalorg -o frontend1 -o frontend2 -p -C top -bid 1 par_emul -S frontend1 -S frontend2 top -bid 1 orgalorg -o frontend1 -o frontend2 -p -er /tmp -n 'md5sum /tmp/bigfile' -S bigfile par_emul -S frontend1 -S frontend2 --basefile bigfile --workdir /tmp md5sum /tmp/bigfile B has a progress indicator for the transferring of a file. GNU B does not. =head2 DIFFERENCES BETWEEN Rust parallel AND GNU Parallel Rust parallel focuses on speed. It is almost as fast as B. It implements a few features from GNU B, but lacks many functions. All these fail: # Show what would be executed parallel --dry-run echo ::: a # Read arguments from file parallel -a file echo # Changing the delimiter parallel -d _ echo ::: a_b_c_ These do something different from GNU B # Read more arguments at a time -n parallel -n 2 echo ::: 1 a 2 b # -q to protect quoted $ and space parallel -q perl -e '$a=shift; print "$a"x10000000' ::: a b c # Generation of combination of inputs parallel echo {1} {2} ::: red green blue ::: S M L XL XXL # {= perl expression =} replacement string parallel echo '{= s/new/old/ =}' ::: my.new your.new # --pipe seq 100000 | parallel --pipe wc # linked arguments parallel echo ::: S M L :::+ small medium large ::: R G B :::+ red green blue # Run different shell dialects zsh -c 'parallel echo \={} ::: zsh && true' csh -c 'parallel echo \$\{\} ::: shell && true' bash -c 'parallel echo \$\({}\) ::: pwd && true' # Rust parallel does not start before the last argument is read (seq 10; sleep 5; echo 2) | time parallel -j2 'sleep 2; echo' tail -f /var/log/syslog | parallel echo Rust parallel has no remote facilities. It uses /tmp/parallel for tmp files and does not clean up if terminated abrubtly. If another user on the system uses Rust parallel, then /tmp/parallel will have the wrong permissions and Rust parallel will fail. A malicious user can setup the right permissions and symlink the output file to one of the user's files and next time the user uses Rust parallel it will overwrite this file. If /tmp/parallel runs full during the run, Rust parallel does not report this, but finishes with success - thereby risking data loss. =head2 DIFFERENCES BETWEEN Rush AND GNU Parallel B (https://github.com/shenwei356/rush) is written in Go and based on B. Just like GNU B B buffers in temporary files. But opposite GNU B B does not clean up, if the process dies abnormally. B has some string manipulations that can be emulated by putting this into ~/.parallel/config (/ is used instead of %, and % is used instead of ^ as that is closer to bash's ${var%postfix}): --rpl '{:} s:(\.[^/]+)*$::' --rpl '{:%([^}]+?)} s:$$1(\.[^/]+)*$::' --rpl '{/:%([^}]*?)} s:.*/(.*)$$1(\.[^/]+)*$:$1:' --rpl '{/:} s:(.*/)?([^/.]+)(\.[^/]+)*$:$2:' --rpl '{@(.*?)} /$$1/ and $_=$1;' Here are the examples from B's website with the equivalent command in GNU B. =head3 EXAMPLES B<1. Simple run, quoting is not necessary> $ seq 1 3 | rush echo {} $ seq 1 3 | parallel echo {} B<2. Read data from file (`-i`)> $ rush echo {} -i data1.txt -i data2.txt $ cat data1.txt data2.txt | parallel echo {} B<3. Keep output order (`-k`)> $ seq 1 3 | rush 'echo {}' -k $ seq 1 3 | parallel -k echo {} B<4. Timeout (`-t`)> $ time seq 1 | rush 'sleep 2; echo {}' -t 1 $ time seq 1 | parallel --timeout 1 'sleep 2; echo {}' B<5. Retry (`-r`)> $ seq 1 | rush 'python unexisted_script.py' -r 1 $ seq 1 | parallel --retries 2 'python unexisted_script.py' Use B<-u> to see it is really run twice: $ seq 1 | parallel -u --retries 2 'python unexisted_script.py' B<6. Dirname (`{/}`) and basename (`{%}`) and remove custom suffix (`{^suffix}`)> $ echo dir/file_1.txt.gz | rush 'echo {/} {%} {^_1.txt.gz}' $ echo dir/file_1.txt.gz | parallel --plus echo {//} {/} {%_1.txt.gz} B<7. Get basename, and remove last (`{.}`) or any (`{:}`) extension> $ echo dir.d/file.txt.gz | rush 'echo {.} {:} {%.} {%:}' $ echo dir.d/file.txt.gz | parallel 'echo {.} {:} {/.} {/:}' B<8. Job ID, combine fields index and other replacement strings> $ echo 12 file.txt dir/s_1.fq.gz | rush 'echo job {#}: {2} {2.} {3%:^_1}' $ echo 12 file.txt dir/s_1.fq.gz | parallel --colsep ' ' 'echo job {#}: {2} {2.} {3/:%_1}' B<9. Capture submatch using regular expression (`{@regexp}`)> $ echo read_1.fq.gz | rush 'echo {@(.+)_\d}' $ echo read_1.fq.gz | parallel 'echo {@(.+)_\d}' B<10. Custom field delimiter (`-d`)> $ echo a=b=c | rush 'echo {1} {2} {3}' -d = $ echo a=b=c | parallel -d = echo {1} {2} {3} B<11. Send multi-lines to every command (`-n`)> $ seq 5 | rush -n 2 -k 'echo "{}"; echo' $ seq 5 | parallel -n 2 -k \ 'echo {=-1 $_=join"\n",@arg[1..$#arg] =}; echo' $ seq 5 | rush -n 2 -k 'echo "{}"; echo' -J ' ' $ seq 5 | parallel -n 2 -k 'echo {}; echo' B<12. Custom record delimiter (`-D`), note that empty records are not used.> $ echo a b c d | rush -D " " -k 'echo {}' $ echo a b c d | parallel -d " " -k 'echo {}' $ echo abcd | rush -D "" -k 'echo {}' Cannot be done by GNU Parallel $ cat fasta.fa >seq1 tag >seq2 cat gat >seq3 attac a cat $ cat fasta.fa | rush -D ">" \ 'echo FASTA record {#}: name: {1} sequence: {2}' -k -d "\n" # rush fails to join the multiline sequences $ cat fasta.fa | (read -n1 ignore_first_char; parallel -d '>' --colsep '\n' echo FASTA record {#}: \ name: {1} sequence: '{=2 $_=join"",@arg[2..$#arg]=}' ) B<13. Assign value to variable, like `awk -v` (`-v`)> $ seq 1 | rush 'echo Hello, {fname} {lname}!' -v fname=Wei -v lname=Shen $ seq 1 | parallel -N0 \ 'fname=Wei; lname=Shen; echo Hello, ${fname} ${lname}!' $ for var in a b; do \ $ seq 1 3 | rush -k -v var=$var 'echo var: {var}, data: {}'; \ $ done In GNU B you would typically do: $ seq 1 3 | parallel -k echo var: {1}, data: {2} ::: a b :::: - If you I want the var: $ seq 1 3 | parallel -k var={1} ';echo var: $var, data: {}' ::: a b :::: - If you I want the B-loop: $ for var in a b; do > export var; > seq 1 3 | parallel -k 'echo var: $var, data: {}'; > done Contrary to B this also works if the value is complex like: My brother's 12" records B<14. B (`-v`), avoid repeatedly writing verbose replacement strings> # naive way $ echo read_1.fq.gz | rush 'echo {:^_1} {:^_1}_2.fq.gz' $ echo read_1.fq.gz | parallel 'echo {:%_1} {:%_1}_2.fq.gz' # macro + removing suffix $ echo read_1.fq.gz | rush -v p='{:^_1}' 'echo {p} {p}_2.fq.gz' $ echo read_1.fq.gz | parallel 'p={:%_1}; echo $p ${p}_2.fq.gz' # macro + regular expression $ echo read_1.fq.gz | rush -v p='{@(.+?)_\d}' 'echo {p} {p}_2.fq.gz' $ echo read_1.fq.gz | parallel 'p={@(.+?)_\d}; echo $p ${p}_2.fq.gz' Contrary to B GNU B works with complex values: echo "My brother's 12\"read_1.fq.gz" | parallel 'p={@(.+?)_\d}; echo $p ${p}_2.fq.gz' B<15. Interrupt jobs by `Ctrl-C`, rush will stop unfinished commands and exit.> $ seq 1 20 | rush 'sleep 1; echo {}' ^C $ seq 1 20 | parallel 'sleep 1; echo {}' ^C B<16. Continue/resume jobs (`-c`). When some jobs failed (by execution failure, timeout, or cancelling by user with `Ctrl + C`), please switch flag `-c/--continue` on and run again, so that `rush` can save successful commands and ignore them in I run.> $ seq 1 3 | rush 'sleep {}; echo {}' -t 3 -c $ cat successful_cmds.rush $ seq 1 3 | rush 'sleep {}; echo {}' -t 3 -c $ seq 1 3 | parallel --joblog mylog --timeout 2 \ 'sleep {}; echo {}' $ cat mylog $ seq 1 3 | parallel --joblog mylog --retry-failed \ 'sleep {}; echo {}' Multi-line jobs: $ seq 1 3 | rush 'sleep {}; echo {}; \ echo finish {}' -t 3 -c -C finished.rush $ cat finished.rush $ seq 1 3 | rush 'sleep {}; echo {}; \ echo finish {}' -t 3 -c -C finished.rush $ seq 1 3 | parallel --joblog mylog --timeout 2 'sleep {}; echo {}; \ echo finish {}' $ cat mylog $ seq 1 3 | parallel --joblog mylog --retry-failed 'sleep {}; echo {}; \ echo finish {}' B<17. A comprehensive example: downloading 1K+ pages given by three URL list files using `phantomjs save_page.js` (some page contents are dynamicly generated by Javascript, so `wget` does not work). Here I set max jobs number (`-j`) as `20`, each job has a max running time (`-t`) of `60` seconds and `3` retry changes (`-r`). Continue flag `-c` is also switched on, so we can continue unfinished jobs. Luckily, it's accomplished in one run :)> $ for f in $(seq 2014 2016); do \ $ /bin/rm -rf $f; mkdir -p $f; \ $ cat $f.html.txt | rush -v d=$f -d = \ 'phantomjs save_page.js "{}" > {d}/{3}.html' \ -j 20 -t 60 -r 3 -c; \ $ done GNU B can append to an existing joblog with '+': $ rm mylog $ for f in $(seq 2014 2016); do /bin/rm -rf $f; mkdir -p $f; cat $f.html.txt | parallel -j20 --timeout 60 --retries 4 --joblog +mylog \ --colsep = \ phantomjs save_page.js {1}={2}={3} '>' $f/{3}.html done B<18. A bioinformatics example: mapping with `bwa`, and processing result with `samtools`:> $ ref=ref/xxx.fa $ threads=25 $ ls -d raw.cluster.clean.mapping/* \ | rush -v ref=$ref -v j=$threads -v p='{}/{%}' \ 'bwa mem -t {j} -M -a {ref} {p}_1.fq.gz {p}_2.fq.gz > {p}.sam; \ samtools view -bS {p}.sam > {p}.bam; \ samtools sort -T {p}.tmp -@ {j} {p}.bam -o {p}.sorted.bam; \ samtools index {p}.sorted.bam; \ samtools flagstat {p}.sorted.bam > {p}.sorted.bam.flagstat; \ /bin/rm {p}.bam {p}.sam;' \ -j 2 --verbose -c -C mapping.rush GNU B would use a function: $ ref=ref/xxx.fa $ export ref $ thr=25 $ export thr $ bwa_sam() { p="$1" bam="$p".bam sam="$p".sam sortbam="$p".sorted.bam bwa mem -t $thr -M -a $ref ${p}_1.fq.gz ${p}_2.fq.gz > "$sam" samtools view -bS "$sam" > "$bam" samtools sort -T ${p}.tmp -@ $thr "$bam" -o "$sortbam" samtools index "$sortbam" samtools flagstat "$sortbam" > "$sortbam".flagstat /bin/rm "$bam" "$sam" } $ export -f bwa_sam $ ls -d raw.cluster.clean.mapping/* | parallel -j 2 --verbose --joblog mylog bwa_sam =head3 Other B features B has: =over 4 =item * B like custom defined variables (B<-v>) With GNU B you would simply simply set a shell variable: parallel 'v={}; echo "$v"' ::: foo echo foo | rush -v v={} 'echo {v}' Also B does not like special chars. So these B: echo does not work | rush -v v=\" 'echo {v}' echo "My brother's 12\" records" | rush -v v={} 'echo {v}' Whereas the corresponding GNU B version works: parallel 'v=\"; echo "$v"' ::: works parallel 'v={}; echo "$v"' ::: "My brother's 12\" records" =item * Exit on first error(s) (-e) This is called B<--halt now,fail=1> (or shorter: B<--halt 2>) when used with GNU B. =item * Settable records sending to every command (B<-n>, default 1) This is also called B<-n> in GNU B. =item * Practical replacement strings =over 4 =item {:} remove any extension With GNU B this can be emulated by: parallel --plus echo '{/\..*/}' ::: foo.ext.bar.gz =item {^suffix}, remove suffix With GNU B this can be emulated by: parallel --plus echo '{%.bar.gz}' ::: foo.ext.bar.gz =item {@regexp}, capture submatch using regular expression With GNU B this can be emulated by: parallel --rpl '{@(.*?)} /$$1/ and $_=$1;' \ echo '{@\d_(.*).gz}' ::: 1_foo.gz =item {%.}, {%:}, basename without extension With GNU B this can be emulated by: parallel echo '{= s:.*/::;s/\..*// =}' ::: dir/foo.bar.gz And if you need it often, you define a B<--rpl> in B<$HOME/.parallel/config>: --rpl '{%.} s:.*/::;s/\..*//' --rpl '{%:} s:.*/::;s/\..*//' Then you can use them as: parallel echo {%.} {%:} ::: dir/foo.bar.gz =back =item * Preset variable (macro) E.g. echo foosuffix | rush -v p={^suffix} 'echo {p}_new_suffix' With GNU B this can be emulated by: echo foosuffix | parallel --plus 'p={%suffix}; echo ${p}_new_suffix' Opposite B GNU B works fine if the input contains double space, ' and ": echo "1'6\" foosuffix" | parallel --plus 'p={%suffix}; echo "${p}"_new_suffix' =item * Commands of multi-lines While you I use multi-lined commands in GNU B, to improve readibilty GNU B discourages the use of multi-line commands. In most cases it can be written as a function: seq 1 3 | parallel --timeout 2 --joblog my.log 'sleep {}; echo {}; \ echo finish {}' Could be written as: doit() { sleep "$1" echo "$1" echo finish "$1" } export -f doit seq 1 3 | parallel --timeout 2 --joblog my.log doit The failed commands can be resumed with: seq 1 3 | parallel --resume-failed --joblog my.log 'sleep {}; echo {};\ echo finish {}' =back =head2 DIFFERENCES BETWEEN ClusterSSH AND GNU Parallel ClusterSSH solves a different problem than GNU B. ClusterSSH opens a terminal window for each computer and using a master window you can run the same command on all the computers. This is typically used for administrating several computers that are almost identical. GNU B runs the same (or different) commands with different arguments in parallel possibly using remote computers to help computing. If more than one computer is listed in B<-S> GNU B may only use one of these (e.g. if there are 8 jobs to be run and one computer has 8 cores). GNU B can be used as a poor-man's version of ClusterSSH: B =head2 DIFFERENCES BETWEEN coshell AND GNU Parallel B only accepts full commands on standard input. Any quoting needs to be done by the user. Commands are run in B so any B/B/B specific syntax will not work. Output can be buffered by using B<-d>. Output is buffered in memory, so big output can cause swapping and therefore be terrible slow or even cause out of memory. =head2 DIFFERENCES BETWEEN spread AND GNU Parallel B runs commands on all directories. It can be emulated with GNU B using this Bash function: spread() { _cmds() { perl -e '$"=" && ";print "@ARGV"' "cd {}" "$@" } parallel $(_cmds "$@")'|| echo exit status $?' ::: */ } This works execpt for the B<--exclude> option. =head2 Todo machma. Requires Go >= 1.7. https://github.com/k-bx/par requires Haskell to work. This limits the number of platforms this can work on. https://github.com/otonvm/Parallel https://github.com/flesler/parallel https://github.com/kou1okada/lesser-parallel https://github.com/Julian/Verge https://github.com/amattn/paral =head1 TESTING OTHER TOOLS There are certain issues that are very common on parallelizing tools. Here are a few stress tests. Be warned: If the tool is badly coded it may overload you machine. =head2 A: Output mixes Output from 2 jobs should not mix. If the tool does not buffer, output will most likely mix. #!/bin/bash paralleltool=parallel cat <<-EOF > mycommand #!/bin/bash # If 'a', 'b' and 'c' mix: Very bad perl -e 'print "a"x3000_000," "' perl -e 'print "b"x3000_000," "' perl -e 'print "c"x3000_000," "' echo EOF chmod +x mycommand # Run 30 jobs in parallel seq 30 | $paralleltool -j0 ./mycommand | tr -s abc # 'a b c' should always stay together # and there should only be a single line per job =head2 B: Output limited by RAM Some tools cache output in RAM. This makes them extremely slow if the output is bigger than physical memory and crash if the the output is bigger than the virtual memory. #!/bin/bash paralleltool=parallel cat <<'EOF' > mycommand #!/bin/bash # Generate 1 GB output yes "`perl -e 'print \"c\"x30_000'`" | head -c 1G EOF chmod +x mycommand # Run 20 jobs in parallel # Adjust 20 to be > physical RAM and < free space on /tmp seq 20 | time $paralleltool -j0 ./mycommand | wc -c =head2 C: Leaving tmp files at unexpected death Some tools do not clean up tmp files if they are killed. If the tool buffers on disk, they may not clean up, if they are killed. #!/bin/bash paralleltool=parallel ls /tmp >/tmp/before seq 10 | $paralleltool sleep & pid=$! # Give the tool time to start up sleep 1 # Kill it without giving it a chance to cleanup kill -9 $! # Should be empty: No files should be left behind diff <(ls /tmp) /tmp/before =head2 D: Dealing badly with special file names. It is not uncommon for users to create files like: My brother's 12" records cost $$$.txt Some tools break on this. #!/bin/bash paralleltool=parallel touch "My brother's 12\" records cost \$\$\$.txt" ls My*txt | $paralleltool echo =head2 E: Composed commands do not work Some tools require you to wrap composed commands into B. echo bar | $paralleltool echo foo';' echo {} =head2 F: Only one replacement string allowed Some tools can only insert the argument once. echo bar | $paralleltool echo {} foo {} =head2 G: Speed depends on number of words Some tools become very slow if output lines have many words. #!/bin/bash paralleltool=parallel cat <<-EOF > mycommand #!/bin/bash # 10 MB of lines with 1000 words yes "`seq 1000`" | head -c 10M EOF chmod +x mycommand # Run 30 jobs in parallel seq 30 | time $paralleltool -j0 ./mycommand > /dev/null =head1 AUTHOR When using GNU B for a publication please cite: O. Tange (2011): GNU Parallel - The Command-Line Power Tool, ;login: The USENIX Magazine, February 2011:42-47. This helps funding further development; and it won't cost you a cent. If you pay 10000 EUR you should feel free to use GNU Parallel without citing. Copyright (C) 2007-10-18 Ole Tange, http://ole.tange.dk Copyright (C) 2008,2009,2010 Ole Tange, http://ole.tange.dk Copyright (C) 2010,2011,2012,2013,2014,2015,2016,2017 Ole Tange, http://ole.tange.dk and Free Software Foundation, Inc. Parts of the manual concerning B compatibility is inspired by the manual of B from GNU findutils 4.4.2. =head1 LICENSE Copyright (C) 2007,2008,2009,2010,2011,2012,2013,2014,2015,2016,2017 Free Software Foundation, Inc. This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3 of the License, or at your option any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program. If not, see . =head2 Documentation license I Permission is granted to copy, distribute and/or modify this documentation under the terms of the GNU Free Documentation License, Version 1.3 or any later version published by the Free Software Foundation; with no Invariant Sections, with no Front-Cover Texts, and with no Back-Cover Texts. A copy of the license is included in the file fdl.txt. =head2 Documentation license II You are free: =over 9 =item B to copy, distribute and transmit the work =item B to adapt the work =back Under the following conditions: =over 9 =item B You must attribute the work in the manner specified by the author or licensor (but not in any way that suggests that they endorse you or your use of the work). =item B If you alter, transform, or build upon this work, you may distribute the resulting work only under the same, similar or a compatible license. =back With the understanding that: =over 9 =item B Any of the above conditions can be waived if you get permission from the copyright holder. =item B Where the work or any of its elements is in the public domain under applicable law, that status is in no way affected by the license. =item B In no way are any of the following rights affected by the license: =over 2 =item * Your fair dealing or fair use rights, or other applicable copyright exceptions and limitations; =item * The author's moral rights; =item * Rights other persons may have either in the work itself or in how the work is used, such as publicity or privacy rights. =back =back =over 9 =item B For any reuse or distribution, you must make clear to others the license terms of this work. =back A copy of the full license is included in the file as cc-by-sa.txt. =head1 DEPENDENCIES GNU B uses Perl, and the Perl modules Getopt::Long, IPC::Open3, Symbol, IO::File, POSIX, and File::Temp. For remote usage it also uses rsync with ssh. =head1 SEE ALSO B(1), B(1), B(1), B(1), B(1), B(1), B(1), B(1), B(1) =cut