mirror of
https://git.savannah.gnu.org/git/parallel.git
synced 2024-11-23 06:27:55 +00:00
2328 lines
53 KiB
Plaintext
2328 lines
53 KiB
Plaintext
#!/usr/bin/perl -w
|
|
|
|
=head1 GNU Parallel Tutorial
|
|
|
|
This tutorial shows off much of GNU Parallel's functionality. The
|
|
tutorial is meant to learn the options in GNU Parallel. The tutorial
|
|
is not to show realistic examples from the real world.
|
|
|
|
Spend an hour walking through the tutorial. Your command line will
|
|
love you for it.
|
|
|
|
=head1 Prerequisites
|
|
|
|
To run this tutorial you must have the following:
|
|
|
|
=over 9
|
|
|
|
=item parallel >= version 20140622
|
|
|
|
Install the newest version with:
|
|
|
|
(wget -O - pi.dk/3 || curl pi.dk/3/ || fetch -o - http://pi.dk/3) | bash
|
|
|
|
This will also install the newest version of the tutorial:
|
|
|
|
man parallel_tutorial
|
|
|
|
Most of the tutorial will work on older versions, too.
|
|
|
|
|
|
=item abc-file:
|
|
|
|
The file can be generated by:
|
|
|
|
parallel -k echo ::: A B C > abc-file
|
|
|
|
=item def-file:
|
|
|
|
The file can be generated by:
|
|
|
|
parallel -k echo ::: D E F > def-file
|
|
|
|
=item abc0-file:
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'printf "A\0B\0C\0"' > abc0-file
|
|
|
|
=item abc_-file:
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'printf "A_B_C_"' > abc_-file
|
|
|
|
=item tsv-file.tsv
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'printf "f1\tf2\nA\tB\nC\tD\n"' > tsv-file.tsv
|
|
|
|
=item num8
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'for(1..8){print "$_\n"}' > num8
|
|
|
|
=item num128
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'for(1..128){print "$_\n"}' > num128
|
|
|
|
=item num30000
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'for(1..30000){print "$_\n"}' > num30000
|
|
|
|
=item num1000000
|
|
|
|
The file can be generated by:
|
|
|
|
perl -e 'for(1..1000000){print "$_\n"}' > num1000000
|
|
|
|
=item num_%header
|
|
|
|
The file can be generated by:
|
|
|
|
(echo %head1; echo %head2; perl -e 'for(1..10){print "$_\n"}') > num_%header
|
|
|
|
=item For remote running: ssh login on 2 servers with no password in
|
|
$SERVER1 and $SERVER2
|
|
|
|
SERVER1=server.example.com
|
|
SERVER2=server2.example.net
|
|
|
|
You must be able to:
|
|
|
|
ssh $SERVER1 echo works
|
|
ssh $SERVER2 echo works
|
|
|
|
It can be setup by running 'ssh-keygen -t dsa; ssh-copy-id $SERVER1'
|
|
and using an empty pass phrase.
|
|
|
|
=back
|
|
|
|
|
|
=head1 Input sources
|
|
|
|
GNU Parallel reads input from input sources. These can be files, the
|
|
command line, and stdin (standard input or a pipe).
|
|
|
|
=head2 A single input source
|
|
|
|
Input can be read from the command line:
|
|
|
|
parallel echo ::: A B C
|
|
|
|
Output (the order may be different because the jobs are run in
|
|
parallel):
|
|
|
|
A
|
|
B
|
|
C
|
|
|
|
The input source can be a file:
|
|
|
|
parallel -a abc-file echo
|
|
|
|
Output: Same as above.
|
|
|
|
STDIN (standard input) can be the input source:
|
|
|
|
cat abc-file | parallel echo
|
|
|
|
Output: Same as above.
|
|
|
|
|
|
=head2 Multiple input sources
|
|
|
|
GNU Parallel can take multiple input sources given on the command
|
|
line. GNU Parallel then generates all combinations of the input
|
|
sources:
|
|
|
|
parallel echo ::: A B C ::: D E F
|
|
|
|
Output (the order may be different):
|
|
|
|
A D
|
|
A E
|
|
A F
|
|
B D
|
|
B E
|
|
B F
|
|
C D
|
|
C E
|
|
C F
|
|
|
|
The input sources can be files:
|
|
|
|
parallel -a abc-file -a def-file echo
|
|
|
|
Output: Same as above.
|
|
|
|
STDIN (standard input) can be one of the input sources using '-':
|
|
|
|
cat abc-file | parallel -a - -a def-file echo
|
|
|
|
Output: Same as above.
|
|
|
|
Instead of -a files can be given after '::::':
|
|
|
|
cat abc-file | parallel echo :::: - def-file
|
|
|
|
Output: Same as above.
|
|
|
|
::: and :::: can be mixed:
|
|
|
|
parallel echo ::: A B C :::: def-file
|
|
|
|
Output: Same as above.
|
|
|
|
=head3 Matching arguments from all input sources
|
|
|
|
With --xapply you can get one argument from each input source:
|
|
|
|
parallel --xapply echo ::: A B C ::: D E F
|
|
|
|
Output (the order may be different):
|
|
|
|
A D
|
|
B E
|
|
C F
|
|
|
|
If one of the input sources is too short, its values will wrap:
|
|
|
|
parallel --xapply echo ::: A B C D E ::: F G
|
|
|
|
Output (the order may be different):
|
|
|
|
A F
|
|
B G
|
|
C F
|
|
D G
|
|
E F
|
|
|
|
=head2 Changing the argument separator.
|
|
|
|
GNU Parallel can use other separators than ::: or ::::. This is
|
|
typically useful if ::: or :::: is used in the command to run:
|
|
|
|
parallel --arg-sep ,, echo ,, A B C :::: def-file
|
|
|
|
Output (the order may be different):
|
|
|
|
A D
|
|
A E
|
|
A F
|
|
B D
|
|
B E
|
|
B F
|
|
C D
|
|
C E
|
|
C F
|
|
|
|
Changing the argument file separator:
|
|
|
|
parallel --arg-file-sep // echo ::: A B C // def-file
|
|
|
|
Output: Same as above.
|
|
|
|
|
|
=head2 Changing the argument delimiter
|
|
|
|
GNU Parallel will normally treat a full line as a single argument: It
|
|
uses \n as argument delimiter. This can be changed with -d:
|
|
|
|
parallel -d _ echo :::: abc_-file
|
|
|
|
Output (the order may be different):
|
|
|
|
A
|
|
B
|
|
C
|
|
|
|
NULL can be given as \0:
|
|
|
|
parallel -d '\0' echo :::: abc0-file
|
|
|
|
Output: Same as above.
|
|
|
|
A shorthand for -d '\0' is -0 (this will often be used to read files
|
|
from find ... -print0):
|
|
|
|
parallel -0 echo :::: abc0-file
|
|
|
|
Output: Same as above.
|
|
|
|
=head2 End-of-file value for input source
|
|
|
|
GNU Parallel can stop reading when it encounters a certain value:
|
|
|
|
parallel -E stop echo ::: A B stop C D
|
|
|
|
Output:
|
|
|
|
A
|
|
B
|
|
|
|
=head2 Skipping empty lines
|
|
|
|
Using --no-run-if-empty GNU Parallel will skip empty lines.
|
|
|
|
(echo 1; echo; echo 2) | parallel --no-run-if-empty echo
|
|
|
|
Output:
|
|
|
|
1
|
|
2
|
|
|
|
|
|
=head1 Building the command line
|
|
|
|
=head2 No command means arguments are commands
|
|
|
|
If no command is given after parallel the arguments themselves are
|
|
treated as commands:
|
|
|
|
parallel ::: ls 'echo foo' pwd
|
|
|
|
Output (the order may be different):
|
|
|
|
[list of files in current dir]
|
|
foo
|
|
[/path/to/current/working/dir]
|
|
|
|
The command can be a script, a binary or a Bash function if the function is
|
|
exported using 'export -f':
|
|
|
|
# Only works in Bash
|
|
my_func() {
|
|
echo in my_func $1
|
|
}
|
|
export -f my_func
|
|
parallel my_func ::: 1 2 3
|
|
|
|
Output (the order may be different):
|
|
|
|
in my_func 1
|
|
in my_func 2
|
|
in my_func 3
|
|
|
|
=head2 Replacement strings
|
|
|
|
=head3 The 7 predefined replacement strings
|
|
|
|
GNU Parallel has several replacement strings. If no replacement
|
|
strings are used the default is to append {}:
|
|
|
|
parallel echo ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A/B.C
|
|
|
|
The default replacement string is {}:
|
|
|
|
parallel echo {} ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A/B.C
|
|
|
|
The replacement string {.} removes the extension:
|
|
|
|
parallel echo {.} ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A/B
|
|
|
|
The replacement string {/} removes the path:
|
|
|
|
parallel echo {/} ::: A/B.C
|
|
|
|
Output:
|
|
|
|
B.C
|
|
|
|
The replacement string {//} keeps only the path:
|
|
|
|
parallel echo {//} ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A
|
|
|
|
The replacement string {/.} removes the path and the extension:
|
|
|
|
parallel echo {/.} ::: A/B.C
|
|
|
|
Output:
|
|
|
|
B
|
|
|
|
The replacement string {#} gives the job number:
|
|
|
|
parallel echo {#} ::: A B C
|
|
|
|
Output (the order may be different):
|
|
|
|
1
|
|
2
|
|
3
|
|
|
|
The replacement string {%} gives the job slot number (between 1 and
|
|
number of jobs to run in parallel):
|
|
|
|
parallel -j 2 echo {%} ::: A B C
|
|
|
|
Output (the order may be different and 1 and 2 may be swapped):
|
|
|
|
1
|
|
2
|
|
1
|
|
|
|
=head3 Changing the replacement strings
|
|
|
|
The replacement string {} can be changed with -I:
|
|
|
|
parallel -I ,, echo ,, ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A/B.C
|
|
|
|
The replacement string {.} can be changed with --extensionreplace:
|
|
|
|
parallel --extensionreplace ,, echo ,, ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A/B
|
|
|
|
The replacement string {/} can be replaced with --basenamereplace:
|
|
|
|
parallel --basenamereplace ,, echo ,, ::: A/B.C
|
|
|
|
Output:
|
|
|
|
B.C
|
|
|
|
The replacement string {//} can be changed with --dirnamereplace:
|
|
|
|
parallel --dirnamereplace ,, echo ,, ::: A/B.C
|
|
|
|
Output:
|
|
|
|
A
|
|
|
|
The replacement string {/.} can be changed with --basenameextensionreplace:
|
|
|
|
parallel --basenameextensionreplace ,, echo ,, ::: A/B.C
|
|
|
|
Output:
|
|
|
|
B
|
|
|
|
The replacement string {#} can be changed with --seqreplace:
|
|
|
|
parallel --seqreplace ,, echo ,, ::: A B C
|
|
|
|
Output (the order may be different):
|
|
|
|
1
|
|
2
|
|
3
|
|
|
|
The replacement string {%} can be changed with --slotreplace:
|
|
|
|
parallel -j2 --slotreplace ,, echo ,, ::: A B C
|
|
|
|
Output (the order may be different and 1 and 2 may be swapped):
|
|
|
|
1
|
|
2
|
|
1
|
|
|
|
=head3 Perl expression replacement string
|
|
|
|
When predefined replacement strings are not flexible enough a perl
|
|
expression can be used instead. One example is to remove two
|
|
extensions: foo.tar.gz -> foo
|
|
|
|
parallel echo '{= s:\.[^.]+$::;s:\.[^.]+$::; =}' ::: foo.tar.gz
|
|
|
|
Output:
|
|
|
|
foo
|
|
|
|
If the strings B<{=> and B<=}> cause problems they can be replaced with --parens:
|
|
|
|
parallel --parens ,,,, echo ',, s:\.[^.]+$::;s:\.[^.]+$::; ,,' ::: foo.tar.gz
|
|
|
|
Output: Same as above.
|
|
|
|
To define a short hand replacement string use B<--rpl>:
|
|
|
|
parallel --rpl '.. s:\.[^.]+$::;s:\.[^.]+$::;' echo '..' ::: foo.tar.gz
|
|
|
|
Output: Same as above.
|
|
|
|
If the short hand starts with '{' it can be used as a positional
|
|
replacement string, too:
|
|
|
|
parallel --rpl '{..} s:\.[^.]+$::;s:\.[^.]+$::;' echo '{..}' ::: foo.tar.gz
|
|
|
|
Output: Same as above.
|
|
|
|
GNU B<parallel>'s 7 replacement strings are implemented as:
|
|
|
|
--rpl '{} '
|
|
--rpl '{#} $_=$job->seq()'
|
|
--rpl '{%} $_=$job->slot()'
|
|
--rpl '{/} s:.*/::'
|
|
--rpl '{//} $Global::use{"File::Basename"} ||= eval "use File::Basename; 1;"; $_ = dirname($_);'
|
|
--rpl '{/.} s:.*/::; s:\.[^/.]+$::;'
|
|
--rpl '{.} s:\.[^/.]+$::'
|
|
|
|
=head3 Positional replacement strings
|
|
|
|
With multiple input sources the argument from the individual input
|
|
sources can be access with {number}:
|
|
|
|
parallel echo {1} and {2} ::: A B ::: C D
|
|
|
|
Output (the order may be different):
|
|
|
|
A and C
|
|
A and D
|
|
B and C
|
|
B and D
|
|
|
|
The positional replacement strings can also be modified using / // /. and .:
|
|
|
|
parallel echo /={1/} //={1//} /.={1/.} .={1.} ::: A/B.C D/E.F
|
|
|
|
Output (the order may be different):
|
|
|
|
/=B.C //=A /.=B .=A/B
|
|
/=E.F //=D /.=E .=D/E
|
|
|
|
If a position is negative, it will refer to the input source counted
|
|
from behind:
|
|
|
|
parallel echo 1={1} 2={2} 3={3} -1={-1} -2={-2} -3={-3} ::: A B ::: C D ::: E F
|
|
|
|
Output (the order may be different):
|
|
|
|
1=A 2=C 3=E -1=E -2=C -3=A
|
|
1=A 2=C 3=F -1=F -2=C -3=A
|
|
1=A 2=D 3=E -1=E -2=D -3=A
|
|
1=A 2=D 3=F -1=F -2=D -3=A
|
|
1=B 2=C 3=E -1=E -2=C -3=B
|
|
1=B 2=C 3=F -1=F -2=C -3=B
|
|
1=B 2=D 3=E -1=E -2=D -3=B
|
|
1=B 2=D 3=F -1=F -2=D -3=B
|
|
|
|
|
|
=head3 Positional perl expression replacement string
|
|
|
|
To use a perl expression as a positional replacement string simply
|
|
prepend the perl expression with number and space:
|
|
|
|
parallel echo '{=2 s:\.[^.]+$::;s:\.[^.]+$::; =} {1}' ::: bar ::: foo.tar.gz
|
|
|
|
Output:
|
|
|
|
foo bar
|
|
|
|
If a defined short hand starts with '{' it can be used as a positional
|
|
replacement string, too:
|
|
|
|
parallel --rpl '{..} s:\.[^.]+$::;s:\.[^.]+$::;' echo '{2..} {1}' ::: bar ::: foo.tar.gz
|
|
|
|
Output: Same as above.
|
|
|
|
|
|
=head3 Input from columns
|
|
|
|
The columns in a file can be bound to positional replacement strings
|
|
using --colsep. Here the columns are separated with TAB (\t):
|
|
|
|
parallel --colsep '\t' echo 1={1} 2={2} :::: tsv-file.tsv
|
|
|
|
Output (the order may be different):
|
|
|
|
1=f1 2=f2
|
|
1=A 2=B
|
|
1=C 2=D
|
|
|
|
=head3 Header defined replacement strings
|
|
|
|
With --header GNU Parallel will use the first value of the input
|
|
source as the name of the replacement string. Only the non-modified
|
|
version {} is supported:
|
|
|
|
parallel --header : echo f1={f1} f2={f2} ::: f1 A B ::: f2 C D
|
|
|
|
Output (the order may be different):
|
|
|
|
f1=A f2=C
|
|
f1=A f2=D
|
|
f1=B f2=C
|
|
f1=B f2=D
|
|
|
|
It is useful with --colsep for processing files with TAB separated values:
|
|
|
|
parallel --header : --colsep '\t' echo f1={f1} f2={f2} :::: tsv-file.tsv
|
|
|
|
Output (the order may be different):
|
|
|
|
f1=A f2=B
|
|
f1=C f2=D
|
|
|
|
=head2 More than one argument
|
|
|
|
With --xargs will GNU Parallel fit as many arguments as possible on a
|
|
single line:
|
|
|
|
cat num30000 | parallel --xargs echo | wc -l
|
|
|
|
Output (if you run this under Bash on GNU/Linux):
|
|
|
|
2
|
|
|
|
The 30000 arguments fitted on 2 lines.
|
|
|
|
The maximal length of a single line can be set with -s. With a maximal
|
|
line length of 10000 chars 17 commands will be run:
|
|
|
|
cat num30000 | parallel --xargs -s 10000 echo | wc -l
|
|
|
|
Output:
|
|
|
|
17
|
|
|
|
For better parallelism GNU Parallel can distribute the arguments
|
|
between all the parallel jobs when end of file is met.
|
|
|
|
Below GNU Parallel reads the last argument when generating the second
|
|
job. When GNU Parallel reads the last argument, it spreads all the
|
|
arguments for the second job over 4 jobs instead, as 4 parallel jobs
|
|
are requested.
|
|
|
|
The first job will be the same as the --xargs example above, but the
|
|
second job will be split into 4 evenly sized jobs, resulting in a
|
|
total of 5 jobs:
|
|
|
|
cat num30000 | parallel --jobs 4 -m echo | wc -l
|
|
|
|
Output (if you run this under Bash on GNU/Linux):
|
|
|
|
5
|
|
|
|
This is even more visible when running 4 jobs with 10 arguments. The
|
|
10 arguments are being spread over 4 jobs:
|
|
|
|
parallel --jobs 4 -m echo ::: 1 2 3 4 5 6 7 8 9 10
|
|
|
|
Output:
|
|
|
|
1 2 3
|
|
4 5 6
|
|
7 8 9
|
|
10
|
|
|
|
A replacement string can be part of a word. -m will not repeat the context:
|
|
|
|
parallel --jobs 4 -m echo pre-{}-post ::: A B C D E F G
|
|
|
|
Output (the order may be different):
|
|
|
|
pre-A B-post
|
|
pre-C D-post
|
|
pre-E F-post
|
|
pre-G-post
|
|
|
|
To repeat the context use -X which otherwise works like -m:
|
|
|
|
parallel --jobs 4 -X echo pre-{}-post ::: A B C D E F G
|
|
|
|
Output (the order may be different):
|
|
|
|
pre-A-post pre-B-post
|
|
pre-C-post pre-D-post
|
|
pre-E-post pre-F-post
|
|
pre-G-post
|
|
|
|
To limit the number of arguments use -N:
|
|
|
|
parallel -N3 echo ::: A B C D E F G H
|
|
|
|
Output (the order may be different):
|
|
|
|
A B C
|
|
D E F
|
|
G H
|
|
|
|
-N also sets the positional replacement strings:
|
|
|
|
parallel -N3 echo 1={1} 2={2} 3={3} ::: A B C D E F G H
|
|
|
|
Output (the order may be different):
|
|
|
|
1=A 2=B 3=C
|
|
1=D 2=E 3=F
|
|
1=G 2=H 3=
|
|
|
|
-N0 reads 1 argument but inserts none:
|
|
|
|
parallel -N0 echo foo ::: 1 2 3
|
|
|
|
Output:
|
|
|
|
foo
|
|
foo
|
|
foo
|
|
|
|
=head2 Quoting
|
|
|
|
Command lines that contain special characters may need to be protected from the shell.
|
|
|
|
The perl program 'print "@ARGV\n"' basically works like echo.
|
|
|
|
perl -e 'print "@ARGV\n"' A
|
|
|
|
Output:
|
|
|
|
A
|
|
|
|
To run that in parallel the command needs to be quoted:
|
|
|
|
parallel perl -e 'print "@ARGV\n"' ::: This wont work
|
|
|
|
Output:
|
|
|
|
[Nothing]
|
|
|
|
To quote the command use -q:
|
|
|
|
parallel -q perl -e 'print "@ARGV\n"' ::: This works
|
|
|
|
Output (the order may be different):
|
|
|
|
This
|
|
works
|
|
|
|
Or you can quote the critical part using \':
|
|
|
|
parallel perl -e \''print "@ARGV\n"'\' ::: This works, too
|
|
|
|
Output (the order may be different):
|
|
|
|
This
|
|
works,
|
|
too
|
|
|
|
GNU Parallel can also \-quote full lines. Simply run:
|
|
|
|
parallel --shellquote
|
|
parallel: Warning: Input is read from the terminal. Only experts do this on purpose. Press CTRL-D to exit.
|
|
perl -e 'print "@ARGV\n"'
|
|
[CTRL-D]
|
|
|
|
Output:
|
|
|
|
perl\ -e\ \'print\ \"@ARGV\\n\"\'
|
|
|
|
This can then be used as the command:
|
|
|
|
parallel perl\ -e\ \'print\ \"@ARGV\\n\"\' ::: This also works
|
|
|
|
Output (the order may be different):
|
|
|
|
This
|
|
also
|
|
works
|
|
|
|
|
|
=head2 Trimming space
|
|
|
|
Space can be trimmed on the arguments using --trim:
|
|
|
|
parallel --trim r echo pre-{}-post ::: ' A '
|
|
|
|
Output:
|
|
|
|
pre- A-post
|
|
|
|
To trim on the left side:
|
|
|
|
parallel --trim l echo pre-{}-post ::: ' A '
|
|
|
|
Output:
|
|
|
|
pre-A -post
|
|
|
|
To trim on the both sides:
|
|
|
|
parallel --trim lr echo pre-{}-post ::: ' A '
|
|
|
|
Output:
|
|
|
|
pre-A-post
|
|
|
|
|
|
=head1 Controlling the output
|
|
|
|
The output can prefixed with the argument:
|
|
|
|
parallel --tag echo foo-{} ::: A B C
|
|
|
|
Output (the order may be different):
|
|
|
|
A foo-A
|
|
B foo-B
|
|
C foo-C
|
|
|
|
To prefix it with another string use --tagstring:
|
|
|
|
parallel --tagstring {}-bar echo foo-{} ::: A B C
|
|
|
|
Output (the order may be different):
|
|
|
|
A-bar foo-A
|
|
B-bar foo-B
|
|
C-bar foo-C
|
|
|
|
To see what commands will be run without running them:
|
|
|
|
parallel --dryrun echo {} ::: A B C
|
|
|
|
Output (the order may be different):
|
|
|
|
echo A
|
|
echo B
|
|
echo C
|
|
|
|
To print the command before running them use --verbose:
|
|
|
|
parallel --verbose echo {} ::: A B C
|
|
|
|
Output (the order may be different):
|
|
|
|
echo A
|
|
echo B
|
|
A
|
|
echo C
|
|
B
|
|
C
|
|
|
|
GNU Parallel will postpone the output until the command completes:
|
|
|
|
parallel -j2 'printf "%s-start\n%s" {} {};sleep {};printf "%s\n" -middle;echo {}-end' ::: 4 2 1
|
|
|
|
Output:
|
|
|
|
2-start
|
|
2-middle
|
|
2-end
|
|
1-start
|
|
1-middle
|
|
1-end
|
|
4-start
|
|
4-middle
|
|
4-end
|
|
|
|
To get the output immediately use --ungroup:
|
|
|
|
parallel -j2 --ungroup 'printf "%s-start\n%s" {} {};sleep {};printf "%s\n" -middle;echo {}-end' ::: 4 2 1
|
|
|
|
Output:
|
|
|
|
4-start
|
|
42-start
|
|
2-middle
|
|
2-end
|
|
1-start
|
|
1-middle
|
|
1-end
|
|
-middle
|
|
4-end
|
|
|
|
--ungroup is fast, but can cause half a line from one job to be mixed
|
|
with half a line of another job. That has happend in the second line,
|
|
where the line '4-middle' is mixed with '2-start'.
|
|
|
|
To avoid this use --linebuffer:
|
|
|
|
parallel -j2 --linebuffer 'printf "%s-start\n%s" {} {};sleep {};printf "%s\n" -middle;echo {}-end' ::: 4 2 1
|
|
|
|
Output:
|
|
|
|
4-start
|
|
2-start
|
|
2-middle
|
|
2-end
|
|
1-start
|
|
1-middle
|
|
1-end
|
|
4-middle
|
|
4-end
|
|
|
|
To force the output in the same order as the arguments use --keep-order/-k:
|
|
|
|
parallel -j2 -k 'printf "%s-start\n%s" {} {};sleep {};printf "%s\n" -middle;echo {}-end' ::: 4 2 1
|
|
|
|
Output:
|
|
|
|
4-start
|
|
4-middle
|
|
4-end
|
|
2-start
|
|
2-middle
|
|
2-end
|
|
1-start
|
|
1-middle
|
|
1-end
|
|
|
|
=head2 Saving output into files
|
|
|
|
GNU Parallel can save the output of each job into files:
|
|
|
|
parallel --files echo ::: A B C
|
|
|
|
Output will be similar to:
|
|
|
|
/tmp/pAh6uWuQCg.par
|
|
/tmp/opjhZCzAX4.par
|
|
/tmp/W0AT_Rph2o.par
|
|
|
|
By default GNU Parallel will cache the output in files in /tmp. This
|
|
can be changed by setting $TMPDIR or --tmpdir:
|
|
|
|
parallel --tmpdir /var/tmp --files echo ::: A B C
|
|
|
|
Output will be similar to:
|
|
|
|
/var/tmp/N_vk7phQRc.par
|
|
/var/tmp/7zA4Ccf3wZ.par
|
|
/var/tmp/LIuKgF_2LP.par
|
|
|
|
Or:
|
|
|
|
TMPDIR=/var/tmp parallel --files echo ::: A B C
|
|
|
|
Output: Same as above.
|
|
|
|
The output files can be saved in a structured way using --results:
|
|
|
|
parallel --results outdir echo ::: A B C
|
|
|
|
Output:
|
|
|
|
A
|
|
B
|
|
C
|
|
|
|
but also these files were generated containing the standard output
|
|
(stdout) and standard error (stderr):
|
|
|
|
outdir/1/A/stderr
|
|
outdir/1/A/stdout
|
|
outdir/1/B/stderr
|
|
outdir/1/B/stdout
|
|
outdir/1/C/stderr
|
|
outdir/1/C/stdout
|
|
|
|
This is useful if you are running multiple variables:
|
|
|
|
parallel --header : --results outdir echo ::: f1 A B ::: f2 C D
|
|
|
|
Generated files:
|
|
|
|
outdir/f1/A/f2/C/stderr
|
|
outdir/f1/A/f2/C/stdout
|
|
outdir/f1/A/f2/D/stderr
|
|
outdir/f1/A/f2/D/stdout
|
|
outdir/f1/B/f2/C/stderr
|
|
outdir/f1/B/f2/C/stdout
|
|
outdir/f1/B/f2/D/stderr
|
|
outdir/f1/B/f2/D/stdout
|
|
|
|
The directories are named after the variables and their values.
|
|
|
|
=head1 Control the execution
|
|
|
|
=head2 Number of simultaneous jobs
|
|
|
|
The number of concurrent jobs is given with --jobs/-j:
|
|
|
|
/usr/bin/time parallel -N0 -j64 sleep 1 :::: num128
|
|
|
|
With 64 jobs in parallel the 128 sleeps will take 2-8 seconds to run -
|
|
depending on how fast your machine is.
|
|
|
|
By default --jobs is the same as the number of CPU cores. So this:
|
|
|
|
/usr/bin/time parallel -N0 sleep 1 :::: num128
|
|
|
|
should take twice the time of running 2 jobs per CPU core:
|
|
|
|
/usr/bin/time parallel -N0 --jobs 200% sleep 1 :::: num128
|
|
|
|
--jobs 0 will run as many jobs in parallel as possible:
|
|
|
|
/usr/bin/time parallel -N0 --jobs 0 sleep 1 :::: num128
|
|
|
|
which should take 1-7 seconds depending on how fast your machine is.
|
|
|
|
--jobs can read from a file which is re-read when a job finishes:
|
|
|
|
echo 50% > my_jobs
|
|
/usr/bin/time parallel -N0 --jobs my_jobs sleep 1 :::: num128 &
|
|
sleep 1
|
|
echo 0 > my_jobs
|
|
wait
|
|
|
|
The first second only 50% of the CPU cores will run a job. The '0' is
|
|
put into my_jobs and then the rest of the jobs will be started in
|
|
parallel.
|
|
|
|
Instead of basing the percentage on the number of CPU cores
|
|
GNU Parallel can base it on the number of CPUs:
|
|
|
|
parallel --use-cpus-instead-of-cores -N0 sleep 1 :::: num8
|
|
|
|
=head2 Shuffle job order
|
|
|
|
If you have many jobs (e.g. by multiple combinations of input
|
|
sources), it can be handy to shuffle the jobs, so you get different
|
|
values run.
|
|
|
|
parallel --shuf echo ::: 1 2 3 ::: a b c ::: A B C
|
|
|
|
Output:
|
|
|
|
All combinations but different order for each run.
|
|
|
|
=head2 Interactivity
|
|
|
|
GNU Parallel can ask the user if a command should be run using --interactive:
|
|
|
|
parallel --interactive echo ::: 1 2 3
|
|
|
|
Output:
|
|
|
|
echo 1 ?...y
|
|
echo 2 ?...n
|
|
1
|
|
echo 3 ?...y
|
|
3
|
|
|
|
GNU Parallel can be used to put arguments on the command line for an
|
|
interactive command such as emacs to edit one file at a time:
|
|
|
|
parallel --tty emacs ::: 1 2 3
|
|
|
|
Or give multiple argument in one go to open multiple files:
|
|
|
|
parallel -X --tty vi ::: 1 2 3
|
|
|
|
=head2 A terminal for every job
|
|
|
|
Using tmux GNU Parallel can start a terminal for every job run:
|
|
|
|
seq 10 20 | parallel --tmux 'echo start {}; sleep {}; echo done {}'
|
|
|
|
This will tell you to run something similar to:
|
|
|
|
tmux -S /tmp/tmsrPrO0 attach
|
|
|
|
Using normal tmux keystrokes (CTRL-b n or CTRL-b p) you can cycle
|
|
between windows of the running jobs. When a job is finished it will
|
|
pause for 10 seconds before closing the window.
|
|
|
|
=head2 Timing
|
|
|
|
Some jobs do heavy I/O when they start. To avoid a thundering herd GNU
|
|
Parallel can delay starting new jobs. --delay X will make sure there is
|
|
at least X seconds between each start:
|
|
|
|
parallel --delay 2.5 echo Starting {}\;date ::: 1 2 3
|
|
|
|
Output:
|
|
|
|
Starting 1
|
|
Thu Aug 15 16:24:33 CEST 2013
|
|
Starting 2
|
|
Thu Aug 15 16:24:35 CEST 2013
|
|
Starting 3
|
|
Thu Aug 15 16:24:38 CEST 2013
|
|
|
|
|
|
If jobs taking more than a certain amount of time are known to fail,
|
|
they can be stopped with --timeout. The accuracy of --timeout is 2
|
|
seconds:
|
|
|
|
parallel --timeout 4.1 sleep {}\; echo {} ::: 2 4 6 8
|
|
|
|
Output:
|
|
|
|
2
|
|
4
|
|
|
|
GNU Parallel can compute the median runtime for jobs and kill those
|
|
that take more than 200% of the median runtime:
|
|
|
|
parallel --timeout 200% sleep {}\; echo {} ::: 2.1 2.2 3 7 2.3
|
|
|
|
Output:
|
|
|
|
2.1
|
|
2.2
|
|
3
|
|
2.3
|
|
|
|
Based on the runtime of completed jobs GNU Parallel can estimate the
|
|
total runtime:
|
|
|
|
parallel --eta sleep ::: 1 3 2 2 1 3 3 2 1
|
|
|
|
Output:
|
|
|
|
Computers / CPU cores / Max jobs to run
|
|
1:local / 2 / 2
|
|
|
|
Computer:jobs running/jobs completed/%of started jobs/Average seconds to complete
|
|
ETA: 2s 0left 1.11avg local:0/9/100%/1.1s
|
|
|
|
=head2 Progress
|
|
|
|
GNU Parallel can give progress information with --progress:
|
|
|
|
parallel --progress sleep ::: 1 3 2 2 1 3 3 2 1
|
|
|
|
Output:
|
|
|
|
Computers / CPU cores / Max jobs to run
|
|
1:local / 2 / 2
|
|
|
|
Computer:jobs running/jobs completed/%of started jobs/Average seconds to complete
|
|
local:0/9/100%/1.1s
|
|
|
|
A progress bar can be shown with --bar:
|
|
|
|
parallel --bar sleep ::: 1 3 2 2 1 3 3 2 1
|
|
|
|
And a graphic bar can be shown with --bar and zenity:
|
|
|
|
seq 1000 | parallel -j10 --bar '(echo -n {};sleep 0.1)' 2> >(zenity --progress --auto-kill)
|
|
|
|
A logfile of the jobs completed so far can be generated with --joblog:
|
|
|
|
parallel --joblog /tmp/log exit ::: 1 2 3 0
|
|
cat /tmp/log
|
|
|
|
Output:
|
|
|
|
Seq Host Starttime Runtime Send Receive Exitval Signal Command
|
|
1 : 1376577364.974 0.008 0 0 1 0 exit 1
|
|
2 : 1376577364.982 0.013 0 0 2 0 exit 2
|
|
3 : 1376577364.990 0.013 0 0 3 0 exit 3
|
|
4 : 1376577365.003 0.003 0 0 0 0 exit 0
|
|
|
|
The log contains the job sequence, which host the job was run on, the
|
|
start time and run time, how much data was transferred if the job was
|
|
run on a remote host, the exit value, the signal that killed the job,
|
|
and finally the command being run.
|
|
|
|
With a joblog GNU Parallel can be stopped and later pickup where it
|
|
left off. It it important that the input of the completed jobs is
|
|
unchanged.
|
|
|
|
parallel --joblog /tmp/log exit ::: 1 2 3 0
|
|
cat /tmp/log
|
|
parallel --resume --joblog /tmp/log exit ::: 1 2 3 0 0 0
|
|
cat /tmp/log
|
|
|
|
Output:
|
|
|
|
Seq Host Starttime Runtime Send Receive Exitval Signal Command
|
|
1 : 1376580069.544 0.008 0 0 1 0 exit 1
|
|
2 : 1376580069.552 0.009 0 0 2 0 exit 2
|
|
3 : 1376580069.560 0.012 0 0 3 0 exit 3
|
|
4 : 1376580069.571 0.005 0 0 0 0 exit 0
|
|
|
|
Seq Host Starttime Runtime Send Receive Exitval Signal Command
|
|
1 : 1376580069.544 0.008 0 0 1 0 exit 1
|
|
2 : 1376580069.552 0.009 0 0 2 0 exit 2
|
|
3 : 1376580069.560 0.012 0 0 3 0 exit 3
|
|
4 : 1376580069.571 0.005 0 0 0 0 exit 0
|
|
5 : 1376580070.028 0.009 0 0 0 0 exit 0
|
|
6 : 1376580070.038 0.007 0 0 0 0 exit 0
|
|
|
|
Note how the start time of the last 2 jobs is clearly from the second run.
|
|
|
|
With --resume-failed GNU Parallel will re-run the jobs that failed:
|
|
|
|
parallel --resume-failed --joblog /tmp/log exit ::: 1 2 3 0 0 0
|
|
cat /tmp/log
|
|
|
|
Output:
|
|
|
|
Seq Host Starttime Runtime Send Receive Exitval Signal Command
|
|
1 : 1376580069.544 0.008 0 0 1 0 exit 1
|
|
2 : 1376580069.552 0.009 0 0 2 0 exit 2
|
|
3 : 1376580069.560 0.012 0 0 3 0 exit 3
|
|
4 : 1376580069.571 0.005 0 0 0 0 exit 0
|
|
5 : 1376580070.028 0.009 0 0 0 0 exit 0
|
|
6 : 1376580070.038 0.007 0 0 0 0 exit 0
|
|
1 : 1376580154.433 0.010 0 0 1 0 exit 1
|
|
2 : 1376580154.444 0.022 0 0 2 0 exit 2
|
|
3 : 1376580154.466 0.005 0 0 3 0 exit 3
|
|
|
|
Note how seq 1 2 3 have been repeated because they had exit value != 0.
|
|
|
|
=head2 Termination
|
|
|
|
For certain jobs there is no need to continue if one of the jobs fails
|
|
and has an exit code != 0. GNU Parallel will stop spawning new jobs
|
|
with --halt soon,fail=1:
|
|
|
|
parallel -j2 --halt soon,fail=1 echo {}\; exit {} ::: 0 0 1 2 3
|
|
|
|
Output:
|
|
|
|
0
|
|
0
|
|
1
|
|
parallel: Starting no more jobs. Waiting for 2 jobs to finish. This job failed:
|
|
echo 1; exit 1
|
|
2
|
|
parallel: Starting no more jobs. Waiting for 1 jobs to finish. This job failed:
|
|
echo 2; exit 2
|
|
|
|
With --halt now,fail=1 the running jobs will be killed immediately:
|
|
|
|
parallel -j2 --halt now,fail=1 echo {}\; exit {} ::: 0 0 1 2 3
|
|
|
|
Output:
|
|
|
|
0
|
|
0
|
|
1
|
|
parallel: This job failed:
|
|
echo 1; exit 1
|
|
|
|
If --halt is given a percentage this percentage of the jobs must fail
|
|
before GNU Parallel stops spawning more jobs:
|
|
|
|
parallel -j2 --halt soon,fail=20% echo {}\; exit {} ::: 0 1 2 3 4 5 6 7 8 9
|
|
|
|
Output:
|
|
|
|
0
|
|
1
|
|
parallel: This job failed:
|
|
echo 1; exit 1
|
|
2
|
|
parallel: This job failed:
|
|
echo 2; exit 2
|
|
parallel: Starting no more jobs. Waiting for 1 jobs to finish.
|
|
3
|
|
parallel: This job failed:
|
|
echo 3; exit 3
|
|
|
|
If you are looking for success instead of failures, you can use success:
|
|
|
|
parallel -j2 --halt soon,success=1 echo {}\; exit {} ::: 1 2 3 0 4 5 6
|
|
|
|
Output:
|
|
|
|
1
|
|
2
|
|
3
|
|
0
|
|
parallel: This job succeeded:
|
|
echo 0; exit 0
|
|
parallel: Starting no more jobs. Waiting for 1 jobs to finish.
|
|
4
|
|
|
|
GNU Parallel can retry the command with --retries. This is useful if a
|
|
command fails for unknown reasons now and then.
|
|
|
|
parallel -k --retries 3 'echo tried {} >>/tmp/runs; echo completed {}; exit {}' ::: 1 2 0
|
|
cat /tmp/runs
|
|
|
|
Output:
|
|
|
|
completed 1
|
|
completed 2
|
|
completed 0
|
|
|
|
tried 1
|
|
tried 2
|
|
tried 1
|
|
tried 2
|
|
tried 1
|
|
tried 2
|
|
tried 0
|
|
|
|
Note how job 1 and 2 were tried 3 times, but 0 was not retried because it had exit code 0.
|
|
|
|
=head2 Limiting the resources
|
|
|
|
To avoid overloading systems GNU Parallel can look at the system load
|
|
before starting another job:
|
|
|
|
parallel --load 100% echo load is less than {} job per cpu ::: 1
|
|
|
|
Output:
|
|
|
|
[when then load is less than the number of cpu cores]
|
|
load is less than 1 job per cpu
|
|
|
|
GNU Parallel can also check if the system is swapping.
|
|
|
|
parallel --noswap echo the system is not swapping ::: now
|
|
|
|
Output:
|
|
|
|
[when then system is not swapping]
|
|
the system is not swapping now
|
|
|
|
GNU Parallel can run the jobs with a nice value. This will work both
|
|
locally and remotely.
|
|
|
|
parallel --nice 17 echo this is being run with nice -n ::: 17
|
|
|
|
Output:
|
|
|
|
this is being run with nice -n 17
|
|
|
|
=head1 Remote execution
|
|
|
|
GNU Parallel can run jobs on remote servers. It uses ssh to
|
|
communicate with the remote machines.
|
|
|
|
=head2 Sshlogin
|
|
|
|
The most basic sshlogin is -S host:
|
|
|
|
parallel -S $SERVER1 echo running on ::: $SERVER1
|
|
|
|
Output:
|
|
|
|
running on [$SERVER1]
|
|
|
|
To use a different username prepend the server with username@
|
|
|
|
parallel -S username@$SERVER1 echo running on ::: username@$SERVER1
|
|
|
|
Output:
|
|
|
|
running on [username@$SERVER1]
|
|
|
|
The special sshlogin ':' is the local machine:
|
|
|
|
parallel -S : echo running on ::: the_local_machine
|
|
|
|
Output:
|
|
|
|
running on the_local_machine
|
|
|
|
If ssh is not in $PATH it can be prepended to $SERVER1:
|
|
|
|
parallel -S '/usr/bin/ssh '$SERVER1 echo custom ::: ssh
|
|
|
|
Output:
|
|
|
|
custom ssh
|
|
|
|
Several servers can be given using multiple -S:
|
|
|
|
parallel -S $SERVER1 -S $SERVER2 echo ::: running on more hosts
|
|
|
|
Output (the order may be different):
|
|
|
|
running
|
|
on
|
|
more
|
|
hosts
|
|
|
|
Or they can be separated by ,:
|
|
|
|
parallel -S $SERVER1,$SERVER2 echo ::: running on more hosts
|
|
|
|
Output: Same as above.
|
|
|
|
Or newline:
|
|
|
|
# This gives a \n between $SERVER1 and $SERVER2
|
|
SERVERS="`echo $SERVER1; echo $SERVER2`"
|
|
parallel -S "$SERVERS" echo ::: running on more hosts
|
|
|
|
The can also be read from a file (replace user@ with the user on $SERVER2):
|
|
|
|
echo $SERVER1 > nodefile
|
|
# Force 4 cores, special ssh-command, username
|
|
echo 4//usr/bin/ssh user@$SERVER2 >> nodefile
|
|
parallel --sshloginfile nodefile echo ::: running on more hosts
|
|
|
|
Output: Same as above.
|
|
|
|
The special --sshloginfile '..' reads from ~/.parallel/sshloginfile.
|
|
|
|
To force GNU Parallel to treat a server having a given number of CPU
|
|
cores prepend #/ to the sshlogin:
|
|
|
|
parallel -S 4/$SERVER1 echo force {} cpus on server ::: 4
|
|
|
|
Output:
|
|
|
|
force 4 cpus on server
|
|
|
|
Servers can be put into groups by prepending '@groupname' to the
|
|
server and the group can then be selected by appending '@groupname' to
|
|
the argument if using '--hostgrp'.
|
|
|
|
parallel --hostgrp -S @grp1/$SERVER1 -S @grp2/SERVER2 echo {} ::: run_on_grp1@grp1 run_on_grp2@grp2
|
|
|
|
Output:
|
|
|
|
run_on_grp1
|
|
run_on_grp2
|
|
|
|
A host can be in multiple groups by separating groups with '+', and
|
|
you can force GNU B<parallel> to limit the groups on which the command
|
|
can be run with '-S @groupname':
|
|
|
|
parallel -S @grp1 -S @grp1+grp2/$SERVER1 -S @grp2/SERVER2 echo {} ::: run_on_grp1 also_grp1
|
|
|
|
Output:
|
|
|
|
run_on_grp1
|
|
also_grp1
|
|
|
|
=head2 Transferring files
|
|
|
|
GNU Parallel can transfer the files to be processed to the remote
|
|
host. It does that using rsync.
|
|
|
|
echo This is input_file > input_file
|
|
parallel -S $SERVER1 --transfer cat ::: input_file
|
|
|
|
Output:
|
|
|
|
This is input_file
|
|
|
|
If the files is processed into another file, the resulting file can be
|
|
transferred back:
|
|
|
|
echo This is input_file > input_file
|
|
parallel -S $SERVER1 --transfer --return {}.out cat {} ">"{}.out ::: input_file
|
|
cat input_file.out
|
|
|
|
Output: Same as above.
|
|
|
|
To remove the input and output file on the remote server use --cleanup:
|
|
|
|
echo This is input_file > input_file
|
|
parallel -S $SERVER1 --transfer --return {}.out --cleanup cat {} ">"{}.out ::: input_file
|
|
cat input_file.out
|
|
|
|
Output: Same as above.
|
|
|
|
There is a short hand for --transfer --return --cleanup called --trc:
|
|
|
|
echo This is input_file > input_file
|
|
parallel -S $SERVER1 --trc {}.out cat {} ">"{}.out ::: input_file
|
|
cat input_file.out
|
|
|
|
Output: Same as above.
|
|
|
|
Some jobs need a common database for all jobs. GNU Parallel can
|
|
transfer that using --basefile which will transfer the file before the
|
|
first job:
|
|
|
|
echo common data > common_file
|
|
parallel --basefile common_file -S $SERVER1 cat common_file\; echo {} ::: foo
|
|
|
|
Output:
|
|
|
|
common data
|
|
foo
|
|
|
|
To remove it from the remote host after the last job use --cleanup.
|
|
|
|
|
|
=head2 Working dir
|
|
|
|
The default working dir on the remote machines is the login dir. This
|
|
can be changed with --workdir I<mydir>.
|
|
|
|
Files transferred using --transfer and --return will be relative
|
|
to I<mydir> on remote computers, and the command will be executed in
|
|
the dir I<mydir>.
|
|
|
|
The special I<mydir> value ... will create working dirs under
|
|
~/.parallel/tmp/ on the remote computers. If --cleanup is given
|
|
these dirs will be removed.
|
|
|
|
The special I<mydir> value . uses the current working dir. If the
|
|
current working dir is beneath your home dir, the value . is
|
|
treated as the relative path to your home dir. This means that if your
|
|
home dir is different on remote computers (e.g. if your login is
|
|
different) the relative path will still be relative to your home dir.
|
|
|
|
parallel -S $SERVER1 pwd ::: ""
|
|
parallel --workdir . -S $SERVER1 pwd ::: ""
|
|
parallel --workdir ... -S $SERVER1 pwd ::: ""
|
|
|
|
Output:
|
|
|
|
[the login dir on $SERVER1]
|
|
[current dir relative on $SERVER1]
|
|
[a dir in ~/.parallel/tmp/...]
|
|
|
|
|
|
=head2 Avoid overloading sshd
|
|
|
|
If many jobs are started on the same server, sshd can be
|
|
overloaded. GNU Parallel can insert a delay between each job run on
|
|
the same server:
|
|
|
|
parallel -S $SERVER1 --sshdelay 0.2 echo ::: 1 2 3
|
|
|
|
Output (the order may be different):
|
|
|
|
1
|
|
2
|
|
3
|
|
|
|
Sshd will be less overloaded if using --controlmaster, which will
|
|
multiplex ssh connections:
|
|
|
|
parallel --controlmaster -S $SERVER1 echo ::: 1 2 3
|
|
|
|
Output: Same as above.
|
|
|
|
=head2 Ignore hosts that are down
|
|
|
|
In clusters with many hosts a few of the are often down. GNU Parallel
|
|
can ignore those hosts. In this case the host 173.194.32.46 is down:
|
|
|
|
parallel --filter-hosts -S 173.194.32.46,$SERVER1 echo ::: bar
|
|
|
|
Output:
|
|
|
|
bar
|
|
|
|
=head2 Running the same commands on all hosts
|
|
|
|
GNU Parallel can run the same command on all the hosts:
|
|
|
|
parallel --onall -S $SERVER1,$SERVER2 echo ::: foo bar
|
|
|
|
Output (the order may be different):
|
|
|
|
foo
|
|
bar
|
|
foo
|
|
bar
|
|
|
|
Often you will just want to run a single command on all hosts with out
|
|
arguments. --nonall is a no argument --onall:
|
|
|
|
parallel --nonall -S $SERVER1,$SERVER2 echo foo bar
|
|
|
|
Output:
|
|
|
|
foo bar
|
|
foo bar
|
|
|
|
When --tag is used with --nonall and --onall the --tagstring is the host:
|
|
|
|
parallel --nonall --tag -S $SERVER1,$SERVER2 echo foo bar
|
|
|
|
Output (the order may be different):
|
|
|
|
$SERVER1 foo bar
|
|
$SERVER2 foo bar
|
|
|
|
--jobs sets the number of servers to log in to in parallel.
|
|
|
|
=head2 Transfer environment variables and functions
|
|
|
|
Using --env GNU Parallel can transfer an environment variable to the
|
|
remote system.
|
|
|
|
MYVAR='foo bar'
|
|
export MYVAR
|
|
parallel --env MYVAR -S $SERVER1 echo '$MYVAR' ::: baz
|
|
|
|
Output:
|
|
|
|
foo bar baz
|
|
|
|
This works for functions too if your shell is Bash:
|
|
|
|
# This only works in Bash
|
|
my_func() {
|
|
echo in my_func $1
|
|
}
|
|
export -f my_func
|
|
parallel --env my_func -S $SERVER1 my_func ::: baz
|
|
|
|
Output:
|
|
|
|
in my_func baz
|
|
|
|
GNU Parallel can copy all defined variables and functions to the
|
|
remote system. It just needs to record which ones to ignore in
|
|
~/.parallel/ignored_vars. Do that by running this once:
|
|
|
|
parallel --record-env
|
|
cat ~/.parallel/ignored_vars
|
|
|
|
Output:
|
|
|
|
[list of variables to ignore - including $PATH and $HOME]
|
|
|
|
Now all new variables and functions defined will be copied when using
|
|
--env _:
|
|
|
|
# The function is only copied if using Bash
|
|
my_func2() {
|
|
echo in my_func2 $VAR $1
|
|
}
|
|
export -f my_func2
|
|
VAR=foo
|
|
export VAR
|
|
|
|
parallel --env _ -S $SERVER1 'echo $VAR; my_func2' ::: bar
|
|
|
|
Output:
|
|
|
|
foo
|
|
in my_func2 foo bar
|
|
|
|
=head2 Showing what is actually run
|
|
|
|
--verbose will show the command that would be run on the local
|
|
machine. When a job is run on a remote machine this is wrapped with
|
|
ssh and possibly transferring files and environment variables, setting
|
|
the workdir, and setting --nice value. -vv shows all of this.
|
|
|
|
parallel -vv -S $SERVER1 echo ::: bar
|
|
|
|
Output:
|
|
|
|
ssh lo exec perl\ -e\ \\\$ENV\\\{\\\"PARALLEL_PID\\\"\\\}=\\\"2554030\\\"\\\;\
|
|
\\$ENV\\\{\\\"PARALLEL_SEQ\\\"\\\}=\\\"1\\\"\\\;\\\$bashfunc\\\ =\\\ \\\"\\\"\
|
|
\\;@ARGV=\\\"echo\\\ bar\\\"\\\;\\\$SIG\\\{CHLD\\\}=sub\\\{\\\$done=1\\\;\\\}\
|
|
\\;\\\$pid=fork\\\;unless\\\(\\\$pid\\\)\\\{setpgrp\\\;exec\\\$ENV\\\{SHELL\\\
|
|
},\\\"-c\\\",\\\(\\\$bashfunc.\\\"@ARGV\\\"\\\)\\\;die\\\"exec:\\\$\\\!\\\\n\\
|
|
\"\\\;\\\}do\\\{\\\$s=\\\$s\\\<1\\\?0.001+\\\$s\\\*1.03:\\\$s\\\;select\\\(
|
|
undef,undef,undef,\\\$s\\\)\\\;\\\}until\\\(\\\$done\\\|\\\|getppid==1\\\)\\\;
|
|
kill\\\(SIGHUP,-\\\$\\\{pid\\\}\\\)unless\\\$done\\\;wait\\\;exit\\\(\\\$\\\?\
|
|
\\&127\\\?128+\\\(\\\$\\\?\\\&127\\\):1+\\\$\\\?\\\>\\\>8\\\);
|
|
bar
|
|
|
|
When the command gets more complex, the output is so hard to read, that it is only useful for debugging:
|
|
|
|
my_func3() {
|
|
echo in my_func $1 > $1.out
|
|
}
|
|
export -f my_func3
|
|
parallel -vv --workdir ... --nice 17 --env _ --trc {}.out -S $SERVER1 my_func3 {} ::: abc-file
|
|
|
|
Output will be similar to:
|
|
|
|
( ssh lo mkdir -p ./.parallel/tmp/aspire-2554425-1;rsync --protocol 30 -rlDzR
|
|
-essh ./abc-file lo:./.parallel/tmp/aspire-2554425-1 );ssh lo exec perl -e \''
|
|
@GNU_Parallel=("use","IPC::Open3;","use","MIME::Base64");eval"@GNU_Parallel";
|
|
$SIG{CHLD}="IGNORE";my$zip=(grep{-x$_}"/usr/local/bin/bzip2")[0]||"bzip2";my(
|
|
$in,$out,$eval);open3($in,$out,">&STDERR",$zip,"-dc");if(my$perlpid=fork){
|
|
close$in;$eval=join"",<$out>;close$out;}else{close$out;print$in(decode_base64(
|
|
join"",@ARGV));close$in;exit;}wait;eval$eval;'\' QlpoOTFBWSZTWayP388AAbdfgAAQd
|
|
X/+3//l/wS/7//vQAIq3U6bauIpk1NPUwnqGgND1NGI9TTQ0A0ADIDQNITU9NGqfonpPJRvU0ZQZAA
|
|
GhpoBoABpqaFNBMh+kGoybUaHogwCYBAADCSgRNqek1TxTeknqemhGyBqek8pk2jRPSeo002mQRo5f
|
|
oSZYrgzQFDd3HNWaskbx+MxNR89BdDzESFbADOJkI+QhIlnojHCWRVuGc2j2lzMzE41wC7auAMQ06c
|
|
S3AlqQfKcdo0gd506U0HzAAxMkGJBHjDCZULOMpVbowhIVxxaQz7yansTsBgurEZaGO/6K0Nc4iodr
|
|
BW4m9SXErqRbLNy5eANDvZ+TIt2c2GBcWSlmYuloxY5u2bGUdU/dGsO5EhyrvKCpZMhIgmQFAQhcwR
|
|
mD+jMKRawkRFJSGyTNC3PqWnE51ucPyx29Yxjnkyub98lytpyk+v8BUc4eA3xz98dMYjxvb0pgWksh
|
|
oHZ7HwGQRq1vuDyzKgkwPL9lwGIdL+WPNJFSljlVAahIhQpDCAOJpTqDhgmfoRQcy54PC9T0T3iMnV
|
|
JeTUdL8P0/s18NqDSUavMNV3qD0CtYi6entl0neNsOQN2VDSLHj0xOMls65LNPo+Wh28rJtVoh2JgE
|
|
7Q9Qo/XBr6krGIsYpQR6nRDuJCD/5aaQBBFFQGtv2VoFTwkXiUTxFP1CC4AGBznAaMklgWQvVtKguJ
|
|
zQnPqr9ABtHwbB5GTzPOQ4iWAmrUxvl4j5wqrVchOZcs3NYUQmGO2+VYBimFVxhGcaxDALMZ6bWEUo
|
|
yt8eC8W5o1ObFtTnHAvjOQgYEL/nHTcxU0G57QMKCzJcASQWFNpe2CpQcgYlBxIN4kwtfxdyRThQkK
|
|
yP388;_EXIT_status=$?; mkdir -p ./.; rsync --protocol 30 --rsync-path=cd\
|
|
./.parallel/tmp/aspire-2554425-1/./.\;\ rsync -rlDzR -essh lo:./abc-file.out
|
|
./.;ssh lo \(rm\ -f\ ./.parallel/tmp/aspire-2554425-1/abc-file\;\ sh\ -c\ \'
|
|
rmdir\ ./.parallel/tmp/aspire-2554425-1/\ ./.parallel/tmp/\ ./.parallel/\ 2\>
|
|
/dev/null\'\;rm\ -rf\ ./.parallel/tmp/aspire-2554425-1\;\);ssh lo \(rm\ -f\
|
|
./.parallel/tmp/aspire-2554425-1/abc-file.out\;\ sh\ -c\ \'rmdir\ ./.parallel
|
|
/tmp/aspire-2554425-1/\ ./.parallel/tmp/\ ./.parallel/\ 2\>/dev/null\'\;rm\
|
|
-rf\ ./.parallel/tmp/aspire-2554425-1\;\);ssh lo rm -rf .parallel/tmp/
|
|
aspire-2554425-1; exit $_EXIT_status;
|
|
|
|
=head1 --pipe
|
|
|
|
The --pipe functionality puts GNU Parallel in a different mode:
|
|
Instead of treating the data on stdin (standard input) as arguments
|
|
for a command to run, the data will be sent to stdin (standard input)
|
|
of the command.
|
|
|
|
The typical situation is:
|
|
|
|
command_A | command_B | command_C
|
|
|
|
where command_B is slow, and you want to speed up command_B.
|
|
|
|
=head2 Chunk size
|
|
|
|
By default GNU Parallel will start an instance of command_B, read a
|
|
chunk of 1 MB, and pass that to the instance. Then start another
|
|
instance, read another chunk, and pass that to the second instance.
|
|
|
|
cat num1000000 | parallel --pipe wc
|
|
|
|
Output (the order may be different):
|
|
|
|
165668 165668 1048571
|
|
149797 149797 1048579
|
|
149796 149796 1048572
|
|
149797 149797 1048579
|
|
149797 149797 1048579
|
|
149796 149796 1048572
|
|
85349 85349 597444
|
|
|
|
The size of the chunk is not exactly 1 MB because GNU Parallel only
|
|
passes full lines - never half a line, thus the blocksize is only
|
|
average 1 MB. You can change the block size to 2 MB with --block:
|
|
|
|
cat num1000000 | parallel --pipe --block 2M wc
|
|
|
|
Output (the order may be different):
|
|
|
|
315465 315465 2097150
|
|
299593 299593 2097151
|
|
299593 299593 2097151
|
|
85349 85349 597444
|
|
|
|
GNU Parallel treats each line as a record. If the order of record is
|
|
unimportant (e.g. you need all lines processed, but you do not care
|
|
which is processed first), then you can use --round-robin. Without
|
|
--round-robin GNU Parallel will start a command per block; with
|
|
--round-robin only the requested number of jobs will be started
|
|
(--jobs). The records will then be distributed between the running
|
|
jobs:
|
|
|
|
cat num1000000 | parallel --pipe -j4 --round-robin wc
|
|
|
|
Output will be similar to:
|
|
|
|
149797 149797 1048579
|
|
299593 299593 2097151
|
|
315465 315465 2097150
|
|
235145 235145 1646016
|
|
|
|
One of the 4 instances got a single record, 2 instances got 2 full
|
|
records each, and one instance got 1 full and 1 partial record.
|
|
|
|
=head2 Records
|
|
|
|
GNU Parallel sees the input as records. The default record is a single
|
|
line.
|
|
|
|
Using -N140000 GNU Parallel will read 140000 records at a time:
|
|
|
|
cat num1000000 | parallel --pipe -N140000 wc
|
|
|
|
Output (the order may be different):
|
|
|
|
140000 140000 868895
|
|
140000 140000 980000
|
|
140000 140000 980000
|
|
140000 140000 980000
|
|
140000 140000 980000
|
|
140000 140000 980000
|
|
140000 140000 980000
|
|
20000 20000 140001
|
|
|
|
Notice that the last job could not get the full 140000 lines, but only
|
|
20000 lines.
|
|
|
|
If a record is 75 lines -L can be used:
|
|
|
|
cat num1000000 | parallel --pipe -L75 wc
|
|
|
|
Output (the order may be different):
|
|
|
|
165600 165600 1048095
|
|
149850 149850 1048950
|
|
149775 149775 1048425
|
|
149775 149775 1048425
|
|
149850 149850 1048950
|
|
149775 149775 1048425
|
|
85350 85350 597450
|
|
25 25 176
|
|
|
|
Notice GNU Parallel still reads a block of around 1 MB; but instead of
|
|
passing full lines to 'wc' it passes full 75 lines at a time. This
|
|
of course does not hold for the last job (which in this case got 25
|
|
lines).
|
|
|
|
=head2 Record separators
|
|
|
|
GNU Parallel uses separators to determine where two record split.
|
|
|
|
--recstart gives the string that starts a record; --recend gives the
|
|
string that ends a record. The default is --recend '\n' (newline).
|
|
|
|
If both --recend and --recstart are given, then the record will only
|
|
split if the recend string is immediately followed by the recstart
|
|
string.
|
|
|
|
Here the --recend is set to ', ':
|
|
|
|
echo /foo, bar/, /baz, qux/, | parallel -kN1 --recend ', ' --pipe echo JOB{#}\;cat\;echo END
|
|
|
|
Output:
|
|
|
|
JOB1
|
|
/foo, END
|
|
JOB2
|
|
bar/, END
|
|
JOB3
|
|
/baz, END
|
|
JOB4
|
|
qux/,
|
|
END
|
|
|
|
Here the --recstart is set to '/':
|
|
|
|
echo /foo, bar/, /baz, qux/, | parallel -kN1 --recstart '/' --pipe echo JOB{#}\;cat\;echo END
|
|
|
|
Output:
|
|
|
|
JOB1
|
|
/foo, barEND
|
|
JOB2
|
|
/, END
|
|
JOB3
|
|
/baz, quxEND
|
|
JOB4
|
|
/,
|
|
END
|
|
|
|
Here both --recend and --recstart are set:
|
|
|
|
echo /foo, bar/, /baz, qux/, | parallel -kN1 --recend ', ' --recstart '/' --pipe echo JOB{#}\;cat\;echo END
|
|
|
|
Output:
|
|
|
|
JOB1
|
|
/foo, bar/, END
|
|
JOB2
|
|
/baz, qux/,
|
|
END
|
|
|
|
Note the difference between setting one string and setting both strings.
|
|
|
|
With --regexp the --recend and --recstart will be treated as a regular expression:
|
|
|
|
echo foo,bar,_baz,__qux, | parallel -kN1 --regexp --recend ',_+' --pipe echo JOB{#}\;cat\;echo END
|
|
|
|
Output:
|
|
|
|
JOB1
|
|
foo,bar,_END
|
|
JOB2
|
|
baz,__END
|
|
JOB3
|
|
qux,
|
|
END
|
|
|
|
GNU Parallel can remove the record separators with --remove-rec-sep/--rrs:
|
|
|
|
echo foo,bar,_baz,__qux, | parallel -kN1 --rrs --regexp --recend ',_+' --pipe echo JOB{#}\;cat\;echo END
|
|
|
|
Output:
|
|
|
|
JOB1
|
|
foo,barEND
|
|
JOB2
|
|
bazEND
|
|
JOB3
|
|
qux,
|
|
END
|
|
|
|
=head2 Header
|
|
|
|
If the input data has a header, the header can be repeated for each
|
|
job by matching the header with --header. If headers start with %:
|
|
|
|
cat num_%header | parallel --header '(%.*\n)*' --pipe -N3 echo JOB{#}\;cat
|
|
|
|
Output (the order may be different):
|
|
|
|
JOB1
|
|
%head1
|
|
%head2
|
|
1
|
|
2
|
|
3
|
|
JOB2
|
|
%head1
|
|
%head2
|
|
4
|
|
5
|
|
6
|
|
JOB3
|
|
%head1
|
|
%head2
|
|
7
|
|
8
|
|
9
|
|
JOB4
|
|
%head1
|
|
%head2
|
|
10
|
|
|
|
If the header is 2 lines, --header 2 will work:
|
|
|
|
cat num_%header | parallel --header 2 --pipe -N3 echo JOB{#}\;cat
|
|
|
|
Output: Same as above.
|
|
|
|
=head2 --pipepart
|
|
|
|
--pipe is not very efficient. It maxes out at around 500
|
|
MB/s. --pipepart can easily deliver 5 GB/s. But there are a few
|
|
limitations. The input has to be a normal file (not a pipe) given by
|
|
-a or :::: and -L/-l/-N do not work.
|
|
|
|
parallel --pipepart -a num1000000 --block 3m wc
|
|
|
|
Output (the order may be different):
|
|
|
|
444443 444444 3000002
|
|
428572 428572 3000004
|
|
126985 126984 888890
|
|
|
|
|
|
=head1 Shebang
|
|
|
|
=head2 Input data and parallel command in the same file
|
|
|
|
GNU Parallel is often called as:
|
|
|
|
cat input_file | parallel command
|
|
|
|
With --shebang the input_file and parallel can be combined into the same script.
|
|
|
|
UNIX-scripts start with a shebang line like:
|
|
|
|
#!/bin/bash
|
|
|
|
GNU Parallel can do that, too. With --shebang the arguments can be
|
|
listed in the file. The parallel command is the first line of the
|
|
script:
|
|
|
|
#!/usr/bin/parallel --shebang -r echo
|
|
|
|
foo
|
|
bar
|
|
baz
|
|
|
|
Output (the order may be different):
|
|
|
|
foo
|
|
bar
|
|
baz
|
|
|
|
=head2 Parallelizing existing scripts
|
|
|
|
GNU Parallel is often called as:
|
|
|
|
cat input_file | parallel command
|
|
parallel command ::: foo bar
|
|
|
|
If command is a script parallel can be combined into a single file so:
|
|
|
|
cat input_file | command
|
|
command foo bar
|
|
|
|
will run the script in parallel.
|
|
|
|
This perl script perl_echo works like echo:
|
|
|
|
#!/usr/bin/perl
|
|
|
|
print "@ARGV\n"
|
|
|
|
It can be called as:
|
|
|
|
parallel perl_echo ::: foo bar
|
|
|
|
By changing the #!-line it can be run in parallel
|
|
|
|
#!/usr/bin/parallel --shebang-wrap /usr/bin/perl
|
|
|
|
print "@ARGV\n"
|
|
|
|
Thus this will work:
|
|
|
|
perl_echo foo bar
|
|
|
|
Output (the order may be different):
|
|
|
|
foo
|
|
bar
|
|
|
|
This technique can be used for:
|
|
|
|
=over 9
|
|
|
|
=item Perl:
|
|
|
|
#!/usr/bin/parallel --shebang-wrap /usr/bin/perl
|
|
|
|
=item Python:
|
|
|
|
#!/usr/bin/parallel --shebang-wrap /usr/bin/python
|
|
|
|
=item Bash:
|
|
|
|
#!/usr/bin/parallel --shebang-wrap /bin/bash
|
|
|
|
=item R:
|
|
|
|
#!/usr/bin/parallel --shebang-wrap /usr/bin/Rscript --vanilla --slave
|
|
|
|
=item GNUplot:
|
|
|
|
#!/usr/bin/parallel --shebang-wrap ARG={} /usr/bin/gnuplot
|
|
|
|
=item Ruby:
|
|
|
|
#!/usr/bin/parallel --shebang-wrap /usr/bin/ruby
|
|
|
|
=back
|
|
|
|
=head1 Semaphore
|
|
|
|
GNU Parallel can work as a counting semaphore. This is slower and less
|
|
efficient than its normal mode.
|
|
|
|
A counting semaphore is like a row of toilets. People needing a toilet
|
|
can use any toilet, but if there are more people than toilets, they
|
|
will have to wait for one of the toilets to be available.
|
|
|
|
An alias for 'parallel --semaphore' is 'sem'.
|
|
|
|
'sem' will follow a person to the toilets, wait until a toilet is
|
|
available, leave the person in the toilet and exit.
|
|
|
|
'sem --fg' will follow a person to the toilets, wait until a toilet is
|
|
available, stay with the person in the toilet and exit when the person
|
|
exits.
|
|
|
|
'sem --wait' will wait for all persons to leave the toilets.
|
|
|
|
'sem' does not have a queue discipline, so the next person is chosen
|
|
randomly.
|
|
|
|
-j sets the number of toilets. The default is to have only one toilet
|
|
(technically this is called a mutex). The program is started in the
|
|
background and 'sem' exits immediately. Use --wait to wait for all
|
|
'sem's to finish:
|
|
|
|
sem 'sleep 1; echo The first finished' &&
|
|
echo The first is now running in the background &&
|
|
sem 'sleep 1; echo The second finished' &&
|
|
echo The second is now running in the background
|
|
sem --wait
|
|
|
|
Output:
|
|
|
|
The first is now running in the background
|
|
The first finished
|
|
The second is now running in the background
|
|
The second finished
|
|
|
|
The command can be run in the foreground with --fg, which will only
|
|
exit when the command completes:
|
|
|
|
sem --fg 'sleep 1; echo The first finished' &&
|
|
echo The first finished running in the foreground &&
|
|
sem --fg 'sleep 1; echo The second finished' &&
|
|
echo The second finished running in the foreground
|
|
sem --wait
|
|
|
|
The difference between this and just running the command, is that a
|
|
mutex is set, so if other sems were running in the background only one
|
|
would run at the same time.
|
|
|
|
To tell the difference between which semaphore is used, use
|
|
--semaphorename/--id. Run this in one terminal:
|
|
|
|
sem --id my_id -u 'echo First started; sleep 10; echo The first finished'
|
|
|
|
and simultaneously this in another terminal:
|
|
|
|
sem --id my_id -u 'echo Second started; sleep 10; echo The second finished'
|
|
|
|
Note how the second will only be started when the first has finished.
|
|
|
|
=head2 Counting semaphore
|
|
|
|
A mutex is like having a single toilet: When it is in use everyone
|
|
else will have to wait. A counting semaphore is like having multiple
|
|
toilets: Several people can use the toilets, but when they all are in
|
|
use, everyone else will have to wait.
|
|
|
|
sem can emulate a counting semaphore. Use --jobs to set the number of
|
|
toilets:
|
|
|
|
sem --jobs 3 --id my_id -u 'echo First started; sleep 5; echo The first finished' &&
|
|
sem --jobs 3 --id my_id -u 'echo Second started; sleep 6; echo The second finished' &&
|
|
sem --jobs 3 --id my_id -u 'echo Third started; sleep 7; echo The third finished' &&
|
|
sem --jobs 3 --id my_id -u 'echo Fourth started; sleep 8; echo The fourth finished' &&
|
|
sem --wait --id my_id
|
|
|
|
Output:
|
|
|
|
First started
|
|
Second started
|
|
Third started
|
|
The first finished
|
|
Fourth started
|
|
The second finished
|
|
The third finished
|
|
The fourth finished
|
|
|
|
=head2 Timeout
|
|
|
|
With --semaphoretimeout you can force running the command anyway after
|
|
a period (postive number) or give up (negative number):
|
|
|
|
sem --id foo -u 'echo Slow started; sleep 5; echo Slow ended' &&
|
|
sem --id foo --semaphoretimeout 1 'echo Force this running after 1 sec' &&
|
|
sem --id foo --semaphoretimeout -2 'echo Give up after 1 sec'
|
|
sem --id foo --wait
|
|
|
|
Output:
|
|
|
|
Slow started
|
|
parallel: Warning: Semaphore timed out. Stealing the semaphore.
|
|
Force this running after 1 sec
|
|
Slow ended
|
|
parallel: Warning: Semaphore timed out. Exiting.
|
|
|
|
Note how the 'Give up' was not run.
|
|
|
|
=head1 Informational
|
|
|
|
GNU Parallel has some options to give short information about the
|
|
configuration.
|
|
|
|
--help will print a summary of the most important options:
|
|
|
|
parallel --help
|
|
|
|
Output:
|
|
|
|
Usage:
|
|
parallel [options] [command [arguments]] < list_of_arguments
|
|
parallel [options] [command [arguments]] (::: arguments|:::: argfile(s))...
|
|
cat ... | parallel --pipe [options] [command [arguments]]
|
|
|
|
-j n Run n jobs in parallel
|
|
-k Keep same order
|
|
-X Multiple arguments with context replace
|
|
--colsep regexp Split input on regexp for positional replacements
|
|
{} {.} {/} {/.} {#} Replacement strings
|
|
{3} {3.} {3/} {3/.} Positional replacement strings
|
|
|
|
-S sshlogin Example: foo@server.example.com
|
|
--slf .. Use ~/.parallel/sshloginfile as the list of sshlogins
|
|
--trc {}.bar Shorthand for --transfer --return {}.bar --cleanup
|
|
--onall Run the given command with argument on all sshlogins
|
|
--nonall Run the given command with no arguments on all sshlogins
|
|
|
|
--pipe Split stdin (standard input) to multiple jobs.
|
|
--recend str Record end separator for --pipe.
|
|
--recstart str Record start separator for --pipe.
|
|
|
|
See 'man parallel' for details
|
|
|
|
When using GNU Parallel for a publication please cite:
|
|
|
|
O. Tange (2011): GNU Parallel - The Command-Line Power Tool,
|
|
;login: The USENIX Magazine, February 2011:42-47.
|
|
|
|
When asking for help, always report the full output of:
|
|
|
|
parallel --version
|
|
|
|
Output:
|
|
|
|
GNU parallel 20130822
|
|
Copyright (C) 2007,2008,2009,2010,2011,2012,2013 Ole Tange and Free Software Foundation, Inc.
|
|
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>
|
|
This is free software: you are free to change and redistribute it.
|
|
GNU parallel comes with no warranty.
|
|
|
|
Web site: http://www.gnu.org/software/parallel
|
|
|
|
When using GNU Parallel for a publication please cite:
|
|
|
|
O. Tange (2011): GNU Parallel - The Command-Line Power Tool,
|
|
;login: The USENIX Magazine, February 2011:42-47.
|
|
|
|
In scripts --minversion can be used to ensure the user has at least
|
|
this version:
|
|
|
|
parallel --minversion 20130722 && echo Your version is at least 20130722.
|
|
|
|
Output:
|
|
|
|
20130722
|
|
Your version is at least 20130722.
|
|
|
|
If using GNU Parallel for research the BibTeX citation can be
|
|
generated using --bibtex.
|
|
|
|
parallel --bibtex
|
|
|
|
Output:
|
|
|
|
@article{Tange2011a,
|
|
title = {GNU Parallel - The Command-Line Power Tool},
|
|
author = {O. Tange},
|
|
address = {Frederiksberg, Denmark},
|
|
journal = {;login: The USENIX Magazine},
|
|
month = {Feb},
|
|
number = {1},
|
|
volume = {36},
|
|
url = {http://www.gnu.org/s/parallel},
|
|
year = {2011},
|
|
pages = {42-47}
|
|
}
|
|
|
|
With --max-line-length-allowed GNU Parallel will report the maximal
|
|
size of the command line:
|
|
|
|
parallel --max-line-length-allowed
|
|
|
|
Output (may vary on different systems):
|
|
|
|
131071
|
|
|
|
--number-of-cpus and --number-of-cores run system specific code to
|
|
determine the number of CPUs and CPU cores on the system. On
|
|
unsupported platforms they will return 1:
|
|
|
|
parallel --number-of-cpus
|
|
parallel --number-of-cores
|
|
|
|
Output (may vary on different systems):
|
|
|
|
4
|
|
64
|
|
|
|
=head1 Profiles
|
|
|
|
The defaults for GNU Parallel can be changed systemwise by putting the
|
|
command line options in /etc/parallel/config. They can be changed for
|
|
a user by putting them in ~/.parallel/config.
|
|
|
|
Profiles work the same way, but have to be referred to with --profile:
|
|
|
|
echo '--nice 17' > ~/.parallel/nicetimeout
|
|
echo '--timeout 300%' >> ~/.parallel/nicetimeout
|
|
parallel --profile nicetimeout echo ::: A B C
|
|
|
|
Output:
|
|
|
|
A
|
|
B
|
|
C
|
|
|
|
Profiles can be combined:
|
|
|
|
echo '-vv --dry-run' > ~/.parallel/dryverbose
|
|
parallel --profile dryverbose --profile nicetimeout echo ::: A B C
|
|
|
|
Output:
|
|
|
|
\nice -n17 /bin/bash -c echo\ A
|
|
\nice -n17 /bin/bash -c echo\ B
|
|
\nice -n17 /bin/bash -c echo\ C
|
|
|
|
|
|
=head1 Spread the word
|
|
|
|
I hope you have learned something from this tutorial.
|
|
|
|
If you like GNU Parallel:
|
|
|
|
=over 2
|
|
|
|
=item *
|
|
|
|
(Re-)walk through the tutorial if you have not done so in the past year
|
|
(http://www.gnu.org/software/parallel/parallel_tutorial.html)
|
|
|
|
=item *
|
|
|
|
Give a demo at your local user group/team/colleagues
|
|
|
|
=item *
|
|
|
|
Post the intro videos and the tutorial on Reddit, Diaspora*,
|
|
forums, blogs, Identi.ca, Google+, Twitter, Facebook, Linkedin,
|
|
mailing lists
|
|
|
|
=item *
|
|
|
|
Request or write a review for your favourite blog or magazine
|
|
|
|
=item *
|
|
|
|
Invite me for your next conference
|
|
|
|
=back
|
|
|
|
If you use GNU Parallel for research:
|
|
|
|
=over 2
|
|
|
|
=item *
|
|
|
|
Please cite GNU Parallel in you publications (use --bibtex)
|
|
|
|
=back
|
|
|
|
If GNU Parallel saves you money:
|
|
|
|
=over 2
|
|
|
|
=item *
|
|
|
|
(Have your company) donate to FSF or become a member
|
|
https://my.fsf.org/donate/
|
|
|
|
=back
|
|
|
|
(C) 2013,2014,2015 Ole Tange, GPLv3
|
|
|
|
|
|
=cut
|