mirror of
https://git.savannah.gnu.org/git/parallel.git
synced 2024-11-22 05:57:54 +00:00
Released as 20120220 alpha.
This commit is contained in:
parent
211ffd7898
commit
587c5e23be
34
NEWS
34
NEWS
|
@ -1,3 +1,37 @@
|
||||||
|
20120222
|
||||||
|
|
||||||
|
* --workdir . will use the current working dir. If the current working
|
||||||
|
dir is beneath your home dir, the value . is treated as the relative
|
||||||
|
path to your home dir. This means that if your home dir is different
|
||||||
|
on remote computers (e.g. if your login is different) the relative
|
||||||
|
path will still be relative to your home dir.
|
||||||
|
|
||||||
|
* A texinfo file is now generated from the POD file using pod2texi.
|
||||||
|
|
||||||
|
* The bioinformatics article "Fast large-scale clustering of protein
|
||||||
|
structures using Gauss integrals" mentions GNU Parallel. They forgot
|
||||||
|
to add the reference. Please remember --bibtex if you use GNU
|
||||||
|
Parallel in an article.
|
||||||
|
http://bioinformatics.oxfordjournals.org/content/28/4/510.short
|
||||||
|
|
||||||
|
* The first supporter received his merchandise. Did you order yours?
|
||||||
|
https://www.gnu.org/software/parallel/merchandise.html
|
||||||
|
|
||||||
|
* How to use GNU Parallel with samtools.
|
||||||
|
http://zvfak.blogspot.com/2012/02/samtools-in-parallel.html
|
||||||
|
|
||||||
|
* How to use GNU Parallel for vacuuming PostgreSQL tables.
|
||||||
|
http://blog.timmattison.com/archives/2012/01/24/mini-hack-parallel-vacuuming-in-postgresql/
|
||||||
|
|
||||||
|
* Converting e-books with Calibre and GNU Parallel.
|
||||||
|
http://www.linkhal.de/blog/2012/01/converting-massive-e-book-collections-with-calibre-and-gnu-parrallel/
|
||||||
|
|
||||||
|
* Using GNU Parallel for tailing logfiles from multiple servers.
|
||||||
|
http://silviud.blogspot.com/2012/02/shell-parallel-processing.html
|
||||||
|
|
||||||
|
* Bug fixes and man page updates.
|
||||||
|
|
||||||
|
|
||||||
20120122
|
20120122
|
||||||
|
|
||||||
* --header : uses the first input line as column names and you can
|
* --header : uses the first input line as column names and you can
|
||||||
|
|
20
configure
vendored
20
configure
vendored
|
@ -1,6 +1,6 @@
|
||||||
#! /bin/sh
|
#! /bin/sh
|
||||||
# Guess values for system-dependent variables and create Makefiles.
|
# Guess values for system-dependent variables and create Makefiles.
|
||||||
# Generated by GNU Autoconf 2.68 for parallel 20120124.
|
# Generated by GNU Autoconf 2.68 for parallel 20120220.
|
||||||
#
|
#
|
||||||
# Report bugs to <bug-parallel@gnu.org>.
|
# Report bugs to <bug-parallel@gnu.org>.
|
||||||
#
|
#
|
||||||
|
@ -559,8 +559,8 @@ MAKEFLAGS=
|
||||||
# Identity of this package.
|
# Identity of this package.
|
||||||
PACKAGE_NAME='parallel'
|
PACKAGE_NAME='parallel'
|
||||||
PACKAGE_TARNAME='parallel'
|
PACKAGE_TARNAME='parallel'
|
||||||
PACKAGE_VERSION='20120124'
|
PACKAGE_VERSION='20120220'
|
||||||
PACKAGE_STRING='parallel 20120124'
|
PACKAGE_STRING='parallel 20120220'
|
||||||
PACKAGE_BUGREPORT='bug-parallel@gnu.org'
|
PACKAGE_BUGREPORT='bug-parallel@gnu.org'
|
||||||
PACKAGE_URL=''
|
PACKAGE_URL=''
|
||||||
|
|
||||||
|
@ -1176,7 +1176,7 @@ if test "$ac_init_help" = "long"; then
|
||||||
# Omit some internal or obsolete options to make the list less imposing.
|
# Omit some internal or obsolete options to make the list less imposing.
|
||||||
# This message is too long to be a string in the A/UX 3.1 sh.
|
# This message is too long to be a string in the A/UX 3.1 sh.
|
||||||
cat <<_ACEOF
|
cat <<_ACEOF
|
||||||
\`configure' configures parallel 20120124 to adapt to many kinds of systems.
|
\`configure' configures parallel 20120220 to adapt to many kinds of systems.
|
||||||
|
|
||||||
Usage: $0 [OPTION]... [VAR=VALUE]...
|
Usage: $0 [OPTION]... [VAR=VALUE]...
|
||||||
|
|
||||||
|
@ -1242,7 +1242,7 @@ fi
|
||||||
|
|
||||||
if test -n "$ac_init_help"; then
|
if test -n "$ac_init_help"; then
|
||||||
case $ac_init_help in
|
case $ac_init_help in
|
||||||
short | recursive ) echo "Configuration of parallel 20120124:";;
|
short | recursive ) echo "Configuration of parallel 20120220:";;
|
||||||
esac
|
esac
|
||||||
cat <<\_ACEOF
|
cat <<\_ACEOF
|
||||||
|
|
||||||
|
@ -1309,7 +1309,7 @@ fi
|
||||||
test -n "$ac_init_help" && exit $ac_status
|
test -n "$ac_init_help" && exit $ac_status
|
||||||
if $ac_init_version; then
|
if $ac_init_version; then
|
||||||
cat <<\_ACEOF
|
cat <<\_ACEOF
|
||||||
parallel configure 20120124
|
parallel configure 20120220
|
||||||
generated by GNU Autoconf 2.68
|
generated by GNU Autoconf 2.68
|
||||||
|
|
||||||
Copyright (C) 2010 Free Software Foundation, Inc.
|
Copyright (C) 2010 Free Software Foundation, Inc.
|
||||||
|
@ -1326,7 +1326,7 @@ cat >config.log <<_ACEOF
|
||||||
This file contains any messages produced by compilers while
|
This file contains any messages produced by compilers while
|
||||||
running configure, to aid debugging if configure makes a mistake.
|
running configure, to aid debugging if configure makes a mistake.
|
||||||
|
|
||||||
It was created by parallel $as_me 20120124, which was
|
It was created by parallel $as_me 20120220, which was
|
||||||
generated by GNU Autoconf 2.68. Invocation command line was
|
generated by GNU Autoconf 2.68. Invocation command line was
|
||||||
|
|
||||||
$ $0 $@
|
$ $0 $@
|
||||||
|
@ -2141,7 +2141,7 @@ fi
|
||||||
|
|
||||||
# Define the identity of the package.
|
# Define the identity of the package.
|
||||||
PACKAGE='parallel'
|
PACKAGE='parallel'
|
||||||
VERSION='20120124'
|
VERSION='20120220'
|
||||||
|
|
||||||
|
|
||||||
cat >>confdefs.h <<_ACEOF
|
cat >>confdefs.h <<_ACEOF
|
||||||
|
@ -2704,7 +2704,7 @@ cat >>$CONFIG_STATUS <<\_ACEOF || ac_write_fail=1
|
||||||
# report actual input values of CONFIG_FILES etc. instead of their
|
# report actual input values of CONFIG_FILES etc. instead of their
|
||||||
# values after options handling.
|
# values after options handling.
|
||||||
ac_log="
|
ac_log="
|
||||||
This file was extended by parallel $as_me 20120124, which was
|
This file was extended by parallel $as_me 20120220, which was
|
||||||
generated by GNU Autoconf 2.68. Invocation command line was
|
generated by GNU Autoconf 2.68. Invocation command line was
|
||||||
|
|
||||||
CONFIG_FILES = $CONFIG_FILES
|
CONFIG_FILES = $CONFIG_FILES
|
||||||
|
@ -2766,7 +2766,7 @@ _ACEOF
|
||||||
cat >>$CONFIG_STATUS <<_ACEOF || ac_write_fail=1
|
cat >>$CONFIG_STATUS <<_ACEOF || ac_write_fail=1
|
||||||
ac_cs_config="`$as_echo "$ac_configure_args" | sed 's/^ //; s/[\\""\`\$]/\\\\&/g'`"
|
ac_cs_config="`$as_echo "$ac_configure_args" | sed 's/^ //; s/[\\""\`\$]/\\\\&/g'`"
|
||||||
ac_cs_version="\\
|
ac_cs_version="\\
|
||||||
parallel config.status 20120124
|
parallel config.status 20120220
|
||||||
configured by $0, generated by GNU Autoconf 2.68,
|
configured by $0, generated by GNU Autoconf 2.68,
|
||||||
with options \\"\$ac_cs_config\\"
|
with options \\"\$ac_cs_config\\"
|
||||||
|
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
AC_INIT([parallel], [20120124], [bug-parallel@gnu.org])
|
AC_INIT([parallel], [20120220], [bug-parallel@gnu.org])
|
||||||
AM_INIT_AUTOMAKE([-Wall -Werror foreign])
|
AM_INIT_AUTOMAKE([-Wall -Werror foreign])
|
||||||
AC_CONFIG_HEADERS([config.h])
|
AC_CONFIG_HEADERS([config.h])
|
||||||
AC_CONFIG_FILES([
|
AC_CONFIG_FILES([
|
||||||
|
|
|
@ -41,8 +41,6 @@ surprising file names containing space, ', or " in the filename.
|
||||||
xargs can give nasty surprises due to the separator problem
|
xargs can give nasty surprises due to the separator problem
|
||||||
http://pi.dk/5
|
http://pi.dk/5
|
||||||
|
|
||||||
GNU Parallel http://nd.gd/0s may be better.
|
|
||||||
|
|
||||||
@jaylyerly @stevenf xargs will bite you if file names contain
|
@jaylyerly @stevenf xargs will bite you if file names contain
|
||||||
space http://pi.dk/5. Use GNU Parallel instead: http://pi.dk/0
|
space http://pi.dk/5. Use GNU Parallel instead: http://pi.dk/0
|
||||||
|
|
||||||
|
|
|
@ -193,12 +193,21 @@ New in this release:
|
||||||
Parallel in an article.
|
Parallel in an article.
|
||||||
http://bioinformatics.oxfordjournals.org/content/28/4/510.short
|
http://bioinformatics.oxfordjournals.org/content/28/4/510.short
|
||||||
|
|
||||||
* First supporter received his merchandise. Did you order yours?
|
* The first supporter received his merchandise. Did you order yours?
|
||||||
https://www.gnu.org/software/parallel/merchandise.html
|
https://www.gnu.org/software/parallel/merchandise.html
|
||||||
|
|
||||||
|
* How to use GNU Parallel with samtools.
|
||||||
|
http://zvfak.blogspot.com/2012/02/samtools-in-parallel.html
|
||||||
|
|
||||||
* How to use GNU Parallel for vacuuming PostgreSQL tables.
|
* How to use GNU Parallel for vacuuming PostgreSQL tables.
|
||||||
http://blog.timmattison.com/archives/2012/01/24/mini-hack-parallel-vacuuming-in-postgresql/
|
http://blog.timmattison.com/archives/2012/01/24/mini-hack-parallel-vacuuming-in-postgresql/
|
||||||
|
|
||||||
|
* Converting e-books with Calibre and GNU Parallel.
|
||||||
|
http://www.linkhal.de/blog/2012/01/converting-massive-e-book-collections-with-calibre-and-gnu-parrallel/
|
||||||
|
|
||||||
|
* Using GNU Parallel for tailing logfiles from multiple servers.
|
||||||
|
http://silviud.blogspot.com/2012/02/shell-parallel-processing.html
|
||||||
|
|
||||||
* Bug fixes and man page updates.
|
* Bug fixes and man page updates.
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -38,16 +38,20 @@ niceload.html: niceload.pod Makefile sql.html
|
||||||
rm -f $(srcdir)/pod2htm*
|
rm -f $(srcdir)/pod2htm*
|
||||||
|
|
||||||
parallel.texi: parallel.pod
|
parallel.texi: parallel.pod
|
||||||
pod2texi --output=$(srcdir)/parallel.texi $(srcdir)/parallel.pod
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/parallel.texi $(srcdir)/parallel.pod || true
|
||||||
|
|
||||||
sem.texi: sem.pod
|
sem.texi: sem.pod
|
||||||
pod2texi --output=$(srcdir)/sem.texi $(srcdir)/sem.pod
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/sem.texi $(srcdir)/sem.pod || true
|
||||||
|
|
||||||
sql.texi: sql
|
sql.texi: sql
|
||||||
pod2texi --output=$(srcdir)/sql.texi $(srcdir)/sql
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/sql.texi $(srcdir)/sql || true
|
||||||
|
|
||||||
niceload.texi: niceload.pod
|
niceload.texi: niceload.pod
|
||||||
pod2texi --output=$(srcdir)/niceload.texi $(srcdir)/niceload.pod
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/niceload.texi $(srcdir)/niceload.pod || true
|
||||||
|
|
||||||
sem: parallel
|
sem: parallel
|
||||||
ln -fs parallel sem
|
ln -fs parallel sem
|
||||||
|
|
|
@ -483,16 +483,20 @@ niceload.html: niceload.pod Makefile sql.html
|
||||||
rm -f $(srcdir)/pod2htm*
|
rm -f $(srcdir)/pod2htm*
|
||||||
|
|
||||||
parallel.texi: parallel.pod
|
parallel.texi: parallel.pod
|
||||||
pod2texi --output=$(srcdir)/parallel.texi $(srcdir)/parallel.pod
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/parallel.texi $(srcdir)/parallel.pod || true
|
||||||
|
|
||||||
sem.texi: sem.pod
|
sem.texi: sem.pod
|
||||||
pod2texi --output=$(srcdir)/sem.texi $(srcdir)/sem.pod
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/sem.texi $(srcdir)/sem.pod || true
|
||||||
|
|
||||||
sql.texi: sql
|
sql.texi: sql
|
||||||
pod2texi --output=$(srcdir)/sql.texi $(srcdir)/sql
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/sql.texi $(srcdir)/sql || true
|
||||||
|
|
||||||
niceload.texi: niceload.pod
|
niceload.texi: niceload.pod
|
||||||
pod2texi --output=$(srcdir)/niceload.texi $(srcdir)/niceload.pod
|
# If pod2texi is not installed: Forget about it
|
||||||
|
pod2texi --output=$(srcdir)/niceload.texi $(srcdir)/niceload.pod || true
|
||||||
|
|
||||||
sem: parallel
|
sem: parallel
|
||||||
ln -fs parallel sem
|
ln -fs parallel sem
|
||||||
|
|
|
@ -24,7 +24,7 @@
|
||||||
use strict;
|
use strict;
|
||||||
use Getopt::Long;
|
use Getopt::Long;
|
||||||
$Global::progname="niceload";
|
$Global::progname="niceload";
|
||||||
$Global::version = 20120124;
|
$Global::version = 20120220;
|
||||||
Getopt::Long::Configure("bundling","require_order");
|
Getopt::Long::Configure("bundling","require_order");
|
||||||
get_options_from_array(\@ARGV) || die_usage();
|
get_options_from_array(\@ARGV) || die_usage();
|
||||||
if($::opt_version) {
|
if($::opt_version) {
|
||||||
|
|
|
@ -543,7 +543,7 @@ sub get_options_from_array {
|
||||||
sub parse_options {
|
sub parse_options {
|
||||||
# Returns: N/A
|
# Returns: N/A
|
||||||
# Defaults:
|
# Defaults:
|
||||||
$Global::version = 20120215;
|
$Global::version = 20120220;
|
||||||
$Global::progname = 'parallel';
|
$Global::progname = 'parallel';
|
||||||
$Global::infinity = 2**31;
|
$Global::infinity = 2**31;
|
||||||
$Global::debug = 0;
|
$Global::debug = 0;
|
||||||
|
|
|
@ -500,7 +500,7 @@ status will be the exit status from the failing job.
|
||||||
=back
|
=back
|
||||||
|
|
||||||
|
|
||||||
=item B<--header> I<regexp> (alpha testing)
|
=item B<--header> I<regexp> (beta testing)
|
||||||
|
|
||||||
Use upto regexp as header. For normal usage the matched header
|
Use upto regexp as header. For normal usage the matched header
|
||||||
(typically the first line: B<--header '\n'>) will be split using
|
(typically the first line: B<--header '\n'>) will be split using
|
||||||
|
@ -743,9 +743,9 @@ Instead of printing the output to stdout (standard output) the output
|
||||||
of each job is saved in a file and the filename is then printed.
|
of each job is saved in a file and the filename is then printed.
|
||||||
|
|
||||||
|
|
||||||
=item B<--pipe> (beta testing)
|
=item B<--pipe>
|
||||||
|
|
||||||
=item B<--spreadstdin> (beta testing)
|
=item B<--spreadstdin>
|
||||||
|
|
||||||
Spread input to jobs on stdin (standard input). Read a block of data
|
Spread input to jobs on stdin (standard input). Read a block of data
|
||||||
from stdin (standard input) and give one block of data as input to one
|
from stdin (standard input) and give one block of data as input to one
|
||||||
|
@ -937,7 +937,7 @@ it to the command.
|
||||||
Only used with B<--pipe>.
|
Only used with B<--pipe>.
|
||||||
|
|
||||||
|
|
||||||
=item B<--resume> (alpha testing)
|
=item B<--resume> (beta testing)
|
||||||
|
|
||||||
Resumes from the last unfinished job. By reading B<--joblog> GNU
|
Resumes from the last unfinished job. By reading B<--joblog> GNU
|
||||||
B<parallel> will figure out the last unfinished job and continue from
|
B<parallel> will figure out the last unfinished job and continue from
|
||||||
|
@ -1114,9 +1114,9 @@ B<--sshlogin> is often used with B<--transfer>, B<--return>,
|
||||||
B<--cleanup>, and B<--trc>.
|
B<--cleanup>, and B<--trc>.
|
||||||
|
|
||||||
|
|
||||||
=item B<--sshloginfile> I<filename> (alpha testing)
|
=item B<--sshloginfile> I<filename> (beta testing)
|
||||||
|
|
||||||
=item B<--slf> I<filename> (alpha testing)
|
=item B<--slf> I<filename> (beta testing)
|
||||||
|
|
||||||
File with sshlogins. The file consists of sshlogins on separate
|
File with sshlogins. The file consists of sshlogins on separate
|
||||||
lines. Empty lines and lines starting with '#' are ignored. Example:
|
lines. Empty lines and lines starting with '#' are ignored. Example:
|
||||||
|
@ -1196,7 +1196,7 @@ different dir for the files. Setting B<--tmpdir> is equivalent to
|
||||||
setting $TMPDIR.
|
setting $TMPDIR.
|
||||||
|
|
||||||
|
|
||||||
=item B<--timeout> I<sec> (beta testing)
|
=item B<--timeout> I<sec>
|
||||||
|
|
||||||
Time out for command. If the command runs for longer than I<sec>
|
Time out for command. If the command runs for longer than I<sec>
|
||||||
seconds it will get killed with SIGTERM, followed by SIGTERM 200 ms
|
seconds it will get killed with SIGTERM, followed by SIGTERM 200 ms
|
||||||
|
@ -1322,9 +1322,9 @@ Use B<-v> B<-v> to print the wrapping ssh command when running remotely.
|
||||||
Print the version GNU B<parallel> and exit.
|
Print the version GNU B<parallel> and exit.
|
||||||
|
|
||||||
|
|
||||||
=item B<--workdir> I<mydir>
|
=item B<--workdir> I<mydir> (alpha testing)
|
||||||
|
|
||||||
=item B<--wd> I<mydir>
|
=item B<--wd> I<mydir> (alpha testing)
|
||||||
|
|
||||||
Files transferred using B<--transfer> and B<--return> will be relative
|
Files transferred using B<--transfer> and B<--return> will be relative
|
||||||
to I<mydir> on remote computers, and the command will be executed in
|
to I<mydir> on remote computers, and the command will be executed in
|
||||||
|
|
|
@ -531,8 +531,8 @@ status will be the exit status from the failing job.
|
||||||
|
|
||||||
@end table
|
@end table
|
||||||
|
|
||||||
@item @strong{--header} @emph{regexp} (alpha testing)
|
@item @strong{--header} @emph{regexp} (beta testing)
|
||||||
@anchor{@strong{--header} @emph{regexp} (alpha testing)}
|
@anchor{@strong{--header} @emph{regexp} (beta testing)}
|
||||||
|
|
||||||
Use upto regexp as header. For normal usage the matched header
|
Use upto regexp as header. For normal usage the matched header
|
||||||
(typically the first line: @strong{--header '\n'}) will be split using
|
(typically the first line: @strong{--header '\n'}) will be split using
|
||||||
|
@ -798,11 +798,11 @@ all the output from one server will be grouped together.
|
||||||
Instead of printing the output to stdout (standard output) the output
|
Instead of printing the output to stdout (standard output) the output
|
||||||
of each job is saved in a file and the filename is then printed.
|
of each job is saved in a file and the filename is then printed.
|
||||||
|
|
||||||
@item @strong{--pipe} (beta testing)
|
@item @strong{--pipe}
|
||||||
@anchor{@strong{--pipe} (beta testing)}
|
@anchor{@strong{--pipe}}
|
||||||
|
|
||||||
@item @strong{--spreadstdin} (beta testing)
|
@item @strong{--spreadstdin}
|
||||||
@anchor{@strong{--spreadstdin} (beta testing)}
|
@anchor{@strong{--spreadstdin}}
|
||||||
|
|
||||||
Spread input to jobs on stdin (standard input). Read a block of data
|
Spread input to jobs on stdin (standard input). Read a block of data
|
||||||
from stdin (standard input) and give one block of data as input to one
|
from stdin (standard input) and give one block of data as input to one
|
||||||
|
@ -1002,8 +1002,8 @@ it to the command.
|
||||||
|
|
||||||
Only used with @strong{--pipe}.
|
Only used with @strong{--pipe}.
|
||||||
|
|
||||||
@item @strong{--resume} (alpha testing)
|
@item @strong{--resume} (beta testing)
|
||||||
@anchor{@strong{--resume} (alpha testing)}
|
@anchor{@strong{--resume} (beta testing)}
|
||||||
|
|
||||||
Resumes from the last unfinished job. By reading @strong{--joblog} GNU
|
Resumes from the last unfinished job. By reading @strong{--joblog} GNU
|
||||||
@strong{parallel} will figure out the last unfinished job and continue from
|
@strong{parallel} will figure out the last unfinished job and continue from
|
||||||
|
@ -1190,11 +1190,11 @@ The remote host must have GNU @strong{parallel} installed.
|
||||||
@strong{--sshlogin} is often used with @strong{--transfer}, @strong{--return},
|
@strong{--sshlogin} is often used with @strong{--transfer}, @strong{--return},
|
||||||
@strong{--cleanup}, and @strong{--trc}.
|
@strong{--cleanup}, and @strong{--trc}.
|
||||||
|
|
||||||
@item @strong{--sshloginfile} @emph{filename} (alpha testing)
|
@item @strong{--sshloginfile} @emph{filename} (beta testing)
|
||||||
@anchor{@strong{--sshloginfile} @emph{filename} (alpha testing)}
|
@anchor{@strong{--sshloginfile} @emph{filename} (beta testing)}
|
||||||
|
|
||||||
@item @strong{--slf} @emph{filename} (alpha testing)
|
@item @strong{--slf} @emph{filename} (beta testing)
|
||||||
@anchor{@strong{--slf} @emph{filename} (alpha testing)}
|
@anchor{@strong{--slf} @emph{filename} (beta testing)}
|
||||||
|
|
||||||
File with sshlogins. The file consists of sshlogins on separate
|
File with sshlogins. The file consists of sshlogins on separate
|
||||||
lines. Empty lines and lines starting with '#' are ignored. Example:
|
lines. Empty lines and lines starting with '#' are ignored. Example:
|
||||||
|
@ -1275,8 +1275,8 @@ into temporary files in /tmp. By setting @strong{--tmpdir} you can use a
|
||||||
different dir for the files. Setting @strong{--tmpdir} is equivalent to
|
different dir for the files. Setting @strong{--tmpdir} is equivalent to
|
||||||
setting $TMPDIR.
|
setting $TMPDIR.
|
||||||
|
|
||||||
@item @strong{--timeout} @emph{sec} (beta testing)
|
@item @strong{--timeout} @emph{sec}
|
||||||
@anchor{@strong{--timeout} @emph{sec} (beta testing)}
|
@anchor{@strong{--timeout} @emph{sec}}
|
||||||
|
|
||||||
Time out for command. If the command runs for longer than @emph{sec}
|
Time out for command. If the command runs for longer than @emph{sec}
|
||||||
seconds it will get killed with SIGTERM, followed by SIGTERM 200 ms
|
seconds it will get killed with SIGTERM, followed by SIGTERM 200 ms
|
||||||
|
@ -1413,11 +1413,11 @@ Use @strong{-v} @strong{-v} to print the wrapping ssh command when running remot
|
||||||
|
|
||||||
Print the version GNU @strong{parallel} and exit.
|
Print the version GNU @strong{parallel} and exit.
|
||||||
|
|
||||||
@item @strong{--workdir} @emph{mydir}
|
@item @strong{--workdir} @emph{mydir} (alpha testing)
|
||||||
@anchor{@strong{--workdir} @emph{mydir}}
|
@anchor{@strong{--workdir} @emph{mydir} (alpha testing)}
|
||||||
|
|
||||||
@item @strong{--wd} @emph{mydir}
|
@item @strong{--wd} @emph{mydir} (alpha testing)
|
||||||
@anchor{@strong{--wd} @emph{mydir}}
|
@anchor{@strong{--wd} @emph{mydir} (alpha testing)}
|
||||||
|
|
||||||
Files transferred using @strong{--transfer} and @strong{--return} will be relative
|
Files transferred using @strong{--transfer} and @strong{--return} will be relative
|
||||||
to @emph{mydir} on remote computers, and the command will be executed in
|
to @emph{mydir} on remote computers, and the command will be executed in
|
||||||
|
|
2
src/sql
2
src/sql
|
@ -556,7 +556,7 @@ $Global::Initfile && unlink $Global::Initfile;
|
||||||
exit ($err);
|
exit ($err);
|
||||||
|
|
||||||
sub parse_options {
|
sub parse_options {
|
||||||
$Global::version = 20120124;
|
$Global::version = 20120220;
|
||||||
$Global::progname = 'sql';
|
$Global::progname = 'sql';
|
||||||
|
|
||||||
# This must be done first as this may exec myself
|
# This must be done first as this may exec myself
|
||||||
|
|
|
@ -49,7 +49,7 @@ niceload -H -l 9.9 uptime | grep ':.[1-9][0-9].[0-9][0-9],' || echo OK
|
||||||
niceload --hard -l 9 uptime | grep ':.[1-9][0-9].[0-9][0-9],' || echo OK
|
niceload --hard -l 9 uptime | grep ':.[1-9][0-9].[0-9][0-9],' || echo OK
|
||||||
|
|
||||||
echo '### -f and --factor'
|
echo '### -f and --factor'
|
||||||
niceload -H -f 0.1 -l6 echo f 0.1 first &
|
niceload -H -f 0.01 -l6 echo f 0.1 first &
|
||||||
niceload -H --factor 10 -l6 echo factor 10 last &
|
niceload -H --factor 10 -l6 echo factor 10 last &
|
||||||
wait
|
wait
|
||||||
|
|
||||||
|
|
|
@ -18,7 +18,7 @@ echo '### bug #34422: parallel -X --eta crashes with div by zero'
|
||||||
seq 2 | stdout parallel -X --eta echo
|
seq 2 | stdout parallel -X --eta echo
|
||||||
|
|
||||||
echo '### --timeout on remote machines'
|
echo '### --timeout on remote machines'
|
||||||
parallel -j0 --timeout 3 --onall -S localhost,parallel@parallel-server1 'sleep {}; echo slept {}' ::: 1 8 9 ; echo jobs failed: $?
|
parallel -j0 --timeout 4 --onall -S localhost,parallel@parallel-server1 'sleep {}; echo slept {}' ::: 1 8 9 ; echo jobs failed: $?
|
||||||
|
|
||||||
echo '### --pipe without command'
|
echo '### --pipe without command'
|
||||||
seq -w 10 | stdout parallel --pipe
|
seq -w 10 | stdout parallel --pipe
|
||||||
|
|
|
@ -85,7 +85,7 @@ echo '### Test -N odd';
|
||||||
seq 1 11 | parallel -j2 -k -N 2 --pipe cat";echo ole;sleep 0.\$PARALLEL_SEQ"
|
seq 1 11 | parallel -j2 -k -N 2 --pipe cat";echo ole;sleep 0.\$PARALLEL_SEQ"
|
||||||
|
|
||||||
echo '### Test -N even+2';
|
echo '### Test -N even+2';
|
||||||
seq 1 12 | parallel -j2 -k -N 2 --pipe cat";echo ole;sleep 0.\$PARALLEL_SEQ"
|
seq 1 12 | parallel -j2 -k -N 2 --pipe cat";echo ole;sleep 1.\$PARALLEL_SEQ"
|
||||||
|
|
||||||
echo '### Test --recstart + --recend';
|
echo '### Test --recstart + --recend';
|
||||||
cat /tmp/blocktest | parallel --block 1m -k --recstart 44 --recend "44" -j10 --pipe sort -n |md5sum
|
cat /tmp/blocktest | parallel --block 1m -k --recstart 44 --recend "44" -j10 --pipe sort -n |md5sum
|
||||||
|
|
Loading…
Reference in a new issue