mirror of
https://git.savannah.gnu.org/git/parallel.git
synced 2024-11-25 15:37:56 +00:00
98 lines
3 KiB
Bash
98 lines
3 KiB
Bash
#!/bin/bash
|
|
|
|
# Simple jobs that never fails
|
|
# Each should be taking 30-100s and be possible to run in parallel
|
|
# I.e.: No race conditions, no logins
|
|
|
|
par_race_condition1() {
|
|
echo '### Test race condition on 8 CPU (my laptop)'
|
|
seq 1 5000000 > /tmp/parallel_race_cond
|
|
seq 1 10 | parallel -k "cat /tmp/parallel_race_cond | parallel --pipe --recend '' -k gzip >/dev/null; echo {}"
|
|
rm /tmp/parallel_race_cond
|
|
}
|
|
|
|
par_tmp_full() {
|
|
# Assume /tmp/shm is easy to fill up
|
|
export SHM=/tmp/shm/parallel
|
|
mkdir -p $SHM
|
|
sudo umount -l $SHM 2>/dev/null
|
|
sudo mount -t tmpfs -o size=10% none $SHM
|
|
|
|
echo "### Test --tmpdir running full. bug #40733 was caused by this"
|
|
stdout parallel -j1 --tmpdir $SHM cat /dev/zero ::: dummy
|
|
}
|
|
|
|
par_bug_48290() {
|
|
echo "### bug #48290: round-robin does not distribute data based on business"
|
|
echo "Jobslot 1 is 256 times slower than jobslot 4 and should get much less data"
|
|
yes "$(seq 1000|xargs)" | head -c 30M |
|
|
parallel --tagstring {%} --linebuffer --compress -j4 --roundrobin --pipe --block 10k \
|
|
pv -qL '{= $_=int( $job->slot()**4/2+1) =}'0000 |
|
|
perl -ne '/^\d+/ and $s{$&}++; END { print map { "$_\n" } sort { $s{$b} <=> $s{$a} } keys %s}'
|
|
}
|
|
|
|
par_memory_leak() {
|
|
a_run() {
|
|
seq $1 |time -v parallel true 2>&1 |
|
|
grep 'Maximum resident' |
|
|
field 6;
|
|
}
|
|
export -f a_run
|
|
echo "### Test for memory leaks"
|
|
echo "Of 30 runs of 1 job at least one should be bigger than a 3000 job run"
|
|
small_max=$(seq 30 | parallel a_run 1 | jq -s max)
|
|
big=$(a_run 3000)
|
|
if [ $small_max -lt $big ] ; then
|
|
echo "Bad: Memleak likely."
|
|
else
|
|
echo "Good: No memleak detected."
|
|
fi
|
|
}
|
|
|
|
par_linebuffer_matters_compress_tag() {
|
|
echo "### (--linebuffer) --compress --tag should give different output"
|
|
random_data_with_id_prepended() {
|
|
perl -pe 's/^/'$1'/' /dev/urandom |
|
|
pv -qL 300000 | head -c 10000000
|
|
}
|
|
export -f random_data_with_id_prepended
|
|
|
|
nolb=$(seq 10 |
|
|
parallel -j0 --compress --tag random_data_with_id_prepended {#} |
|
|
field 1 | uniq)
|
|
lb=$(seq 10 |
|
|
parallel -j0 --linebuffer --compress --tag random_data_with_id_prepended {#} |
|
|
field 1 | uniq)
|
|
if [ "$lb" == "$nolb" ] ; then
|
|
echo "BAD: --linebuffer makes no difference"
|
|
else
|
|
echo "OK: --linebuffer makes a difference"
|
|
fi
|
|
}
|
|
|
|
par_linebuffer_matters_compress() {
|
|
echo "### (--linebuffer) --compress --tag should give different output"
|
|
random_data_with_id_prepended() {
|
|
perl -pe 's/^/'$1'/' /dev/urandom |
|
|
pv -qL 300000 | head -c 1000000
|
|
}
|
|
export -f random_data_with_id_prepended
|
|
|
|
nolb=$(seq 10 |
|
|
parallel -j0 --compress random_data_with_id_prepended {#} |
|
|
field 1 | uniq)
|
|
lb=$(seq 10 |
|
|
parallel -j0 --linebuffer --compress random_data_with_id_prepended {#} |
|
|
field 1 | uniq)
|
|
if [ "$lb" == "$nolb" ] ; then
|
|
echo "BAD: --linebuffer makes no difference"
|
|
else
|
|
echo "OK: --linebuffer makes a difference"
|
|
fi
|
|
}
|
|
|
|
|
|
|
|
export -f $(compgen -A function | grep par_)
|
|
compgen -A function | grep par_ | sort | parallel -j6 --tag -k '{} 2>&1'
|