|
|
| |
MCE::Child(3) |
User Contributed Perl Documentation |
MCE::Child(3) |
MCE::Child - A threads-like parallelization module compatible with Perl 5.8
This document describes MCE::Child version 1.878
use MCE::Child;
MCE::Child->init(
max_workers => 'auto', # default undef, unlimited
# Specify a percentage. MCE::Child 1.876+.
max_workers => '25%', # 4 on HW with 16 lcores
max_workers => '50%', # 8 on HW with 16 lcores
child_timeout => 20, # default undef, no timeout
posix_exit => 1, # default undef, CORE::exit
void_context => 1, # default undef
on_start => sub {
my ( $pid, $ident ) = @_;
...
},
on_finish => sub {
my ( $pid, $exit, $ident, $signal, $error, @ret ) = @_;
...
}
);
MCE::Child->create( sub { print "Hello from child\n" } )->join();
sub parallel {
my ($arg1) = @_;
print "Hello again, $arg1\n" if defined($arg1);
print "Hello again, $_\n"; # same thing
}
MCE::Child->create( \¶llel, $_ ) for 1 .. 3;
my @procs = MCE::Child->list();
my @pids = MCE::Child->list_pids();
my @running = MCE::Child->list_running();
my @joinable = MCE::Child->list_joinable();
my @count = MCE::Child->pending();
# Joining is orderly, e.g. child1 is joined first, child2, child3.
$_->join() for @procs; # (or)
$_->join() for @joinable;
# Joining occurs immediately as child processes complete execution.
1 while MCE::Child->wait_one();
my $child = mce_child { foreach (@files) { ... } };
$child->join();
if ( my $err = $child->error() ) {
warn "Child error: $err\n";
}
# Get a child's object
$child = MCE::Child->self();
# Get a child's ID
$pid = MCE::Child->pid(); # $$
$pid = $child->pid();
$pid = MCE::Child->tid(); # tid is an alias for pid
$pid = $child->tid();
# Test child objects
if ( $child1 == $child2 ) {
...
}
# Give other workers a chance to run
MCE::Child->yield();
MCE::Child->yield(0.05);
# Return context, wantarray aware
my ($value1, $value2) = $child->join();
my $value = $child->join();
# Check child's state
if ( $child->is_running() ) {
sleep 1;
}
if ( $child->is_joinable() ) {
$child->join();
}
# Send a signal to a child
$child->kill('SIGUSR1');
# Exit a child
MCE::Child->exit(0);
MCE::Child->exit(0, @ret);
MCE::Child is a fork of MCE::Hobo for compatibility with Perl 5.8.
A child is a migratory worker inside the machine that carries the
asynchronous gene. Child processes are equipped with
"threads"-like capability for running code
asynchronously. Unlike threads, each child is a unique process to the
underlying OS. The IPC is handled via
"MCE::Channel", which runs on all the
major platforms including Cygwin and Strawberry Perl.
"MCE::Child" may be used as a
standalone or together with "MCE"
including running alongside "threads".
use MCE::Child;
use MCE::Shared;
# synopsis: head -20 file.txt | perl script.pl
my $ifh = MCE::Shared->handle( "<", \*STDIN ); # shared
my $ofh = MCE::Shared->handle( ">", \*STDOUT );
my $ary = MCE::Shared->array();
sub parallel_task {
my ( $id ) = @_;
while ( <$ifh> ) {
printf {$ofh} "[ %4d ] %s", $., $_;
# $ary->[ $. - 1 ] = "[ ID $id ] read line $.\n" ); # dereferencing
$ary->set( $. - 1, "[ ID $id ] read line $.\n" ); # faster via OO
}
}
my $child1 = MCE::Child->new( "parallel_task", 1 );
my $child2 = MCE::Child->new( \¶llel_task, 2 );
my $child3 = MCE::Child->new( sub { parallel_task(3) } );
$_->join for MCE::Child->list(); # ditto: MCE::Child->wait_all();
# search array (total one round-trip via IPC)
my @vals = $ary->vals( "val =~ / ID 2 /" );
print {*STDERR} join("", @vals);
- $child = MCE::Child->create( FUNCTION, ARGS )
- $child = MCE::Child->new( FUNCTION, ARGS )
- This will create a new child process that will begin execution with
function as the entry point, and optionally ARGS for list of parameters.
It will return the corresponding MCE::Child object, or undef if child
creation failed.
FUNCTION may either be the name of a function, an
anonymous subroutine, or a code ref.
my $child = MCE::Child->create( "func_name", ... );
# or
my $child = MCE::Child->create( sub { ... }, ... );
# or
my $child = MCE::Child->create( \&func, ... );
- $child = MCE::Child->create( { options }, FUNCTION, ARGS )
- $child = MCE::Child->create( IDENT, FUNCTION, ARGS )
- Options, excluding "ident", may be
specified globally via the "init"
function. Otherwise, "ident",
"child_timeout",
"posix_exit", and
"void_context" may be set uniquely.
The "ident" option is used
by callback functions "on_start" and
"on_finish" for identifying the
started and finished child process respectively.
my $child1 = MCE::Child->create( { posix_exit => 1 }, sub {
...
} );
$child1->join;
my $child2 = MCE::Child->create( { child_timeout => 3 }, sub {
sleep 1 for ( 1 .. 9 );
} );
$child2->join;
if ( $child2->error() eq "Child timed out\n" ) {
...
}
The "new()" method is an
alias for "create()".
- mce_child { BLOCK } ARGS;
- mce_child { BLOCK };
- "mce_child" runs the block
asynchronously similarly to
"MCE::Child->create()". It returns
the child object, or undef if child creation failed.
my $child = mce_child { foreach (@files) { ... } };
$child->join();
if ( my $err = $child->error() ) {
warn("Child error: $err\n");
}
- $child->join()
- This will wait for the corresponding child process to complete its
execution. In non-voided context,
"join()" will return the value(s) of the
entry point function.
The context (void, scalar or list) for the return value(s) for
"join" is determined at the time of
joining and mostly "wantarray"
aware.
my $child1 = MCE::Child->create( sub {
my @res = qw(foo bar baz);
return (@res);
});
my @res1 = $child1->join(); # ( foo, bar, baz )
my $res1 = $child1->join(); # baz
my $child2 = MCE::Child->create( sub {
return 'foo';
});
my @res2 = $child2->join(); # ( foo )
my $res2 = $child2->join(); # foo
- $child1->equal( $child2 )
- Tests if two child objects are the same child or not. Child comparison is
based on process IDs. This is overloaded to the more natural forms.
if ( $child1 == $child2 ) {
print("Child objects are the same\n");
}
# or
if ( $child1 != $child2 ) {
print("Child objects differ\n");
}
- $child->error()
- Child processes are executed in an
"eval" context. This method will return
"undef" if the child terminates
normally. Otherwise, it returns the value of
$@ associated with the child's execution status in
its "eval" context.
- $child->exit()
- This sends 'SIGQUIT' to the child process,
notifying the child to exit. It returns the child object to allow for
method chaining. It is important to join later if not immediately to not
leave a zombie or defunct process.
$child->exit()->join();
...
$child->join(); # later
- MCE::Child->exit( 0 )
- MCE::Child->exit( 0, @ret )
- A child can exit at any time by calling
"MCE::Child->exit()". Otherwise, the
behavior is the same as "exit(status)"
when called from the main process. The child process may optionally return
data, to be sent via IPC.
- MCE::Child->finish()
- This class method is called automatically by
"END", but may be called explicitly. An
error is emitted via croak if there are active child processes not yet
joined.
MCE::Child->create( 'task1', $_ ) for 1 .. 4;
$_->join for MCE::Child->list();
MCE::Child->create( 'task2', $_ ) for 1 .. 4;
$_->join for MCE::Child->list();
MCE::Child->create( 'task3', $_ ) for 1 .. 4;
$_->join for MCE::Child->list();
MCE::Child->finish();
- MCE::Child->init( options )
- The init function accepts a list of MCE::Child options.
MCE::Child->init(
max_workers => 'auto', # default undef, unlimited
# Specify a percentage. MCE::Child 1.876+.
max_workers => '25%', # 4 on HW with 16 lcores
max_workers => '50%', # 8 on HW with 16 lcores
child_timeout => 20, # default undef, no timeout
posix_exit => 1, # default undef, CORE::exit
void_context => 1, # default undef
on_start => sub {
my ( $pid, $ident ) = @_;
...
},
on_finish => sub {
my ( $pid, $exit, $ident, $signal, $error, @ret ) = @_;
...
}
);
# Identification given as an option or the 1st argument.
for my $key ( 'aa' .. 'zz' ) {
MCE::Child->create( { ident => $key }, sub { ... } );
MCE::Child->create( $key, sub { ... } );
}
MCE::Child->wait_all;
Set "max_workers" if you
want to limit the number of workers by waiting automatically for an
available slot. Specify a percentage or
"auto" to obtain the number of logical
cores via "MCE::Util::get_ncpu()".
Set "child_timeout", in
number of seconds, if you want the child process to terminate after some
time. The default is 0 for no timeout.
Set "posix_exit" to avoid
all END and destructor processing. Constructing MCE::Child inside a
thread implies 1 or if present CGI, FCGI, Coro, Curses, Gearman::Util,
Gearman::XS, LWP::UserAgent, Mojo::IOLoop, STFL, Tk, Wx, or
Win32::GUI.
Set "void_context" to create
the child process in void context for the return value. Otherwise, the
return context is wantarray-aware for
"join()" and
"result()" and determined when
retrieving the data.
The callback options
"on_start" and
"on_finish" are called in the parent
process after starting the worker and later when terminated. The
arguments for the subroutines were inspired by
Parallel::ForkManager.
The parameters for
"on_start" are the following:
- pid of the child process
- identification (ident option or 1st arg to create)
The parameters for
"on_finish" are the following:
- pid of the child process
- program exit code
- identification (ident option or 1st arg to create)
- exit signal id
- error message from eval inside MCE::Child
- returned data
- $child->is_running()
- Returns true if a child is still running.
- $child->is_joinable()
- Returns true if the child has finished running and not yet joined.
- $child->kill( 'SIG...' )
- Sends the specified signal to the child. Returns the child object to allow
for method chaining. As with "exit", it
is important to join eventually if not immediately to not leave a zombie
or defunct process.
$child->kill('SIG...')->join();
The following is a parallel demonstration comparing
"MCE::Shared" against
"Redis" and
"Redis::Fast" on a Fedora 23 VM.
Joining begins after all workers have been notified to quit.
use Time::HiRes qw(time);
use Redis;
use Redis::Fast;
use MCE::Child;
use MCE::Shared;
my $redis = Redis->new();
my $rfast = Redis::Fast->new();
my $array = MCE::Shared->array();
sub parallel_redis {
my ($_redis) = @_;
my ($count, $quit, $len) = (0, 0);
# instead, use a flag to exit loop
$SIG{'QUIT'} = sub { $quit = 1 };
while () {
$len = $_redis->rpush('list', $count++);
last if $quit;
}
$count;
}
sub parallel_array {
my ($count, $quit, $len) = (0, 0);
# do not exit from inside handler
$SIG{'QUIT'} = sub { $quit = 1 };
while () {
$len = $array->push($count++);
last if $quit;
}
$count;
}
sub benchmark_this {
my ($desc, $num_procs, $timeout, $code, @args) = @_;
my ($start, $total) = (time(), 0);
MCE::Child->new($code, @args) for 1..$num_procs;
sleep $timeout;
# joining is not immediate; ok
$_->kill('QUIT') for MCE::Child->list();
# joining later; ok
$total += $_->join() for MCE::Child->list();
printf "$desc <> duration: %0.03f secs, count: $total\n",
time() - $start;
sleep 0.2;
}
benchmark_this('Redis ', 8, 5.0, \¶llel_redis, $redis);
benchmark_this('Redis::Fast', 8, 5.0, \¶llel_redis, $rfast);
benchmark_this('MCE::Shared', 8, 5.0, \¶llel_array);
- MCE::Child->list()
- Returns a list of all child objects not yet joined.
@procs = MCE::Child->list();
- MCE::Child->list_pids()
- Returns a list of all child pids not yet joined (available since 1.849).
@pids = MCE::Child->list_pids();
$SIG{INT} = $SIG{HUP} = $SIG{TERM} = sub {
# Signal workers all at once
CORE::kill('KILL', MCE::Child->list_pids());
exec('reset');
};
- MCE::Child->list_running()
- Returns a list of all child objects that are still running.
@procs = MCE::Child->list_running();
- MCE::Child->list_joinable()
- Returns a list of all child objects that have completed running. Thus,
ready to be joined without blocking.
@procs = MCE::Child->list_joinable();
- MCE::Child->max_workers([ N ])
- Getter and setter for max_workers. Specify a number or 'auto' to acquire
the total number of cores via MCE::Util::get_ncpu. Specify a false value
to set back to no limit.
- MCE::Child->pending()
- Returns a count of all child objects not yet joined.
$count = MCE::Child->pending();
- $child->result()
- Returns the result obtained by "join",
"wait_one", or
"wait_all". If the process has not yet
exited, waits for the corresponding child to complete its execution.
use MCE::Child;
use Time::HiRes qw(sleep);
sub task {
my ($id) = @_;
sleep $id * 0.333;
return $id;
}
MCE::Child->create('task', $_) for ( reverse 1 .. 3 );
# 1 while MCE::Child->wait_one();
while ( my $child = MCE::Child->wait_one() ) {
my $err = $child->error() || 'no error';
my $res = $child->result();
my $pid = $child->pid();
print "[$pid] $err : $res\n";
}
Like "join" described above,
the context (void, scalar or list) for the return value(s) is determined
at the time "result" is called and
mostly "wantarray" aware.
my $child1 = MCE::Child->create( sub {
my @res = qw(foo bar baz);
return (@res);
});
my @res1 = $child1->result(); # ( foo, bar, baz )
my $res1 = $child1->result(); # baz
my $child2 = MCE::Child->create( sub {
return 'foo';
});
my @res2 = $child2->result(); # ( foo )
my $res2 = $child2->result(); # foo
- MCE::Child->self()
- Class method that allows a child to obtain it's own MCE::Child
object.
- $child->pid()
- $child->tid()
- Returns the ID of the child.
pid: $$ process id
tid: $$ alias for pid
- MCE::Child->pid()
- MCE::Child->tid()
- Class methods that allows a child to obtain its own ID.
pid: $$ process id
tid: $$ alias for pid
- MCE::Child->wait_one()
- MCE::Child->waitone()
- MCE::Child->wait_all()
- MCE::Child->waitall()
- Meaningful for the manager process only, waits for one or all child
processes to complete execution. Afterwards, returns the corresponding
child objects. If a child doesn't exist, returns the
"undef" value or an empty list for
"wait_one" and
"wait_all" respectively.
The "waitone" and
"waitall" methods are aliases for
compatibility with "MCE::Hobo".
use MCE::Child;
use Time::HiRes qw(sleep);
sub task {
my $id = shift;
sleep $id * 0.333;
return $id;
}
MCE::Child->create('task', $_) for ( reverse 1 .. 3 );
# join, traditional use case
$_->join() for MCE::Child->list();
# wait_one, simplistic use case
1 while MCE::Child->wait_one();
# wait_one
while ( my $child = MCE::Child->wait_one() ) {
my $err = $child->error() || 'no error';
my $res = $child->result();
my $pid = $child->pid();
print "[$pid] $err : $res\n";
}
# wait_all
my @procs = MCE::Child->wait_all();
for ( @procs ) {
my $err = $_->error() || 'no error';
my $res = $_->result();
my $pid = $_->pid();
print "[$pid] $err : $res\n";
}
- MCE::Child->yield( [ floating_seconds ] )
- Give other workers a chance to run, optionally for given time. Yield
behaves similarly to MCE's interval option. It throttles workers from
running too fast. A demonstration is provided in the next section for
fetching URLs in parallel.
The default
"floating_seconds" is 0.008 and 0.015
on UNIX and Windows, respectively. Pass 0 if simply wanting to give
other workers a chance to run.
# total run time: 1.00 second
MCE::Child->create( sub { MCE::Child->yield(0.25) } ) for 1 .. 4;
MCE::Child->wait_all();
Threads-like detach capability was added starting with the 1.867 release.
A threads example is shown first followed by the MCE::Child
example. All one needs to do is set the CHLD signal handler to IGNORE.
Unfortunately, this works on UNIX platforms only. The child process restores
the CHLD handler to default, so is able to deeply spin workers and reap if
desired.
use threads;
for ( 1 .. 8 ) {
async {
# do something
}->detach;
}
use MCE::Child;
# Have the OS reap workers automatically when exiting.
# The on_finish option is ignored if specified (no-op).
# Ensure not inside a thread on UNIX platforms.
$SIG{CHLD} = 'IGNORE';
for ( 1 .. 8 ) {
mce_child {
# do something
};
}
# Optionally, wait for any remaining workers before leaving.
# This is necessary if workers are consuming shared objects,
# constructed via MCE::Shared.
MCE::Child->wait_all;
The following is another way and works on Windows. Here, the
on_finish handler works as usual.
use MCE::Child;
MCE::Child->init(
on_finish = sub {
...
},
);
for ( 1 .. 8 ) {
$_->join for MCE::Child->list_joinable;
mce_child {
# do something
};
}
MCE::Child->wait_all;
MCE::Child behaves similarly to threads for the most part. It also provides
Parallel::ForkManager-like capabilities. The
"Parallel::ForkManager" example is shown
first followed by a version using
"MCE::Child".
- Parallel::ForkManager
-
use strict;
use warnings;
use Parallel::ForkManager;
use Time::HiRes 'time';
my $start = time;
my $pm = Parallel::ForkManager->new(10);
$pm->set_waitpid_blocking_sleep(0);
$pm->run_on_finish( sub {
my ($pid, $exit_code, $ident, $exit_signal, $core_dumped, $resp) = @_;
print "child $pid completed: $ident => ", $resp->[0], "\n";
});
DATA_LOOP:
foreach my $data ( 1..2000 ) {
# forks and returns the pid for the child
my $pid = $pm->start($data) and next DATA_LOOP;
my $ret = [ $data * 2 ];
$pm->finish(0, $ret);
}
$pm->wait_all_children;
printf STDERR "duration: %0.03f seconds\n", time - $start;
- MCE::Child
-
use strict;
use warnings;
use MCE::Child 1.843;
use Time::HiRes 'time';
my $start = time;
MCE::Child->init(
max_workers => 10,
on_finish => sub {
my ($pid, $exit_code, $ident, $exit_signal, $error, $resp) = @_;
print "child $pid completed: $ident => ", $resp->[0], "\n";
}
);
foreach my $data ( 1..2000 ) {
MCE::Child->create( $data, sub {
[ $data * 2 ];
});
}
MCE::Child->wait_all;
printf STDERR "duration: %0.03f seconds\n", time - $start;
- Time to spin 2,000 workers and obtain results (in seconds).
- Results were obtained on a Macbook Pro (2.6 GHz ~ 3.6 GHz with Turbo
Boost). Parallel::ForkManager 2.02 uses Moo. Therefore, I ran again with
Moo loaded at the top of the script.
MCE::Hobo uses MCE::Shared to retrieve data during reaping.
MCE::Child uses MCE::Channel, no shared-manager.
Version Cygwin Windows Linux macOS FreeBSD
MCE::Child 1.843 19.099s 17.091s 0.965s 1.534s 1.229s
MCE::Hobo 1.843 20.514s 19.594s 1.246s 1.629s 1.613s
P::FM 1.20 19.703s 19.235s 0.875s 1.445s 1.346s
MCE::Child 1.843 20.426s 18.417s 1.116s 1.632s 1.338s Moo loaded
MCE::Hobo 1.843 21.809s 20.810s 1.407s 1.759s 1.722s Moo loaded
P::FM 2.02 21.668s 25.927s 1.882s 2.612s 2.483s Moo used
- Set posix_exit to avoid all END and destructor processing.
- This is helpful for reducing overhead when workers exit. Ditto if using a
Perl module not parallel safe. The option is ignored on Windows
"$^O eq 'MSWin32'".
MCE::Child->init( posix_exit => 1, ... );
MCE::Hobo->init( posix_exit => 1, ... );
Version Cygwin Windows Linux macOS FreeBSD
MCE::Child 1.843 19.815s ignored 0.824s 1.284s 1.245s Moo loaded
MCE::Hobo 1.843 21.029s ignored 0.953s 1.335s 1.439s Moo loaded
This demonstration constructs two queues, two handles, starts the shared-manager
process if needed, and spawns four workers. For this demonstration, am
chunking 64 URLs per job. In reality, one may run with 200 workers and chunk
300 URLs on a 24-way box.
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
# perl demo.pl -- all output
# perl demo.pl >/dev/null -- mngr/child output
# perl demo.pl 2>/dev/null -- show results only
#
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
use strict;
use warnings;
use AnyEvent;
use AnyEvent::HTTP;
use Time::HiRes qw( time );
use MCE::Child;
use MCE::Shared;
# Construct two queues, input and return.
my $que = MCE::Shared->queue();
my $ret = MCE::Shared->queue();
# Construct shared handles for serializing output from many workers
# writing simultaneously. This prevents garbled output.
mce_open my $OUT, ">>", \*STDOUT or die "open error: $!";
mce_open my $ERR, ">>", \*STDERR or die "open error: $!";
# Spawn workers early for minimum memory consumption.
MCE::Child->create({ posix_exit => 1 }, 'task', $_) for 1 .. 4;
# Obtain or generate input data for workers to process.
my ( $count, @urls ) = ( 0 );
push @urls, map { "http://127.0.0.$_/" } 1..254;
push @urls, map { "http://192.168.0.$_/" } 1..254; # 508 URLs total
while ( @urls ) {
my @chunk = splice(@urls, 0, 64);
$que->enqueue( { ID => ++$count, INPUT => \@chunk } );
}
# So that workers leave the loop after consuming the queue.
$que->end();
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
# Loop for the manager process. The manager may do other work if
# need be and periodically check $ret->pending() not shown here.
#
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
my $start = time;
printf {$ERR} "Mngr - entering loop\n";
while ( $count ) {
my ( $result, $failed ) = $ret->dequeue( 2 );
# Remove ID from result, so not treated as a URL item.
printf {$ERR} "Mngr - received job %s\n", delete $result->{ID};
# Display the URL and the size captured.
foreach my $url ( keys %{ $result } ) {
printf {$OUT} "%s: %d\n", $url, length($result->{$url})
if $result->{$url}; # url has content
}
# Display URLs could not reach.
if ( @{ $failed } ) {
foreach my $url ( @{ $failed } ) {
print {$OUT} "Failed: $url\n";
}
}
# Decrement the count.
$count--;
}
MCE::Child->wait_all();
printf {$ERR} "Mngr - exiting loop\n\n";
printf {$ERR} "Duration: %0.3f seconds\n\n", time - $start;
exit;
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
# Child processes enqueue two items ( $result and $failed ) per each
# job for the manager process. Likewise, the manager process dequeues
# two items above. Optionally, child processes may include the ID in
# the result.
#
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
sub task {
my ( $id ) = @_;
printf {$ERR} "Child $id entering loop\n";
while ( my $job = $que->dequeue() ) {
my ( $result, $failed ) = ( { ID => $job->{ID} }, [ ] );
# Walk URLs, provide a hash and array refs for data.
printf {$ERR} "Child $id running job $job->{ID}\n";
walk( $job, $result, $failed );
# Send results to the manager process.
$ret->enqueue( $result, $failed );
}
printf {$ERR} "Child $id exiting loop\n";
}
sub walk {
my ( $job, $result, $failed ) = @_;
# Yielding is critical when running an event loop in parallel.
# Not doing so means that the app may reach contention points
# with the firewall and likely impose unnecessary hardship at
# the OS level. The idea here is not to have multiple workers
# initiate HTTP requests to a batch of URLs at the same time.
# Yielding behaves similarly like scatter to have the child
# process run solo for a fraction of time.
MCE::Child->yield( 0.03 );
my $cv = AnyEvent->condvar();
# Populate the hash ref for the URLs it could reach.
# Do not mix AnyEvent timeout with child timeout.
# Therefore, choose event timeout when available.
foreach my $url ( @{ $job->{INPUT} } ) {
$cv->begin();
http_get $url, timeout => 2, sub {
my ( $data, $headers ) = @_;
$result->{$url} = $data;
$cv->end();
};
}
$cv->recv();
# Populate the array ref for URLs it could not reach.
foreach my $url ( @{ $job->{INPUT} } ) {
push @{ $failed }, $url unless (exists $result->{ $url });
}
return;
}
__END__
$ perl demo.pl
Child 1 entering loop
Child 2 entering loop
Child 3 entering loop
Mngr - entering loop
Child 2 running job 2
Child 3 running job 3
Child 1 running job 1
Child 4 entering loop
Child 4 running job 4
Child 2 running job 5
Mngr - received job 2
Child 3 running job 6
Mngr - received job 3
Child 1 running job 7
Mngr - received job 1
Child 4 running job 8
Mngr - received job 4
http://192.168.0.1/: 3729
Child 2 exiting loop
Mngr - received job 5
Child 3 exiting loop
Mngr - received job 6
Child 1 exiting loop
Mngr - received job 7
Child 4 exiting loop
Mngr - received job 8
Mngr - exiting loop
Duration: 4.131 seconds
Making an executable is possible with the PAR::Packer module. On the Windows
platform, threads, threads::shared, and exiting via threads are necessary for
the binary to exit successfully.
# https://metacpan.org/pod/PAR::Packer
# https://metacpan.org/pod/pp
#
# pp -o demo.exe demo.pl
# ./demo.exe
use strict;
use warnings;
use if $^O eq "MSWin32", "threads";
use if $^O eq "MSWin32", "threads::shared";
# Include minimum dependencies for MCE::Child.
# Add other modules required by your application here.
use Storable ();
use Time::HiRes ();
# use IO::FDPass (); # optional: for condvar, handle, queue
# use Sereal (); # optional: for faster serialization
use MCE::Child;
use MCE::Shared;
# For PAR to work on the Windows platform, one must include manually
# any shared modules used by the application.
# use MCE::Shared::Array; # if using MCE::Shared->array
# use MCE::Shared::Cache; # if using MCE::Shared->cache
# use MCE::Shared::Condvar; # if using MCE::Shared->condvar
# use MCE::Shared::Handle; # if using MCE::Shared->handle, mce_open
# use MCE::Shared::Hash; # if using MCE::Shared->hash
# use MCE::Shared::Minidb; # if using MCE::Shared->minidb
# use MCE::Shared::Ordhash; # if using MCE::Shared->ordhash
# use MCE::Shared::Queue; # if using MCE::Shared->queue
# use MCE::Shared::Scalar; # if using MCE::Shared->scalar
# Et cetera. Only load modules needed for your application.
use MCE::Shared::Sequence; # if using MCE::Shared->sequence
my $seq = MCE::Shared->sequence( 1, 9 );
sub task {
my ( $id ) = @_;
while ( defined ( my $num = $seq->next() ) ) {
print "$id: $num\n";
sleep 1;
}
}
sub main {
MCE::Child->new( \&task, $_ ) for 1 .. 3;
MCE::Child->wait_all();
}
# Main must run inside a thread on the Windows platform or workers
# will fail duing exiting, causing the exe to crash. The reason is
# that PAR or a dependency isn't multi-process safe.
( $^O eq "MSWin32" ) ? threads->create(\&main)->join() : main();
threads->exit(0) if $INC{"threads.pm"};
MCE::Child emits an error when "is_joinable",
"is_running", and
"join" isn't called by the managed process,
where the child was spawned. This is a limitation in MCE::Child only due to
not involving a shared-manager process for IPC.
This use-case is not typical.
The inspiration for "MCE::Child" comes from
wanting "threads"-like behavior for
processes compatible with Perl 5.8. Both can run side-by-side including
safe-use by MCE workers. Likewise, the documentation resembles
"threads".
The inspiration for "wait_all"
and "wait_one" comes from the
"Parallel::WorkUnit" module.
- forks
- forks::BerkeleyDB
- MCE::Hobo
- Parallel::ForkManager
- Parallel::Loops
- Parallel::Prefork
- Parallel::WorkUnit
- Proc::Fork
- Thread::Tie
- threads
MCE, MCE::Channel, MCE::Shared
Mario E. Roy, <marioeroy AT gmail DOT com>
Visit the GSP FreeBSD Man Page Interface. Output converted with ManDoc. |