Provided by: libanyevent-fork-perl_1.32-1_all bug

NAME

       AnyEvent::Fork - everything you wanted to use fork() for, but couldn't

SYNOPSIS

          use AnyEvent::Fork;

          AnyEvent::Fork
             ->new
             ->require ("MyModule")
             ->run ("MyModule::server", my $cv = AE::cv);

          my $fh = $cv->recv;

DESCRIPTION

       This module allows you to create new processes, without actually forking them from your current process
       (avoiding the problems of forking), but preserving most of the advantages of fork.

       It can be used to create new worker processes or new independent subprocesses for short- and long-running
       jobs, process pools (e.g. for use in pre-forked servers) but also to spawn new external processes (such
       as CGI scripts from a web server), which can be faster (and more well behaved) than using fork+exec in
       big processes.

       Special care has been taken to make this module useful from other modules, while still supporting
       specialised environments such as App::Staticperl or PAR::Packer.

   WHAT THIS MODULE IS NOT
       This module only creates processes and lets you pass file handles and strings to it, and run perl code.
       It does not implement any kind of RPC - there is no back channel from the process back to you, and there
       is no RPC or message passing going on.

       If you need some form of RPC, you could use the AnyEvent::Fork::RPC companion module, which adds simple
       RPC/job queueing to a process created by this module.

       And if you need some automatic process pool management on top of AnyEvent::Fork::RPC, you can look at the
       AnyEvent::Fork::Pool companion module.

       Or you can implement it yourself in whatever way you like: use some message-passing module such as
       AnyEvent::MP, some pipe such as AnyEvent::ZeroMQ, use AnyEvent::Handle on both sides to send e.g. JSON or
       Storable messages, and so on.

   COMPARISON TO OTHER MODULES
       There is an abundance of modules on CPAN that do "something fork", such as Parallel::ForkManager,
       AnyEvent::ForkManager, AnyEvent::Worker or AnyEvent::Subprocess. There are modules that implement their
       own process management, such as AnyEvent::DBI.

       The problems that all these modules try to solve are real, however, none of them (from what I have seen)
       tackle the very real problems of unwanted memory sharing, efficiency or not being able to use event
       processing, GUI toolkits or similar modules in the processes they create.

       This module doesn't try to replace any of them - instead it tries to solve the problem of creating
       processes with a minimum of fuss and overhead (and also luxury). Ideally, most of these would use
       AnyEvent::Fork internally, except they were written before AnyEvent:Fork was available, so obviously had
       to roll their own.

   PROBLEM STATEMENT
       There are two traditional ways to implement parallel processing on UNIX like operating systems - fork and
       process, and fork+exec and process. They have different advantages and disadvantages that I describe
       below, together with how this module tries to mitigate the disadvantages.

       Forking from a big process can be very slow.
           A  5GB  process  needs  0.05s to fork on my 3.6GHz amd64 GNU/Linux box. This overhead is often shared
           with exec (because you have to fork first), but in some circumstances  (e.g.  when  vfork  is  used),
           fork+exec can be much faster.

           This module can help here by telling a small(er) helper process to fork, which is faster then forking
           the  main  process,  and  also  uses  vfork  where  possible. This gives the speed of vfork, with the
           flexibility of fork.

       Forking usually creates a copy-on-write copy of the parent process.
           For example, modules or data files that are loaded will not  use  additional  memory  after  a  fork.
           Exec'ing  a  new process, in contrast, means modules and data files might need to be loaded again, at
           extra CPU and memory cost.

           But when forking, you still create a copy of your data structures - if the  program  frees  them  and
           replaces  them  by  new  data, the child processes will retain the old version even if it isn't used,
           which can suddenly and unexpectedly increase memory usage when freeing memory.

           For example, Gtk2::CV is an image viewer optimised for large directories (millions of  pictures).  It
           also  forks  subprocesses  for thumbnail generation, which inherit the data structure that stores all
           file information. If the user changes the directory, it gets freed in the  main  process,  leaving  a
           copy  in  the thumbnailer processes. This can lead to many times the memory usage that would actually
           be required. The  solution  is  to  fork  early  (and  being  unable  to  dynamically  generate  more
           subprocesses or do this from a module)... or to use <AnyEvent:Fork>.

           There  is  a trade-off between more sharing with fork (which can be good or bad), and no sharing with
           exec.

           This module allows the main program to do a controlled fork, and allows  modules  to  exec  processes
           safely  at  any  time. When creating a custom process pool you can take advantage of data sharing via
           fork without risking to share large dynamic data structures that will blow up child memory usage.

           In other words, this module puts you into control over what is being shared and what  isn't,  at  all
           times.

       Exec'ing a new perl process might be difficult.
           For  example,  it  is  not easy to find the correct path to the perl interpreter - $^X might not be a
           perl interpreter at all. Worse, there might not even be a perl binary installed on the system.

           This module tries hard to identify the correct path to the perl interpreter. With a cooperative  main
           program,  exec'ing  the  interpreter might not even be necessary, but even without help from the main
           program, it will still work when used from a module.

       Exec'ing a new perl process might be slow, as all necessary modules have to be loaded from disk again,
       with no guarantees of success.
           Long running processes might run into problems when perl  is  upgraded  and  modules  are  no  longer
           loadable  because  they  refer to a different perl version, or parts of a distribution are newer than
           the ones already loaded.

           This module supports creating pre-initialised perl processes  to  be  used  as  a  template  for  new
           processes at a later time, e.g. for use in a process pool.

       Forking might be impossible when a program is running.
           For  example,  POSIX  makes  it  almost  impossible to fork from a multi-threaded program while doing
           anything useful in the child - in fact, if your perl program uses POSIX threads (even indirectly  via
           e.g.  IO::AIO  or  threads),  you  cannot  call fork on the perl level anymore without risking memory
           corruption or worse on a number of operating systems.

           This module can safely fork helper processes at any time, by calling fork+exec  in  C,  in  a  POSIX-
           compatible way (via Proc::FastSpawn).

       Parallel processing with fork might be inconvenient or difficult to implement. Modules might not work in
       both parent and child.
           For  example,  when a program uses an event loop and creates watchers it becomes very hard to use the
           event loop from a child program, as the watchers already exist but are only meaningful in the parent.
           Worse, a module might want to use such a module, not knowing  whether  another  module  or  the  main
           program also does, leading to problems.

           Apart  from  event loops, graphical toolkits also commonly fall into the "unsafe module" category, or
           just about anything that communicates with the external world, such as network libraries and file I/O
           modules, which usually don't like being copied and then allowed to continue in two processes.

           With this module only the main program is allowed to create new processes by  forking  (because  only
           the  main  program  can  know  when  it is still safe to do so) - all other processes are created via
           fork+exec, which makes it possible to use modules such as event loops or window interfaces safely.

EXAMPLES

       This is where the wall of text ends and code speaks.

   Create a single new process, tell it to run your worker function.
          AnyEvent::Fork
             ->new
             ->require ("MyModule")
             ->run ("MyModule::worker, sub {
                my ($master_filehandle) = @_;

                # now $master_filehandle is connected to the
                # $slave_filehandle in the new process.
             });

       "MyModule" might look like this:

          package MyModule;

          sub worker {
             my ($slave_filehandle) = @_;

             # now $slave_filehandle is connected to the $master_filehandle
             # in the original process. have fun!
          }

   Create a pool of server processes all accepting on the same socket.
          # create listener socket
          my $listener = ...;

          # create a pool template, initialise it and give it the socket
          my $pool = AnyEvent::Fork
                        ->new
                        ->require ("Some::Stuff", "My::Server")
                        ->send_fh ($listener);

          # now create 10 identical workers
          for my $id (1..10) {
             $pool
                ->fork
                ->send_arg ($id)
                ->run ("My::Server::run");
          }

          # now do other things - maybe use the filehandle provided by run
          # to wait for the processes to die. or whatever.

       "My::Server" might look like this:

          package My::Server;

          sub run {
             my ($slave, $listener, $id) = @_;

             close $slave; # we do not use the socket, so close it to save resources

             # we could go ballistic and use e.g. AnyEvent here, or IO::AIO,
             # or anything we usually couldn't do in a process forked normally.
             while (my $socket = $listener->accept) {
                # do sth. with new socket
             }
          }

   use AnyEvent::Fork as a faster fork+exec
       This runs "/bin/echo hi", with standard output redirected to /tmp/log and standard  error  redirected  to
       the  communications  socket.  It  is  usually  faster  than  fork+exec,  but  still  lets you prepare the
       environment.

          open my $output, ">/tmp/log" or die "$!";

          AnyEvent::Fork
             ->new
             ->eval ('
                  # compile a helper function for later use
                  sub run {
                     my ($fh, $output, @cmd) = @_;

                     # perl will clear close-on-exec on STDOUT/STDERR
                     open STDOUT, ">&", $output or die;
                     open STDERR, ">&", $fh or die;

                     exec @cmd;
                  }
               ')
             ->send_fh ($output)
             ->send_arg ("/bin/echo", "hi")
             ->run ("run", my $cv = AE::cv);

          my $stderr = $cv->recv;

   For stingy users: put the worker code into a "DATA" section.
       When you want to be stingy with files, you can put your code into the "DATA" section of your  module  (or
       program):

          use AnyEvent::Fork;

          AnyEvent::Fork
             ->new
             ->eval (do { local $/; <DATA> })
             ->run ("doit", sub { ... });

          __DATA__

          sub doit {
             ... do something!
          }

   For stingy standalone programs: do not rely on external files at all.
       For  single-file  scripts  it  can  be  inconvenient to rely on external files - even when using a "DATA"
       section, you still need to "exec" an external perl interpreter, which might not be available  when  using
       App::Staticperl, Urlader or PAR::Packer for example.

       Two  modules  help  here  -  AnyEvent::Fork::Early  forks  a  template  process  for all further calls to
       "new_exec", and AnyEvent::Fork::Template forks the main program as a template process.

       Here is how your main program should look like:

          #! perl

          # optional, as the very first thing.
          # in case modules want to create their own processes.
          use AnyEvent::Fork::Early;

          # next, load all modules you need in your template process
          use Example::My::Module
          use Example::Whatever;

          # next, put your run function definition and anything else you
          # need, but do not use code outside of BEGIN blocks.
          sub worker_run {
             my ($fh, @args) = @_;
             ...
          }

          # now preserve everything so far as AnyEvent::Fork object
          # in $TEMPLATE.
          use AnyEvent::Fork::Template;

          # do not put code outside of BEGIN blocks until here

          # now use the $TEMPLATE process in any way you like

          # for example: create 10 worker processes
          my @worker;
          my $cv = AE::cv;
          for (1..10) {
             $cv->begin;
             $TEMPLATE->fork->send_arg ($_)->run ("worker_run", sub {
                push @worker, shift;
                $cv->end;
             });
          }
          $cv->recv;

CONCEPTS

       This module can create new processes either by executing a new  perl  process,  or  by  forking  from  an
       existing "template" process.

       All  these  processes  are  called "child processes" (whether they are direct children or not), while the
       process that manages them is called the "parent process".

       Each such process comes with its own file handle that can be used to communicate with it (it's actually a
       socket - one end in the new process, one end in the main process), and among the things you can do in  it
       are load modules, fork new processes, send file handles to it, and execute functions.

       There are multiple ways to create additional processes to execute some jobs:

       fork a new process from the "default" template process, load code, run it
           This  module  has  a  "default"  template process which it executes when it is needed the first time.
           Forking from this process shares the memory used for the perl interpreter with the new  process,  but
           loading modules takes time, and the memory is not shared with anything else.

           This  is  ideal  for  when you only need one extra process of a kind, with the option of starting and
           stopping it on demand.

           Example:

              AnyEvent::Fork
                 ->new
                 ->require ("Some::Module")
                 ->run ("Some::Module::run", sub {
                    my ($fork_fh) = @_;
                 });

       fork a new template process, load code, then fork processes off of it and run the code
           When you need to have a bunch of processes that all execute the same (or very similar) tasks, then  a
           good  way  is  to  create a new template process for them, loading all the modules you need, and then
           create your worker processes from this new template process.

           This way, all code (and data structures) that can be shared (e.g. the modules you loaded)  is  shared
           between the processes, and each new process consumes relatively little memory of its own.

           The  disadvantage of this approach is that you need to create a template process for the sole purpose
           of forking new processes from it, but if you only need a fixed number of  processes  you  can  create
           them, and then destroy the template process.

           Example:

              my $template = AnyEvent::Fork->new->require ("Some::Module");

              for (1..10) {
                 $template->fork->run ("Some::Module::run", sub {
                    my ($fork_fh) = @_;
                 });
              }

              # at this point, you can keep $template around to fork new processes
              # later, or you can destroy it, which causes it to vanish.

       execute a new perl interpreter, load some code, run it
           This is relatively slow, and doesn't allow you to share memory between multiple processes.

           The  only  advantage is that you don't have to have a template process hanging around all the time to
           fork off some new processes, which might be an advantage when there are  long  time  spans  where  no
           extra processes are needed.

           Example:

              AnyEvent::Fork
                 ->new_exec
                 ->require ("Some::Module")
                 ->run ("Some::Module::run", sub {
                    my ($fork_fh) = @_;
                 });

THE "AnyEvent::Fork" CLASS

       This module exports nothing, and only implements a single class - "AnyEvent::Fork".

       There are two class constructors that both create new processes - "new" and "new_exec". The "fork" method
       creates a new process by forking an existing one and could be considered a third constructor.

       Most  of  the remaining methods deal with preparing the new process, by loading code, evaluating code and
       sending data to the new process. They usually return the process object, so you can chain method calls.

       If a process object is destroyed before calling its "run" method, then the process  simply  exits.  After
       "run" is called, all responsibility is passed to the specified function.

       As  long as there is any outstanding work to be done, process objects resist being destroyed, so there is
       no reason to store them unless you need them later - configure and forget works just fine.

       my $proc = new AnyEvent::Fork
           Create  a  new  "empty"  perl  interpreter  process  and  returns  its  process  object  for  further
           manipulation.

           The  new  process  is  forked  from  a template process that is kept around for this purpose. When it
           doesn't exist yet, it is created by a call to "new_exec" first  and  then  stays  around  for  future
           calls.

       $new_proc = $proc->fork
           Forks $proc, creating a new process, and returns the process object of the new process.

           If  any of the "send_" functions have been called before fork, then they will be cloned in the child.
           For example, in a pre-forked server, you might "send_fh"  the  listening  socket  into  the  template
           process, and then keep calling "fork" and "run".

       my $proc = new_exec AnyEvent::Fork
           Create  a  new  "empty"  perl  interpreter  process  and  returns  its  process  object  for  further
           manipulation.

           Unlike the "new" method, this method always spawns a new perl process  (except  in  some  cases,  see
           AnyEvent::Fork::Early  for  details). This reduces the amount of memory sharing that is possible, and
           is also slower.

           You should use "new" whenever possible, except when having a template process around is unacceptable.

           The path to the perl interpreter is divined using various methods - first $^X is investigated to  see
           if  the  path  ends  with  something that looks as if it were the perl interpreter. Failing this, the
           module falls back to using $Config::Config{perlpath}.

           The path to perl can also be overridden by setting the global variable $AnyEvent::Fork::PERL  -  it's
           value will be used for all subsequent invocations.

       $pid = $proc->pid
           Returns the process id of the process iff it is a direct child of the process running AnyEvent::Fork,
           and  "undef"  otherwise.  As  a  general  rule  (that  you  cannot  rely upon), processes created via
           "new_exec", AnyEvent::Fork::Early or AnyEvent::Fork::Template are direct children,  while  all  other
           processes are not.

           Or  in other words, you do not normally have to take care of zombies for processes created via "new",
           but when in doubt, or zombies are a problem, you need to check whether a process is a diretc child by
           calling this method, and possibly creating a child watcher or reap it manually.

       $proc = $proc->eval ($perlcode, @args)
           Evaluates the given $perlcode as ... Perl code, while setting @_ to the strings specified  by  @args,
           in  the  "main"  package  (so  you  can access the args using $_[0] and so on, but not using implicit
           "shit" as the latter works on @ARGV).

           This call is meant to do any custom initialisation that might be required (for example, the "require"
           method uses it). It's not supposed to be used to completely take over  the  process,  use  "run"  for
           that.

           The  code will usually be executed after this call returns, and there is no way to pass anything back
           to the calling process. Any evaluation errors will be reported to stderr and  cause  the  process  to
           exit.

           If  you  want  to  execute  some  code  (that isn't in a module) to take over the process, you should
           compile a function via "eval" first, and then call it via "run". This also gives you  access  to  any
           arguments  passed  via the "send_xxx" methods, such as file handles. See the "use AnyEvent::Fork as a
           faster fork+exec" example to see it in action.

           Returns the process object for easy chaining of method calls.

           It's common to want to call an iniitalisation function with some arguments. Make  sure  you  actually
           pass  @_  to  that  function  (for example by using &name syntax), and do not just specify a function
           name:

              $proc->eval ('&MyModule::init', $string1, $string2);

       $proc = $proc->require ($module, ...)
           Tries to load the given module(s) into the process

           Returns the process object for easy chaining of method calls.

       $proc = $proc->send_fh ($handle, ...)
           Send one or more file handles (not file descriptors) to the process, to prepare a call to "run".

           The process object keeps a reference to the handles until they have been passed over to the  process,
           so  you must not explicitly close the handles. This is most easily accomplished by simply not storing
           the file handles anywhere after passing them to this method - when AnyEvent::Fork is  finished  using
           them, perl will automatically close them.

           Returns the process object for easy chaining of method calls.

           Example:  pass  a  file  handle  to  a  process,  and  release  it without closing. It will be closed
           automatically when it is no longer used.

              $proc->send_fh ($my_fh);
              undef $my_fh; # free the reference if you want, but DO NOT CLOSE IT

       $proc = $proc->send_arg ($string, ...)
           Send one or more argument strings to the process, to prepare a call to "run". The strings can be  any
           octet strings.

           The  protocol is optimised to pass a moderate number of relatively short strings - while you can pass
           up to 4GB of data in one go, this is more meant to pass some ID information or  other  startup  info,
           not big chunks of data.

           Returns the process object for easy chaining of method calls.

       $proc->run ($func, $cb->($fh))
           Enter  the  function  specified  by the function name in $func in the process. The function is called
           with the communication socket as first argument, followed by all file handles  and  string  arguments
           sent earlier via "send_fh" and "send_arg" methods, in the order they were called.

           The process object becomes unusable on return from this function - any further method calls result in
           undefined behaviour.

           The  function  name  should  be  fully qualified, but if it isn't, it will be looked up in the "main"
           package.

           If the called function returns, doesn't exist, or any error occurs, the process exits.

           Preparing the process is done in the background - when all commands have been sent, the  callback  is
           invoked  with  the  local  communications  socket  as argument. At this point you can start using the
           socket in any way you like.

           If the communication socket isn't used, it should be closed on both sides, to save on kernel memory.

           The socket is non-blocking in the parent, and blocking in the newly created  process.  The  close-on-
           exec flag is set in both.

           Even  if not used otherwise, the socket can be a good indicator for the existence of the process - if
           the other process exits, you get a readable event on it,  because  exiting  the  process  closes  the
           socket (if it didn't create any children using fork).

           Compatibility to AnyEvent::Fork::Remote
               If  you  want to write code that works with both this module and AnyEvent::Fork::Remote, you need
               to write your code so that it assumes there are two file handles for communications, which  might
               not be unix domain sockets. The "run" function should start like this:

                  sub run {
                     my ($rfh, @args) = @_; # @args is your normal arguments
                     my $wfh = fileno $rfh ? $rfh : *STDOUT;

                     # now use $rfh for reading and $wfh for writing
                  }

               This  checks  whether  the  passed file handle is, in fact, the process "STDIN" handle. If it is,
               then the function was invoked visa AnyEvent::Fork::Remote, so STDIN should be  used  for  reading
               and "STDOUT" should be used for writing.

               In  all  other  cases, the function was called via this module, and there is only one file handle
               that should be sued for reading and writing.

           Example: create a template for a process pool, pass a few strings, some file handles, then fork, pass
           one more string, and run some code.

              my $pool = AnyEvent::Fork
                            ->new
                            ->send_arg ("str1", "str2")
                            ->send_fh ($fh1, $fh2);

              for (1..2) {
                 $pool
                    ->fork
                    ->send_arg ("str3")
                    ->run ("Some::function", sub {
                       my ($fh) = @_;

                       # fh is nonblocking, but we trust that the OS can accept these
                       # few octets anyway.
                       syswrite $fh, "hi #$_\n";

                       # $fh is being closed here, as we don't store it anywhere
                    });
              }

              # Some::function might look like this - all parameters passed before fork
              # and after will be passed, in order, after the communications socket.
              sub Some::function {
                 my ($fh, $str1, $str2, $fh1, $fh2, $str3) = @_;

                 print scalar <$fh>; # prints "hi #1\n" and "hi #2\n" in any order
              }

   CHILD PROCESS INTERFACE
       This module has a limited API for use in child processes.

       @args = AnyEvent::Fork::Serve::run_args
           This function, which only exists before the "run" method is called, returns the arguments that  would
           be passed to the run function, and clears them.

           This  is  mainly  useful  to  get  any file handles passed via "send_fh", but works for any arguments
           passed via "send_xxx" methods.

   EXPERIMENTAL METHODS
       These methods might go away completely or change behaviour, at any time.

       $proc->to_fh ($cb->($fh))    # EXPERIMENTAL, MIGHT BE REMOVED
           Flushes all commands out to the process and then calls the callback with the communications socket.

           The process object becomes unusable on return from this function - any further method calls result in
           undefined behaviour.

           The point of this method is to give you a file handle that you can pass to another process.  In  that
           other  process, you can call "new_from_fh AnyEvent::Fork $fh" to create a new "AnyEvent::Fork" object
           from it, thereby effectively passing a fork object to another process.

       new_from_fh AnyEvent::Fork $fh    # EXPERIMENTAL, MIGHT BE REMOVED
           Takes a file handle originally rceeived by the "to_fh"  method  and  creates  a  new  "AnyEvent:Fork"
           object.  The child process itself will not change in any way, i.e. it will keep all the modifications
           done to it before calling "to_fh".

           The new object is very much like the original object,  except  that  the  "pid"  method  will  return
           "undef" even if the process is a direct child.

PERFORMANCE

       Now  for  some unscientific benchmark numbers (all done on an amd64 GNU/Linux box). These are intended to
       give you an idea of the relative  performance  you  can  expect,  they  are  not  meant  to  be  absolute
       performance numbers.

       OK, so, I ran a simple benchmark that creates a socket pair, forks, calls exit in the child and waits for
       the  socket  to close in the parent. I did load AnyEvent, EV and AnyEvent::Fork, for a total process size
       of 5100kB.

          2079 new processes per second, using manual socketpair + fork

       Then I did the same thing, but instead of calling fork, I called AnyEvent::Fork->new->run  ("CORE::exit")
       and  then again waited for the socket from the child to close on exit. This does the same thing as manual
       socket pair + fork, except that what is forked is the template process (2440kB), and the socket needs  to
       be passed to the server at the other end of the socket first.

          2307 new processes per second, using AnyEvent::Fork->new

       And  finally,  using  "new_exec"  instead  "new",  using  vforks+execs to exec a new perl interpreter and
       compile the small server each time, I get:

           479 vfork+execs per second, using AnyEvent::Fork->new_exec

       So how can "AnyEvent->new" be faster than a standard fork, even though it uses the same  operations,  but
       adds a lot of overhead?

       The  difference is simply the process size: forking the 5MB process takes so much longer than forking the
       2.5MB template process that the extra overhead is canceled out.

       If the benchmark process grows, the normal fork becomes even slower:

          1340 new processes, manual fork of a 20MB process
           731 new processes, manual fork of a 200MB process
           235 new processes, manual fork of a 2000MB process

       What that means (to me) is that I can use this module without having a  bad  conscience  because  of  the
       extra overhead required to start new processes.

TYPICAL PROBLEMS

       This section lists typical problems that remain. I hope by recognising them, most can be avoided.

       leaked file descriptors for exec'ed processes
           POSIX  systems  inherit  file  descriptors  by default when exec'ing a new process. While perl itself
           laudably sets the close-on-exec flags on new file handles, most C libraries don't care, and  even  if
           all cared, it's often not possible to set the flag in a race-free manner.

           That  means  some  file  descriptors can leak through. And since it isn't possible to know which file
           descriptors are "good" and "necessary" (or even to know which file descriptors are open), there is no
           good way to close the ones that might harm.

           As an example of what "harm" can  be  done  consider  a  web  server  that  accepts  connections  and
           afterwards  some  module  uses  AnyEvent::Fork  for the first time, causing it to fork and exec a new
           process, which might inherit the network socket. When the server closes the socket, it is still  open
           in  the  child  (which  doesn't  even know that) and the client might conclude that the connection is
           still fine.

           For the main program, there are multiple remedies available - AnyEvent::Fork::Early is one,  creating
           a  process  early  and  not  using  "new_exec" is another, as in both cases, the first process can be
           exec'ed well before many random file descriptors are open.

           In general, the solution for these kind of problems is to fix the libraries or the  code  that  leaks
           those file descriptors.

           Fortunately, most of these leaked descriptors do no harm, other than sitting on some resources.

       leaked file descriptors for fork'ed processes
           Normally, AnyEvent::Fork does start new processes by exec'ing them, which closes file descriptors not
           marked for being inherited.

           However,  AnyEvent::Fork::Early and AnyEvent::Fork::Template offer a way to create these processes by
           forking, and this leaks more file descriptors than  exec'ing  them,  as  there  is  no  way  to  mark
           descriptors as "close on fork".

           An  example  would be modules like EV, IO::AIO or Gtk2. Both create pipes for internal uses, and Gtk2
           might open a connection to the X server. EV and IO::AIO can deal  with  fork,  but  Gtk2  might  have
           trouble with a fork.

           The   solution  is  to  either  not  load  these  modules  before  use'ing  AnyEvent::Fork::Early  or
           AnyEvent::Fork::Template, or to  delay  initialising  them,  for  example,  by  calling  "init  Gtk2"
           manually.

       exiting calls object destructors
           This only applies to users of AnyEvent::Fork:Early and AnyEvent::Fork::Template, or when initialising
           code creates objects that reference external resources.

           When  a  process  created  by AnyEvent::Fork exits, it might do so by calling exit, or simply letting
           perl reach the end of the program. At which point Perl runs all destructors.

           Not all destructors are fork-safe - for example, an object that represents the  connection  to  an  X
           display  might  tell  the X server to free resources, which is inconvenient when the "real" object in
           the parent still needs to use them.

           This is obviously not a problem for AnyEvent::Fork::Early, as you used it as the  very  first  thing,
           right?

           It  is a problem for AnyEvent::Fork::Template though - and the solution is to not create objects with
           nontrivial destructors that might have an effect outside of Perl.

PORTABILITY NOTES

       Native win32 perls are somewhat supported (AnyEvent::Fork::Early is a nop, and ::Template is not going to
       work), and it cost a lot of blood and sweat to make it so, mostly due to  the  bloody  broken  perl  that
       nobody  seems  to  care about. The fork emulation is a bad joke - I have yet to see something useful that
       you can do with it without running into memory corruption issues or other braindamage. Hrrrr.

       Since fork is endlessly broken on win32 perls (it doesn't  even  remotely  work  within  it's  documented
       limits)  and  quite obviously it's not getting improved any time soon, the best way to proceed on windows
       would be to always use "new_exec" and thus never rely on perl's fork "emulation".

       Cygwin perl is not supported at the moment due to some hilarious shortcomings of its API - see IO::FDPoll
       for more details. If you never use "send_fh" and always use "new_exec" to  create  processes,  it  should
       work though.

USING AnyEvent::Fork IN SUBPROCESSES

       AnyEvent::Fork  itself  cannot  generally be used in subprocesses. As long as only one process ever forks
       new processes, sharing the template processes is possible (you could use a pipe as a lock  by  writing  a
       byte into it to unlock, and reading the byte to lock for example)

       To  make  concurrent  calls  possible  after  fork,  you  should  get  rid of the template and early fork
       processes. AnyEvent::Fork will create a new template process as needed.

          undef $AnyEvent::Fork::EARLY;
          undef $AnyEvent::Fork::TEMPLATE;

       It doesn't matter whether you get rid of them in the parent or child after a fork.

SEE ALSO

       AnyEvent::Fork::Early, to avoid executing a perl interpreter at all (part of this distribution).

       AnyEvent::Fork::Template, to create a process by forking the main program at a convenient time  (part  of
       this distribution).

       AnyEvent::Fork::Remote,  for another way to create processes that is mostly compatible to this module and
       modules building on top of it, but works better with remote processes.

       AnyEvent::Fork::RPC, for simple RPC to child processes (on CPAN).

       AnyEvent::Fork::Pool, for simple worker process pool (on CPAN).

AUTHOR AND CONTACT INFORMATION

        Marc Lehmann <schmorp@schmorp.de>
        http://software.schmorp.de/pkg/AnyEvent-Fork

perl v5.32.1                                       2022-01-29                                          Fork(3pm)