Herbert Breunung > Perl6-Doc > Synopsis_17

Download:
Perl6-Doc-0.36.tar.gz

Annotate this POD

CPAN RT

Open  0
View/Report Bugs
Source  

NAME ^

Synopsis_17 - Concurrency [DRAFT]

AUTHOR ^

 Elizabeth Mattijsen <liz@dijkmat.nl>
 Audrey Tang <autrijus@autrijus.org>

VERSION ^

 Maintainer: Elizabeth Mattijsen <liz@dijkmat.nl>
 Date: 13 Jun 2005
 Last Modified: 13 Nov 2005
 Number: 0
 Version: 1

SKETCH ^

This is a rough sketch of how concurrency works in Perl 6.

(actually these are just random notes, put here under the release-early release-often principle, slowly being integrated in a more textual format. Patches welcome!)

OVERVIEW ^

Concurrency can take many forms in Perl 6. With varying degrees of explicitness and control capabilities. This document attempts to describe what these capabilities are and in which form they can be accessed in Perl 6.

Processes, threads, fibers?

Concurrency comes in many shapes and forms. Most Perl users are used to the concept of a "process" or a "thread" (usually depending on the OS they work on). Some systems even are familiar with very lightweight threads called "fibers".

When discussing issues about concurrency with different people, it soon becomes apparent that everybody has his own set of "understandings" about what each word means, which doesn't make it any easier to describe Perl 6 concurrency.

It seemed the most natural to use the word "thread" to describe a process which has its own context, but also shares context with 0 or more concurrently running processes. Depending on your OS, or even specific version of your OS, this could still be a single "process" from the OS's point of view. Or it could contain an OS process for each thread. Or any mixture of these two implementations.

In this document we try to be agnostic about this: all we know in Perl 6 are "threads", which have their own context and share context with other concurrently running "threads". Whether they be process, threads or fibres at the OS level should not matter at the Perl 6 level.

And for sake of consistency, an unthreaded "normal" program is considered to be also running in a single thread.

Variables

In the past, there have been two models for concurrent processes in Perl. In general, these are referred to as "5.005 threads" (perldoc perlothrtut) and "ithreads" (perldoc perlthrtut).

The main difference between these two models from a programmer's point of view, is that variables in "5.005 threads" are shared by default. Whereas in the "ithreads" model, only variables that have been indicated to be "shared", are actually shared between threads. All other variable values are actually copies of the variable's value in the "parent" thread.

With regards to variables, the concurrency model of Perl 6 is closer to the "5.005 threads" model than it is to the "ithreads" model. In fact, all variables "visible" to a particular scope in Perl 6 will be accessible and modifiable from all of the concurrent processes that start from that scope. In that sense, one could consider the "ithreads" model as a historical diversion: the Perl 6 concurrency picks up where the "5.005 threads" path left off.

(EM: maybe point out that the "ithreads" behaviour can be simulated with some kind of copy-on-write magic to be automagically added to all variable access inside a thread, except for those with an explicit "is shared" attribute?)

No user accessible locks

Differently from any current concurrent process implementation in Perl, there are no user accessible locks. Instead, the concept of Software Transactionable Memory is used. This is in concept similar to the use of

 BEGIN TRANSACTION
 ... do your uninterruptible actions
 COMMIT

in the database world. More interestingly, this also includes the concept of rollback:

 BEGIN TRANSACTION
 ... do your stuff, but impossible to complete: ROLLBACK

This causes the state of the process to be reverted to the state at the moment the BEGIN TRANSACTION was executed.

Perl 6 supports this concept through Code blocks which are marked "is atomic". These sections are guaranteed to either be completed totally (when the Code block is exited), or have their state reverted to the state at the start of the Code block (with the retry statement).

(EM: maybe point out if / how old style locks can be "simulated", for those needing a migration path?)

Atomic Code blocks

    my ($x, $y);
    sub c is atomic {
        $x -= 3;
        $y += 3;
        if $x < 10 { retry }
    };

    $e = &c.retry_with( &d ); # 
    $e();

    if $i { is atomic; ...  } else { ...; }

A Code block can be marked as "is atomic". This means that code executed inside that scope is guaranteed not to be interrupted in any way.

The start of a block marked "is atomic" also becomes a "checkpoint" to which execution can return (in exactly the same state) if a problem occurs (a.k.a. a retry is done) inside the scope of the Code block.

retry

The retry function basically restores the state of the thread at the last checkpoint and will wait there until an external event allows it to potentially run that atomic section of code again without having to retry again.

If there are no external events possible that could restart execution, an exception will be raised.

The last checkpoint is either the last atomic / non-atomic boundary, or the most immediate caller constructed with retry_with.

retry_with

The retry_with method on an atomic Code object causes a checkpoint to be made for retry, creating an alternate execution path to be followed when a retry is done.

limitations

Because Perl 6 must be able to revert its state to the state it had at the checkpoint, it is not allowed to perform any non-revertable actions. These would include reading / writing from file handles that do not support seek (such as sockets). Attempting to do so will cause a fatal error to occur.

If you're not interested in revertability, but are interested in uninteruptability, you could use the "is critical" trait.

Critical Code blocks

 sub tricky is critical {
     # code accessing external info, not to be interrupted
 }

 if ($update) {
     is critical;
     # code accessing external info, not to be interrupted
 }

A Code block marked "is critical" can not be interrupted in any way. But since it is able to access non-revertible data structures (such as non-seekable file handles), it cannot do a retry as it would be impossible to restore the state to the beginning of the Code block.

Mixing Atomic and Critical

Both "atomic" as well as "critical" propagate down the call chain. This means that any subroutine that in itself is not "atomic" or "critical" becomes uninterruptible if called inside a code block that is marked as "atomic" or "critical".

Atomic Code blocks called inside the call chain of a "critical" code block do not pose a problem, as they are more restrictive.

Any code that attempts to perform any non-revertible action (e.g. reading from a socket) will cause a fatal error when called inside the call chain of an Atomic Code block.

Co-Routines

The execution of co-routine (or "coro" for short) could be considered as a short "side-step" from the normal path of execution, much like the normal calling of a subroutine.

The main difference with a normal subroutine, is that the co-routine supports a special type of return, called "yield".

(EM: not sure whether the "threads->yield" causes so much mental interference that we should use something else for "yield" in the coro context. And whether we should have a seperate "coro" keyword at all: after all, the "yield" could be in a normal subroutine called from a coro, so it's not like the compiler would be allowed to flag "yield" in a sub as an error).

####################################################################### Below here still the more or less unorganized stuff

CORE::GLOBAL::exit; # kills all the threads

# We intententionally do not list cross-machine parallelism Conc:: classes here. # Consult your local 6PAN mirror with a time machine. use Conc::Processes; # fork() or createProcess based implementation use Conc::Threads; # maybe it just exports &async to override the default one, yay use Conc::Multiplex; # this is default

my $thr = async { ...do something... END { } };

Conc::Thread.this Conc::Proc.this

Conc object # name is still up for grabs! - numify to TIDs (as in pugs) - stringify to something sensible (eg. "<Conc:tid=5>"); - enumerable with Conc.list - Conc.yield (if this is to live but deprecated, maybe call it sleep(0)?) - sleep() always respects other threads, thank you very much - standard methods: - .join # wait for invocant to finish (always item cxt) - .die # throw exception in the invocant thread - .alarm # set up alarms - .alarms # query existing alarms - .suspend # pause a thread; fail if already paused - .resume # revive a thread; fail if already running - .detach # survives parent thread demise (promoted to process) # process-local changes no longer affects parent # tentatively, the control methods still applies to it # including wait (which will always return undef) # also needs to discard any atomicity context - attributes: - .started # time - .finished # time - .waiting # suspened (not diff from block on wakeup signal) # waiting on a handle, a condition, a lock, et cetera # otherwise returns false for running threads # if it's finished then it's undef(?) - .current_continuation # the CC currently running in that thread

- "is throttled" trait

    method throttled::trait_auxillary:<is> ($limit=1, :$key=gensym()) {
        # "is throttled" limits max connection to this Code object
        # the throttling is shared among closures with the same key
        # the limit may differ on closures with the same key.
        # if the counter with the "key" equals or exceeds a closure's limit,
        # the closure can't be entered until it's released
        # (this can be trivially implmented using atomic+retry)
    }

    class Foo {
        method a is throttled(:limit(3) :key<blah>) { ... }
        method b is throttled(:limit(2) :key<blah>) { ... }
    }
    my Foo $f .= new;
    async { $f.a }
    async { $f.b }

- Thread::Status - IO objects and containers gets concurrency love! - $obj.wake_on_readable - $obj.wake_on_writable - $obj.wake_on_either_readable_or_writable_or_passed_time(3); # fixme fixme - $obj.wake_on:{.readable} # busy wait, probably

    my @a is Array::Chan = 1..Inf;
    async { @a.push(1) };
    async { @a.blocking_shift({ ... }) };
    async { @a.unshift({ ... }) };

Communication abstractions - shared, transactional variables by default

# program will wait for _all_ threads # unjoined threads will be joined at the beginning of the END block batch # of the parent thread that spawned them

### INTERFACE BARRIER ### module Blah; {

    is atomic;   # retry/orelse/whatever other rollback stuff
                 # limitation: no external IO (without lethal warnings anyway)
                 # can't do anything irreversible

    is critical; # free to do anything irreversible
                 # means "don't interrupt me"
                 # in system with critical section, no interrupts from
                 # other threads will happen during execution
                 # you can't suspend me

    my $boo is export;
    $boo = 1;

    # We decree that this part forms the static interface
    # it's run once during initial compilation under the
    # Separate Compilation doctrine and the syms sealed off
    # to form part fo bytecode syms headers
    %CALLER::<&blah> = { 1 }; # work - adds to export set
    die "Eureka!" if %CALLER::<$sym>; # never dies

    # BEGIN { $boo = time };

    sub IMPORT {
        # VERY DYNAMIC!

        our $i = time;
        %CALLER::<&blah> = { 1 }; # work - adds to export set
        die "Eureka!" if %CALLER::<$sym>; # probes interactively
    }
}
### INTERFACE BARRIER ###

my $sym; threads.new({ use Blah; BEGIN { require(Blah).import }

    my $boo; BEGIN { eval slurp<Blah.pm>; $boo := $Blah::boo };

    ...
});

Signals

Asynchronous exceptions are just like user-initiated exceptions with die, so you can also catch it with regular CATCH blocks as specified in S04.

To declare your main program catches INT signals, put a CATCH block anywhere in the toplevel to handle exceptions like this:

 CATCH {
     when Error::Signal::INT { ... }
 }

Alarm

An alarm is just a pre-arranged exception to be delivered to your program.

By the time alarm has arrived, the current block may have already finished executing, so you would need to set up CATCH blocks in places where an alarm can rise to handle it properly.

You can request an alarm using the number of seconds, or with a target date. It returns a proxy alarm object that you can do interesting things with.

    multi Alarm *alarm (Num $seconds = $CALLER::_, &do = {die Sig::ALARM}, :$repeat = 1)
    multi Alarm *alarm (Date $date, &do = {die Sig::ALARM}, :$repeat = 1)

Perl 6's alarm has three additional features over traditional alarms:

Multiple and Lexical Alarms

One can set up multiple alarms using repeated alarm calls:

    {
        my $a1 = alarm(2);
        my $a2 = alarm(2);
        sleep 10;
        CATCH {
            is critical; # if you don't want $a2 to be raised inside this
            when Sig::ALARM { ... } 
        }
    }

To stop an alarm, call $alarm.stop. The alarms method for Conc objects (including process and threads) returns a list of alarms currently scheduled for that concurrent context.

When an alarm object is garbage collected, the alarm is stopped automatically. Under void context, the implicit alarm object can only be stopped by querying .alarms on the current process.

We are not sure what alarm(0) would mean. Probably a deprecation warning?

Repeated Alarms

If you request a repeated alarm using the repeated named argument, it will attempt to fire off the alarm that many times. However, the alarm will be supressed when inside a CATCH block that's already handling the exception raised by same alarm.

To repeat 0 times is to not fire off any alarms at all. To repeat +Inf times is to repeat over and over again.

Callbacks in Alarms

You can arrange a callback (like JavaScript's setTimeOut) in alarm, which will then be invoked with the then-current code as caller.

If you set up such a callback to another Conc object, what happens is just like when you called .die on behalf of that object -- namely, the callback closure, along with anything it referenced, is shared to the target Conc context.

Unlike in Perl 5's ithreads where you cannot share anything after the fact, this allows passing shared objects in an ad-hoc fashion across concurrent parts of the program. Under the default (multiplexing) concurrency model, this is basically a no-op.

Continuations

Coroutines

## braindump of coro meeting by Liz and Autri, more to follow

- Coros are _like_ processes

coro dbl { yield $_ * 2; yield $_; return }; my @x = 1..10; my %y = map &dbl, @x; # 2 => 2, 6 => 4, 10 => 6, ...

coro perm (@x) { @x.splice(rand(@x),1).yield while @x; }

my &p1 := &perm.start(1..10); my &p2 := &perm.start(1..20);

p1(); p1(); p2(); p2();

coro foo { yield 42 };

(1..10).pick;

coro foo ($x) { yield $x; yield $x+2; cleanup(); while (2) { while (1) { &?SUB.kill; # seppuku } } } # implicit falloff return + return() means startover without yielding # return() means yielding and restart + no implicit falloff (I LIKE THIS)

&foo.finished; # true on return() and false on midway yield()

foo(4); # and that's all she wrote

coro foo ($x) { yield $x; # this point with $x bound to 10 yield $x+1; return 5; ... # this is never reached, I think we all agree }

# If you don't want your variables to get rebound, use "is copy": coro foo ($x is copy) {...} # which is sugar for coro foo ($x) { { my $x := $OUTER::x; ...; # Further calls of &foo rebound $OUTER::x, not $x. } }

sub foo { return undef if rand; ... }

use overload { '&{}' => sub { ... } }

class Coro is Conc::Multiplex does Code { method postcircumfix:<( )> { # start the thread, block stuff (we are in the caller's context) } }

class Hash is extended { method postcircumfix:<( )> (&self: *@_) { &self = ./start(@_); } method start { # remember self # upon return() or normal falloff, restore self } }

%ENV(123);

&foo_continued := &foo.start(10); &foo.start(20);

foo(10); # returns 10

foo(); # be "insufficient param" error or just return 11? foo(20); # returns 21

# continuation coros multi foo () { ...no rebinding... } multi foo ($x) { ...rebinding... }

&foo.kill;

my $first_ret = zoro( type => <even> ); &zoro.variant(:type<even>).kill; &zoro.variant(type => 'even').kill;

zoro( type => <odd> );

zoro( even => 1 ); zoro( odd => 1 );

multi coro zoro ($type where 'even') {} multi coro zoro ($type where 'odd') {}

multi coro zoro ($even is named) {} multi coro zoro ($odd is named) {}

# iblech's thoughts: # Coroutine parameters should never be rebound. Instead, yield(...)s return # value is an Arglist object containing the new arguments: coro bar ($a, $b) { ...; my $new_set_of_args = yield(...); my $sum_of_old_a_and_new_a = $a + $new_set_of_args<$a>; ...; } bar(42, 23); # $a is 42, $b is 23 bar(17, 19); # $a still 42, $b still 19, # $new_set_of_args is \(a => 17, b => 19)

Junctive Autothreading and Hyper Operations

Live in userland for the time being.

Interprocess Communication

I/O Considerations

File Descriptors

Sockets

syntax highlighting: