pthsem man page on DragonFly

Man page or keyword search:  
man Server   44335 pages
apropos Keyword Search (all sections)
Output format
DragonFly logo
[printable version]

.::pth(3)		    pthsem Portable Threads		     .::pth(3)

NAME
       pthsem - GNU Portable Threads

VERSION
       pthsem 2.0.8 based on GNU Pth

SYNOPSIS
       Global Library Management
	   pth_init, pth_kill, pth_ctrl, pth_version.

       Thread Attribute Handling
	   pth_attr_of, pth_attr_new, pth_attr_init, pth_attr_set,
	   pth_attr_get, pth_attr_destroy.

       Thread Control
	   pth_spawn, pth_once, pth_self, pth_suspend, pth_resume, pth_yield,
	   pth_nap, pth_wait, pth_cancel, pth_abort, pth_raise, pth_join,
	   pth_exit.

       Utilities
	   pth_fdmode, pth_time, pth_timeout, pth_int_time, pth_sfiodisc.

       Cancellation Management
	   pth_cancel_point, pth_cancel_state.

       Event Handling
	   pth_event, pth_event_typeof, pth_event_extract, pth_event_concat,
	   pth_event_isolate, pth_event_walk, pth_event_status,
	   pth_event_free.

       Key-Based Storage
	   pth_key_create, pth_key_delete, pth_key_setdata, pth_key_getdata.

       Message Port Communication
	   pth_msgport_create, pth_msgport_destroy, pth_msgport_find,
	   pth_msgport_pending, pth_msgport_put, pth_msgport_get,
	   pth_msgport_reply.

       Thread Cleanups
	   pth_cleanup_push, pth_cleanup_pop.

       Process Forking
	   pth_atfork_push, pth_atfork_pop, pth_fork.

       Synchronization
	   pth_mutex_init, pth_mutex_acquire, pth_mutex_release,
	   pth_rwlock_init, pth_rwlock_acquire, pth_rwlock_release,
	   pth_cond_init, pth_cond_await, pth_cond_notify, pth_barrier_init,
	   pth_barrier_reach.

       Semaphore support
	   pth_sem_init, pth_sem_dec, pth_sem_dec_value, pth_sem_inc,
	   pth_sem_inc_value, pth_sem_set_value, pth_sem_get_value.

       User-Space Context
	   pth_uctx_create, pth_uctx_make, pth_uctx_switch, pth_uctx_destroy.

       Generalized POSIX Replacement API
	   pth_sigwait_ev, pth_accept_ev, pth_connect_ev, pth_select_ev,
	   pth_poll_ev, pth_read_ev, pth_readv_ev, pth_write_ev,
	   pth_writev_ev, pth_recv_ev, pth_recvfrom_ev, pth_send_ev,
	   pth_sendto_ev.

       Standard POSIX Replacement API
	   pth_nanosleep, pth_usleep, pth_sleep, pth_waitpid, pth_system,
	   pth_sigmask, pth_sigwait, pth_accept, pth_connect, pth_select,
	   pth_pselect, pth_poll, pth_read, pth_readv, pth_write, pth_writev,
	   pth_pread, pth_pwrite, pth_recv, pth_recvfrom, pth_send,
	   pth_sendto.

DESCRIPTION
	 ____  _   _
	|  _ \| |_| |__
	| |_) | __| '_ \	 ``Only those who attempt
	|  __/| |_| | | |	   the absurd can achieve
	|_|    \__|_| |_|	   the impossible.''

       Pth is a very portable POSIX/ANSI-C based library for Unix platforms
       which provides non-preemptive priority-based scheduling for multiple
       threads of execution (aka `multithreading') inside event-driven
       applications. All threads run in the same address space of the
       application process, but each thread has its own individual program
       counter, run-time stack, signal mask and "errno" variable.

       The thread scheduling itself is done in a cooperative way, i.e., the
       threads are managed and dispatched by a priority- and event-driven non-
       preemptive scheduler. The intention is that this way both better
       portability and run-time performance is achieved than with preemptive
       scheduling. The event facility allows threads to wait until various
       types of internal and external events occur, including pending I/O on
       file descriptors, asynchronous signals, elapsed timers, pending I/O on
       message ports, thread and process termination, and even results of
       customized callback functions.

       Pth also provides an optional emulation API for POSIX.1c threads
       (`Pthreads') which can be used for backward compatibility to existing
       multithreaded applications. See Pth's pthread(3) manual page for
       details.

   Threading Background
       When programming event-driven applications, usually servers, lots of
       regular jobs and one-shot requests have to be processed in parallel.
       To efficiently simulate this parallel processing on uniprocessor
       machines, we use `multitasking' -- that is, we have the application ask
       the operating system to spawn multiple instances of itself. On Unix,
       typically the kernel implements multitasking in a preemptive and
       priority-based way through heavy-weight processes spawned with fork(2).
       These processes usually do not share a common address space. Instead
       they are clearly separated from each other, and are created by direct
       cloning a process address space (although modern kernels use memory
       segment mapping and copy-on-write semantics to avoid unnecessary
       copying of physical memory).

       The drawbacks are obvious: Sharing data between the processes is
       complicated, and can usually only be done efficiently through shared
       memory (but which itself is not very portable). Synchronization is
       complicated because of the preemptive nature of the Unix scheduler (one
       has to use atomic locks, etc). The machine's resources can be exhausted
       very quickly when the server application has to serve too many long-
       running requests (heavy-weight processes cost memory). And when each
       request spawns a sub-process to handle it, the server performance and
       responsiveness is horrible (heavy-weight processes cost time to spawn).
       Finally, the server application doesn't scale very well with the load
       because of these resource problems. In practice, lots of tricks are
       usually used to overcome these problems - ranging from pre-forked sub-
       process pools to semi-serialized processing, etc.

       One of the most elegant ways to solve these resource- and data-sharing
       problems is to have multiple light-weight threads of execution inside a
       single (heavy-weight) process, i.e., to use multithreading.  Those
       threads usually improve responsiveness and performance of the
       application, often improve and simplify the internal program structure,
       and most important, require less system resources than heavy-weight
       processes. Threads are neither the optimal run-time facility for all
       types of applications, nor can all applications benefit from them. But
       at least event-driven server applications usually benefit greatly from
       using threads.

   The World of Threading
       Even though lots of documents exists which describe and define the
       world of threading, to understand Pth, you need only basic knowledge
       about threading. The following definitions of thread-related terms
       should at least help you understand thread programming enough to allow
       you to use Pth.

       o process vs. thread
	 A process on Unix systems consists of at least the following
	 fundamental ingredients: virtual memory table, program code, program
	 counter, heap memory, stack memory, stack pointer, file descriptor
	 set, signal table. On every process switch, the kernel saves and
	 restores these ingredients for the individual processes. On the other
	 hand, a thread consists of only a private program counter, stack
	 memory, stack pointer and signal table. All other ingredients, in
	 particular the virtual memory, it shares with the other threads of
	 the same process.

       o kernel-space vs. user-space threading
	 Threads on a Unix platform traditionally can be implemented either
	 inside kernel-space or user-space. When threads are implemented by
	 the kernel, the thread context switches are performed by the kernel
	 without the application's knowledge. Similarly, when threads are
	 implemented in user-space, the thread context switches are performed
	 by an application library, without the kernel's knowledge. There also
	 are hybrid threading approaches where, typically, a user-space
	 library binds one or more user-space threads to one or more kernel-
	 space threads (there usually called light-weight processes - or in
	 short LWPs).

	 User-space threads are usually more portable and can perform faster
	 and cheaper context switches (for instance via swapcontext(2) or
	 setjmp(3)/longjmp(3)) than kernel based threads. On the other hand,
	 kernel-space threads can take advantage of multiprocessor machines
	 and don't have any inherent I/O blocking problems. Kernel-space
	 threads are usually scheduled in preemptive way side-by-side with the
	 underlying processes. User-space threads on the other hand use either
	 preemptive or non-preemptive scheduling.

       o preemptive vs. non-preemptive thread scheduling
	 In preemptive scheduling, the scheduler lets a thread execute until a
	 blocking situation occurs (usually a function call which would block)
	 or the assigned timeslice elapses. Then it detracts control from the
	 thread without a chance for the thread to object. This is usually
	 realized by interrupting the thread through a hardware interrupt
	 signal (for kernel-space threads) or a software interrupt signal (for
	 user-space threads), like "SIGALRM" or "SIGVTALRM". In non-preemptive
	 scheduling, once a thread received control from the scheduler it
	 keeps it until either a blocking situation occurs (again a function
	 call which would block and instead switches back to the scheduler) or
	 the thread explicitly yields control back to the scheduler in a
	 cooperative way.

       o concurrency vs. parallelism
	 Concurrency exists when at least two threads are in progress at the
	 same time. Parallelism arises when at least two threads are executing
	 simultaneously. Real parallelism can be only achieved on
	 multiprocessor machines, of course. But one also usually speaks of
	 parallelism or high concurrency in the context of preemptive thread
	 scheduling and of low concurrency in the context of non-preemptive
	 thread scheduling.

       o responsiveness
	 The responsiveness of a system can be described by the user visible
	 delay until the system responses to an external request. When this
	 delay is small enough and the user doesn't recognize a noticeable
	 delay, the responsiveness of the system is considered good. When the
	 user recognizes or is even annoyed by the delay, the responsiveness
	 of the system is considered bad.

       o reentrant, thread-safe and asynchronous-safe functions
	 A reentrant function is one that behaves correctly if it is called
	 simultaneously by several threads and then also executes
	 simultaneously.  Functions that access global state, such as memory
	 or files, of course, need to be carefully designed in order to be
	 reentrant. Two traditional approaches to solve these problems are
	 caller-supplied states and thread-specific data.

	 Thread-safety is the avoidance of data races, i.e., situations in
	 which data is set to either correct or incorrect value depending upon
	 the (unpredictable) order in which multiple threads access and modify
	 the data. So a function is thread-safe when it still behaves
	 semantically correct when called simultaneously by several threads
	 (it is not required that the functions also execute simultaneously).
	 The traditional approach to achieve thread-safety is to wrap a
	 function body with an internal mutual exclusion lock (aka `mutex').
	 As you should recognize, reentrant is a stronger attribute than
	 thread-safe, because it is harder to achieve and results especially
	 in no run-time contention between threads. So, a reentrant function
	 is always thread-safe, but not vice versa.

	 Additionally there is a related attribute for functions named
	 asynchronous-safe, which comes into play in conjunction with signal
	 handlers. This is very related to the problem of reentrant functions.
	 An asynchronous-safe function is one that can be called safe and
	 without side-effects from within a signal handler context. Usually
	 very few functions are of this type, because an application is very
	 restricted in what it can perform from within a signal handler
	 (especially what system functions it is allowed to call). The reason
	 mainly is, because only a few system functions are officially
	 declared by POSIX as guaranteed to be asynchronous-safe.
	 Asynchronous-safe functions usually have to be already reentrant.

   User-Space Threads
       User-space threads can be implemented in various way. The two
       traditional approaches are:

       1. Matrix-based explicit dispatching between small units of execution:

	  Here the global procedures of the application are split into small
	  execution units (each is required to not run for more than a few
	  milliseconds) and those units are implemented by separate functions.
	  Then a global matrix is defined which describes the execution (and
	  perhaps even dependency) order of these functions. The main server
	  procedure then just dispatches between these units by calling one
	  function after each other controlled by this matrix. The threads are
	  created by more than one jump-trail through this matrix and by
	  switching between these jump-trails controlled by corresponding
	  occurred events.

	  This approach gives the best possible performance, because one can
	  fine-tune the threads of execution by adjusting the matrix, and the
	  scheduling is done explicitly by the application itself. It is also
	  very portable, because the matrix is just an ordinary data
	  structure, and functions are a standard feature of ANSI C.

	  The disadvantage of this approach is that it is complicated to write
	  large applications with this approach, because in those applications
	  one quickly gets hundreds(!) of execution units and the control flow
	  inside such an application is very hard to understand (because it is
	  interrupted by function borders and one always has to remember the
	  global dispatching matrix to follow it). Additionally, all threads
	  operate on the same execution stack. Although this saves memory, it
	  is often nasty, because one cannot switch between threads in the
	  middle of a function. Thus the scheduling borders are the function
	  borders.

       2. Context-based implicit scheduling between threads of execution:

	  Here the idea is that one programs the application as with forked
	  processes, i.e., one spawns a thread of execution and this runs from
	  the begin to the end without an interrupted control flow. But the
	  control flow can be still interrupted - even in the middle of a
	  function.  Actually in a preemptive way, similar to what the kernel
	  does for the heavy-weight processes, i.e., every few milliseconds
	  the user-space scheduler switches between the threads of execution.
	  But the thread itself doesn't recognize this and usually (except for
	  synchronization issues) doesn't have to care about this.

	  The advantage of this approach is that it's very easy to program,
	  because the control flow and context of a thread directly follows a
	  procedure without forced interrupts through function borders.
	  Additionally, the programming is very similar to a traditional and
	  well understood fork(2) based approach.

	  The disadvantage is that although the general performance is
	  increased, compared to using approaches based on heavy-weight
	  processes, it is decreased compared to the matrix-approach above.
	  Because the implicit preemptive scheduling does usually a lot more
	  context switches (every user-space context switch costs some
	  overhead even when it is a lot cheaper than a kernel-level context
	  switch) than the explicit cooperative/non-preemptive scheduling.
	  Finally, there is no really portable POSIX/ANSI-C based way to
	  implement user-space preemptive threading. Either the platform
	  already has threads, or one has to hope that some semi-portable
	  package exists for it. And even those semi-portable packages usually
	  have to deal with assembler code and other nasty internals and are
	  not easy to port to forthcoming platforms.

       So, in short: the matrix-dispatching approach is portable and fast, but
       nasty to program. The thread scheduling approach is easy to program,
       but suffers from synchronization and portability problems caused by its
       preemptive nature.

   The Compromise of Pth
       But why not combine the good aspects of both approaches while avoiding
       their bad aspects? That's the goal of Pth. Pth implements easy-to-
       program threads of execution, but avoids the problems of preemptive
       scheduling by using non-preemptive scheduling instead.

       This sounds like, and is, a useful approach. Nevertheless, one has to
       keep the implications of non-preemptive thread scheduling in mind when
       working with Pth. The following list summarizes a few essential points:

       o Pth provides maximum portability, but NOT the fanciest features.

	 This is, because it uses a nifty and portable POSIX/ANSI-C approach
	 for thread creation (and this way doesn't require any platform
	 dependent assembler hacks) and schedules the threads in non-
	 preemptive way (which doesn't require unportable facilities like
	 "SIGVTALRM"). On the other hand, this way not all fancy threading
	 features can be implemented.  Nevertheless the available facilities
	 are enough to provide a robust and full-featured threading system.

       o Pth increases the responsiveness and concurrency of an event-driven
	 application, but NOT the concurrency of number-crunching
	 applications.

	 The reason is the non-preemptive scheduling. Number-crunching
	 applications usually require preemptive scheduling to achieve
	 concurrency because of their long CPU bursts. For them, non-
	 preemptive scheduling (even together with explicit yielding) provides
	 only the old concept of `coroutines'. On the other hand, event driven
	 applications benefit greatly from non-preemptive scheduling. They
	 have only short CPU bursts and lots of events to wait on, and this
	 way run faster under non-preemptive scheduling because no unnecessary
	 context switching occurs, as it is the case for preemptive
	 scheduling. That's why Pth is mainly intended for server type
	 applications, although there is no technical restriction.

       o Pth requires thread-safe functions, but NOT reentrant functions.

	 This nice fact exists again because of the nature of non-preemptive
	 scheduling, where a function isn't interrupted and this way cannot be
	 reentered before it returned. This is a great portability benefit,
	 because thread-safety can be achieved more easily than reentrance
	 possibility. Especially this means that under Pth more existing
	 third-party libraries can be used without side-effects than it's the
	 case for other threading systems.

       o Pth doesn't require any kernel support, but can NOT benefit from
	 multiprocessor machines.

	 This means that Pth runs on almost all Unix kernels, because the
	 kernel does not need to be aware of the Pth threads (because they are
	 implemented entirely in user-space). On the other hand, it cannot
	 benefit from the existence of multiprocessors, because for this,
	 kernel support would be needed. In practice, this is no problem,
	 because multiprocessor systems are rare, and portability is almost
	 more important than highest concurrency.

   The life cycle of a thread
       To understand the Pth Application Programming Interface (API), it helps
       to first understand the life cycle of a thread in the Pth threading
       system. It can be illustrated with the following directed graph:

		    NEW
		     |
		     V
	     +---> READY ---+
	     |	     ^	    |
	     |	     |	    V
	  WAITING <--+-- RUNNING
			    |
	     :		    V
	  SUSPENDED	  DEAD

       When a new thread is created, it is moved into the NEW queue of the
       scheduler. On the next dispatching for this thread, the scheduler picks
       it up from there and moves it to the READY queue. This is a queue
       containing all threads which want to perform a CPU burst. There they
       are queued in priority order. On each dispatching step, the scheduler
       always removes the thread with the highest priority only. It then
       increases the priority of all remaining threads by 1, to prevent them
       from `starving'.

       The thread which was removed from the READY queue is the new RUNNING
       thread (there is always just one RUNNING thread, of course). The
       RUNNING thread is assigned execution control. After this thread yields
       execution (either explicitly by yielding execution or implicitly by
       calling a function which would block) there are three possibilities:
       Either it has terminated, then it is moved to the DEAD queue, or it has
       events on which it wants to wait, then it is moved into the WAITING
       queue. Else it is assumed it wants to perform more CPU bursts and
       immediately enters the READY queue again.

       Before the next thread is taken out of the READY queue, the WAITING
       queue is checked for pending events. If one or more events occurred,
       the threads that are waiting on them are immediately moved to the READY
       queue.

       The purpose of the NEW queue has to do with the fact that in Pth a
       thread never directly switches to another thread. A thread always
       yields execution to the scheduler and the scheduler dispatches to the
       next thread. So a freshly spawned thread has to be kept somewhere until
       the scheduler gets a chance to pick it up for scheduling. That is what
       the NEW queue is for.

       The purpose of the DEAD queue is to support thread joining. When a
       thread is marked to be unjoinable, it is directly kicked out of the
       system after it terminated. But when it is joinable, it enters the DEAD
       queue. There it remains until another thread joins it.

       Finally, there is a special separated queue named SUSPENDED, to where
       threads can be manually moved from the NEW, READY or WAITING queues by
       the application. The purpose of this special queue is to temporarily
       absorb suspended threads until they are again resumed by the
       application. Suspended threads do not cost scheduling or event handling
       resources, because they are temporarily completely out of the
       scheduler's scope. If a thread is resumed, it is moved back to the
       queue from where it originally came and this way again enters the
       schedulers scope.

APPLICATION PROGRAMMING INTERFACE (API)
       In the following the Pth Application Programming Interface (API) is
       discussed in detail. With the knowledge given above, it should now be
       easy to understand how to program threads with this API. In good Unix
       tradition, Pth functions use special return values ("NULL" in pointer
       context, "FALSE" in boolean context and "-1" in integer context) to
       indicate an error condition and set (or pass through) the "errno"
       system variable to pass more details about the error to the caller.

   Global Library Management
       The following functions act on the library as a whole.  They are used
       to initialize and shutdown the scheduler and fetch information from it.

       int pth_init(void);
	   This initializes the Pth library. It has to be the first Pth API
	   function call in an application, and is mandatory. It's usually
	   done at the begin of the main() function of the application. This
	   implicitly spawns the internal scheduler thread and transforms the
	   single execution unit of the current process into a thread (the
	   `main' thread). It returns "TRUE" on success and "FALSE" on error.

       int pth_kill(void);
	   This kills the Pth library. It should be the last Pth API function
	   call in an application, but is not really required. It's usually
	   done at the end of the main function of the application. At least,
	   it has to be called from within the main thread. It implicitly
	   kills all threads and transforms back the calling thread into the
	   single execution unit of the underlying process.  The usual way to
	   terminate a Pth application is either a simple `"pth_exit(0);"' in
	   the main thread (which waits for all other threads to terminate,
	   kills the threading system and then terminates the process) or a
	   `"pth_kill(); exit(0)"' (which immediately kills the threading
	   system and terminates the process). The pth_kill() return
	   immediately with a return code of "FALSE" if it is not called from
	   within the main thread. Else it kills the threading system and
	   returns "TRUE".

       long pth_ctrl(unsigned long query, ...);
	   This is a generalized query/control function for the Pth library.
	   The argument query is a bitmask formed out of one or more
	   "PTH_CTRL_"XXXX queries. Currently the following queries are
	   supported:

	   "PTH_CTRL_GETTHREADS"
	       This returns the total number of threads currently in
	       existence.  This query actually is formed out of the
	       combination of queries for threads in a particular state, i.e.,
	       the "PTH_CTRL_GETTHREADS" query is equal to the OR-combination
	       of all the following specialized queries:

	       "PTH_CTRL_GETTHREADS_NEW" for the number of threads in the new
	       queue (threads created via pth_spawn(3) but still not scheduled
	       once), "PTH_CTRL_GETTHREADS_READY" for the number of threads in
	       the ready queue (threads who want to do CPU bursts),
	       "PTH_CTRL_GETTHREADS_RUNNING" for the number of running threads
	       (always just one thread!), "PTH_CTRL_GETTHREADS_WAITING" for
	       the number of threads in the waiting queue (threads waiting for
	       events), "PTH_CTRL_GETTHREADS_SUSPENDED" for the number of
	       threads in the suspended queue (threads waiting to be resumed)
	       and "PTH_CTRL_GETTHREADS_DEAD" for the number of threads in the
	       new queue (terminated threads waiting for a join).

	   "PTH_CTRL_GETAVLOAD"
	       This requires a second argument of type `"float *"' (pointer to
	       a floating point variable).  It stores a floating point value
	       describing the exponential averaged load of the scheduler in
	       this variable. The load is a function from the number of
	       threads in the ready queue of the schedulers dispatching unit.
	       So a load around 1.0 means there is only one ready thread (the
	       standard situation when the application has no high load). A
	       higher load value means there a more threads ready who want to
	       do CPU bursts. The average load value updates once per second
	       only. The return value for this query is always 0.

	   "PTH_CTRL_GETPRIO"
	       This requires a second argument of type `"pth_t"' which
	       identifies a thread.  It returns the priority (ranging from
	       "PTH_PRIO_MIN" to "PTH_PRIO_MAX") of the given thread.

	   "PTH_CTRL_GETNAME"
	       This requires a second argument of type `"pth_t"' which
	       identifies a thread. It returns the name of the given thread,
	       i.e., the return value of pth_ctrl(3) should be casted to a
	       `"char *"'.

	   "PTH_CTRL_DUMPSTATE"
	       This requires a second argument of type `"FILE *"' to which a
	       summary of the internal Pth library state is written to. The
	       main information which is currently written out is the current
	       state of the thread pool.

	   "PTH_CTRL_FAVOURNEW"
	       This requires a second argument of type `"int"' which specified
	       whether the GNU Pth scheduler favours new threads on startup,
	       i.e., whether they are moved from the new queue to the top
	       (argument is "TRUE") or middle (argument is "FALSE") of the
	       ready queue. The default is to favour new threads to make sure
	       they do not starve already at startup, although this slightly
	       violates the strict priority based scheduling.

	   The function returns "-1" on error.

       long pth_version(void);
	   This function returns a hex-value `0xVRRTLL' which describes the
	   current Pth library version. V is the version, RR the revisions, LL
	   the level and T the type of the level (alphalevel=0, betalevel=1,
	   patchlevel=2, etc). For instance Pth version 1.0b1 is encoded as
	   0x100101.  The reason for this unusual mapping is that this way the
	   version number is steadily increasing. The same value is also
	   available under compile time as "PTH_VERSION".

   Thread Attribute Handling
       Attribute objects are used in Pth for two things: First
       stand-alone/unbound attribute objects are used to store attributes for
       to be spawned threads.  Bounded attribute objects are used to modify
       attributes of already existing threads. The following attribute fields
       exists in attribute objects:

       "PTH_ATTR_PRIO" (read-write) ["int"]
	   Thread Priority between "PTH_PRIO_MIN" and "PTH_PRIO_MAX".  The
	   default is "PTH_PRIO_STD".

       "PTH_ATTR_NAME" (read-write) ["char *"]
	   Name of thread (up to 40 characters are stored only), mainly for
	   debugging purposes.

       "PTH_ATTR_DISPATCHES" (read-write) ["int"]
	   In bounded attribute objects, this field is incremented every time
	   the context is switched to the associated thread.

       "PTH_ATTR_JOINABLE" (read-write> ["int"]
	   The thread detachment type, "TRUE" indicates a joinable thread,
	   "FALSE" indicates a detached thread. When a thread is detached,
	   after termination it is immediately kicked out of the system
	   instead of inserted into the dead queue.

       "PTH_ATTR_CANCEL_STATE" (read-write) ["unsigned int"]
	   The thread cancellation state, i.e., a combination of
	   "PTH_CANCEL_ENABLE" or "PTH_CANCEL_DISABLE" and
	   "PTH_CANCEL_DEFERRED" or "PTH_CANCEL_ASYNCHRONOUS".

       "PTH_ATTR_STACK_SIZE" (read-write) ["unsigned int"]
	   The thread stack size in bytes. Use lower values than 64 KB with
	   great care!

       "PTH_ATTR_STACK_ADDR" (read-write) ["char *"]
	   A pointer to the lower address of a chunk of malloc(3)'ed memory
	   for the stack.

       "PTH_ATTR_TIME_SPAWN" (read-only) ["pth_time_t"]
	   The time when the thread was spawned.  This can be queried only
	   when the attribute object is bound to a thread.

       "PTH_ATTR_TIME_LAST" (read-only) ["pth_time_t"]
	   The time when the thread was last dispatched.  This can be queried
	   only when the attribute object is bound to a thread.

       "PTH_ATTR_TIME_RAN" (read-only) ["pth_time_t"]
	   The total time the thread was running.  This can be queried only
	   when the attribute object is bound to a thread.

       "PTH_ATTR_START_FUNC" (read-only) ["void *(*)(void *)"]
	   The thread start function.  This can be queried only when the
	   attribute object is bound to a thread.

       "PTH_ATTR_START_ARG" (read-only) ["void *"]
	   The thread start argument.  This can be queried only when the
	   attribute object is bound to a thread.

       "PTH_ATTR_STATE" (read-only) ["pth_state_t"]
	   The scheduling state of the thread, i.e., either "PTH_STATE_NEW",
	   "PTH_STATE_READY", "PTH_STATE_WAITING", or "PTH_STATE_DEAD" This
	   can be queried only when the attribute object is bound to a thread.

       "PTH_ATTR_EVENTS" (read-only) ["pth_event_t"]
	   The event ring the thread is waiting for.  This can be queried only
	   when the attribute object is bound to a thread.

       "PTH_ATTR_BOUND" (read-only) ["int"]
	   Whether the attribute object is bound ("TRUE") to a thread or not
	   ("FALSE").

       The following API functions can be used to handle the attribute
       objects:

       pth_attr_t pth_attr_of(pth_t tid);
	   This returns a new attribute object bound to thread tid.  Any
	   queries on this object directly fetch attributes from tid. And
	   attribute modifications directly change tid. Use such attribute
	   objects to modify existing threads.

       pth_attr_t pth_attr_new(void);
	   This returns a new unbound attribute object. An implicit
	   pth_attr_init() is done on it. Any queries on this object just
	   fetch stored attributes from it.  And attribute modifications just
	   change the stored attributes.  Use such attribute objects to pre-
	   configure attributes for to be spawned threads.

       int pth_attr_init(pth_attr_t attr);
	   This initializes an attribute object attr to the default values:
	   "PTH_ATTR_PRIO" := "PTH_PRIO_STD", "PTH_ATTR_NAME" := `"unknown"',
	   "PTH_ATTR_DISPATCHES" := 0, "PTH_ATTR_JOINABLE" := "TRUE",
	   "PTH_ATTR_CANCELSTATE" := "PTH_CANCEL_DEFAULT",
	   "PTH_ATTR_STACK_SIZE" := 64*1024 and "PTH_ATTR_STACK_ADDR" :=
	   "NULL". All other "PTH_ATTR_*" attributes are read-only attributes
	   and don't receive default values in attr, because they exists only
	   for bounded attribute objects.

       int pth_attr_set(pth_attr_t attr, int field, ...);
	   This sets the attribute field field in attr to a value specified as
	   an additional argument on the variable argument list. The following
	   attribute fields and argument pairs can be used:

	    PTH_ATTR_PRIO	    int
	    PTH_ATTR_NAME	    char *
	    PTH_ATTR_DISPATCHES	    int
	    PTH_ATTR_JOINABLE	    int
	    PTH_ATTR_CANCEL_STATE   unsigned int
	    PTH_ATTR_STACK_SIZE	    unsigned int
	    PTH_ATTR_STACK_ADDR	    char *

       int pth_attr_get(pth_attr_t attr, int field, ...);
	   This retrieves the attribute field field in attr and stores its
	   value in the variable specified through a pointer in an additional
	   argument on the variable argument list. The following fields and
	   argument pairs can be used:

	    PTH_ATTR_PRIO	    int *
	    PTH_ATTR_NAME	    char **
	    PTH_ATTR_DISPATCHES	    int *
	    PTH_ATTR_JOINABLE	    int *
	    PTH_ATTR_CANCEL_STATE   unsigned int *
	    PTH_ATTR_STACK_SIZE	    unsigned int *
	    PTH_ATTR_STACK_ADDR	    char **
	    PTH_ATTR_TIME_SPAWN	    pth_time_t *
	    PTH_ATTR_TIME_LAST	    pth_time_t *
	    PTH_ATTR_TIME_RAN	    pth_time_t *
	    PTH_ATTR_START_FUNC	    void *(**)(void *)
	    PTH_ATTR_START_ARG	    void **
	    PTH_ATTR_STATE	    pth_state_t *
	    PTH_ATTR_EVENTS	    pth_event_t *
	    PTH_ATTR_BOUND	    int *

       int pth_attr_destroy(pth_attr_t attr);
	   This destroys a attribute object attr. After this attr is no longer
	   a valid attribute object.

   Thread Control
       The following functions control the threading itself and make up the
       main API of the Pth library.

       pth_t pth_spawn(pth_attr_t attr, void *(*entry)(void *), void *arg);
	   This spawns a new thread with the attributes given in attr (or
	   "PTH_ATTR_DEFAULT" for default attributes - which means that thread
	   priority, joinability and cancel state are inherited from the
	   current thread) with the starting point at routine entry; the
	   dispatch count is not inherited from the current thread if attr is
	   not specified - rather, it is initialized to zero.  This entry
	   routine is called as `pth_exit(entry(arg))' inside the new thread
	   unit, i.e., entry's return value is fed to an implicit pth_exit(3).
	   So the thread can also exit by just returning. Nevertheless the
	   thread can also exit explicitly at any time by calling pth_exit(3).
	   But keep in mind that calling the POSIX function exit(3) still
	   terminates the complete process and not just the current thread.

	   There is no Pth-internal limit on the number of threads one can
	   spawn, except the limit implied by the available virtual memory.
	   Pth internally keeps track of thread in dynamic data structures.
	   The function returns "NULL" on error.

       int pth_once(pth_once_t *ctrlvar, void (*func)(void *), void *arg);
	   This is a convenience function which uses a control variable of
	   type "pth_once_t" to make sure a constructor function func is
	   called only once as `func(arg)' in the system. In other words: Only
	   the first call to pth_once(3) by any thread in the system succeeds.
	   The variable referenced via ctrlvar should be declared as
	   `"pth_once_t" variable-name = "PTH_ONCE_INIT";' before calling this
	   function.

       pth_t pth_self(void);
	   This just returns the unique thread handle of the currently running
	   thread.  This handle itself has to be treated as an opaque entity
	   by the application.	It's usually used as an argument to other
	   functions who require an argument of type "pth_t".

       int pth_suspend(pth_t tid);
	   This suspends a thread tid until it is manually resumed again via
	   pth_resume(3). For this, the thread is moved to the SUSPENDED queue
	   and this way is completely out of the scheduler's event handling
	   and thread dispatching scope. Suspending the current thread is not
	   allowed.  The function returns "TRUE" on success and "FALSE" on
	   errors.

       int pth_resume(pth_t tid);
	   This function resumes a previously suspended thread tid, i.e. tid
	   has to stay on the SUSPENDED queue. The thread is moved to the NEW,
	   READY or WAITING queue (dependent on what its state was when the
	   pth_suspend(3) call were made) and this way again enters the event
	   handling and thread dispatching scope of the scheduler. The
	   function returns "TRUE" on success and "FALSE" on errors.

       int pth_raise(pth_t tid, int sig)
	   This function raises a signal for delivery to thread tid only.
	   When one just raises a signal via raise(3) or kill(2), its
	   delivered to an arbitrary thread which has this signal not blocked.
	   With pth_raise(3) one can send a signal to a thread and its
	   guarantees that only this thread gets the signal delivered. But
	   keep in mind that nevertheless the signals action is still
	   configured process-wide.  When sig is 0 plain thread checking is
	   performed, i.e., `"pth_raise(tid, 0)"' returns "TRUE" when thread
	   tid still exists in the PTH system but doesn't send any signal to
	   it.

       int pth_yield(pth_t tid);
	   This explicitly yields back the execution control to the scheduler
	   thread.  Usually the execution is implicitly transferred back to
	   the scheduler when a thread waits for an event. But when a thread
	   has to do larger CPU bursts, it can be reasonable to interrupt it
	   explicitly by doing a few pth_yield(3) calls to give other threads
	   a chance to execute, too.  This obviously is the cooperating part
	   of Pth.  A thread has not to yield execution, of course. But when
	   you want to program a server application with good response times
	   the threads should be cooperative, i.e., when they should split
	   their CPU bursts into smaller units with this call.

	   Usually one specifies tid as "NULL" to indicate to the scheduler
	   that it can freely decide which thread to dispatch next.  But if
	   one wants to indicate to the scheduler that a particular thread
	   should be favored on the next dispatching step, one can specify
	   this thread explicitly. This allows the usage of the old concept of
	   coroutines where a thread/routine switches to a particular
	   cooperating thread. If tid is not "NULL" and points to a new or
	   ready thread, it is guaranteed that this thread receives execution
	   control on the next dispatching step. If tid is in a different
	   state (that is, not in "PTH_STATE_NEW" or "PTH_STATE_READY") an
	   error is reported.

	   The function usually returns "TRUE" for success and only "FALSE"
	   (with "errno" set to "EINVAL") if tid specified an invalid or still
	   not new or ready thread.

       int pth_nap(pth_time_t naptime);
	   This functions suspends the execution of the current thread until
	   naptime is elapsed. naptime is of type "pth_time_t" and this way
	   has theoretically a resolution of one microsecond. In practice you
	   should neither rely on this nor that the thread is awakened exactly
	   after naptime has elapsed. It's only guarantees that the thread
	   will sleep at least naptime. But because of the non-preemptive
	   nature of Pth it can last longer (when another thread kept the CPU
	   for a long time). Additionally the resolution is dependent of the
	   implementation of timers by the operating system and these usually
	   have only a resolution of 10 microseconds or larger. But usually
	   this isn't important for an application unless it tries to use this
	   facility for real time tasks.

       int pth_wait(pth_event_t ev);
	   This is the link between the scheduler and the event facility (see
	   below for the various pth_event_xxx() functions). It's modeled like
	   select(2), i.e., one gives this function one or more events (in the
	   event ring specified by ev) on which the current thread wants to
	   wait. The scheduler awakes the thread when one ore more of them
	   occurred or failed after tagging them as such. The ev argument is a
	   pointer to an event ring which isn't changed except for the
	   tagging. pth_wait(3) returns the number of occurred or failed
	   events and the application can use pth_event_status(3) to test
	   which events occurred or failed.

       int pth_cancel(pth_t tid);
	   This cancels a thread tid. How the cancellation is done depends on
	   the cancellation state of tid which the thread can configure
	   itself. When its state is "PTH_CANCEL_DISABLE" a cancellation
	   request is just made pending.  When it is "PTH_CANCEL_ENABLE" it
	   depends on the cancellation type what is performed. When its
	   "PTH_CANCEL_DEFERRED" again the cancellation request is just made
	   pending. But when its "PTH_CANCEL_ASYNCHRONOUS" the thread is
	   immediately canceled before pth_cancel(3) returns. The effect of a
	   thread cancellation is equal to implicitly forcing the thread to
	   call `"pth_exit(PTH_CANCELED)"' at one of his cancellation points.
	   In Pth thread enter a cancellation point either explicitly via
	   pth_cancel_point(3) or implicitly by waiting for an event.

       int pth_abort(pth_t tid);
	   This is the cruel way to cancel a thread tid. When it's already
	   dead and waits to be joined it just joins it (via `"pth_join("tid",
	   NULL)"') and this way kicks it out of the system.  Else it forces
	   the thread to be not joinable and to allow asynchronous
	   cancellation and then cancels it via `"pth_cancel("tid")"'.

       int pth_join(pth_t tid, void **value);
	   This joins the current thread with the thread specified via tid.
	   It first suspends the current thread until the tid thread has
	   terminated. Then it is awakened and stores the value of tid's
	   pth_exit(3) call into *value (if value and not "NULL") and returns
	   to the caller. A thread can be joined only when it has the
	   attribute "PTH_ATTR_JOINABLE" set to "TRUE" (the default). A thread
	   can only be joined once, i.e., after the pth_join(3) call the
	   thread tid is completely removed from the system.

       void pth_exit(void *value);
	   This terminates the current thread. Whether it's immediately
	   removed from the system or inserted into the dead queue of the
	   scheduler depends on its join type which was specified at spawning
	   time. If it has the attribute "PTH_ATTR_JOINABLE" set to "FALSE",
	   it's immediately removed and value is ignored. Else the thread is
	   inserted into the dead queue and value remembered for a subsequent
	   pth_join(3) call by another thread.

   Utilities
       Utility functions.

       int pth_fdmode(int fd, int mode);
	   This switches the non-blocking mode flag on file descriptor fd.
	   The argument mode can be "PTH_FDMODE_BLOCK" for switching fd into
	   blocking I/O mode, "PTH_FDMODE_NONBLOCK" for switching fd into non-
	   blocking I/O mode or "PTH_FDMODE_POLL" for just polling the current
	   mode. The current mode is returned (either "PTH_FDMODE_BLOCK" or
	   "PTH_FDMODE_NONBLOCK") or "PTH_FDMODE_ERROR" on error. Keep in mind
	   that since Pth 1.1 there is no longer a requirement to manually
	   switch a file descriptor into non-blocking mode in order to use it.
	   This is automatically done temporarily inside Pth.  Instead when
	   you now switch a file descriptor explicitly into non-blocking mode,
	   pth_read(3) or pth_write(3) will never block the current thread.

       pth_time_t pth_time(long sec, long usec);
	   This is a constructor for a "pth_time_t" structure which is a
	   convenient function to avoid temporary structure values. It returns
	   a pth_time_t structure which holds the absolute time value
	   specified by sec and usec.

       pth_time_t pth_timeout(long sec, long usec);
	   This is a constructor for a "pth_time_t" structure which is a
	   convenient function to avoid temporary structure values.  It
	   returns a pth_time_t structure which holds the absolute time value
	   calculated by adding sec and usec to the current time.

       void pth_int_time(struct timespec *sp);
	   Returns the current time of the pthsem internal time base.

       Sfdisc_t *pth_sfiodisc(void);
	   This functions is always available, but only reasonably usable when
	   Pth was built with Sfio support ("--with-sfio" option) and
	   "PTH_EXT_SFIO" is then defined by "pth.h". It is useful for
	   applications which want to use the comprehensive Sfio I/O library
	   with the Pth threading library. Then this function can be used to
	   get an Sfio discipline structure ("Sfdisc_t") which can be pushed
	   onto Sfio streams ("Sfio_t") in order to let this stream use
	   pth_read(3)/pth_write(2) instead of read(2)/write(2). The benefit
	   is that this way I/O on the Sfio stream does only block the current
	   thread instead of the whole process. The application has to free(3)
	   the "Sfdisc_t" structure when it is no longer needed. The Sfio
	   package can be found at http://www.research.att.com/sw/tools/sfio/.

   Cancellation Management
       Pth supports POSIX style thread cancellation via pth_cancel(3) and the
       following two related functions:

       void pth_cancel_state(int newstate, int *oldstate);
	   This manages the cancellation state of the current thread.  When
	   oldstate is not "NULL" the function stores the old cancellation
	   state under the variable pointed to by oldstate. When newstate is
	   not 0 it sets the new cancellation state. oldstate is created
	   before newstate is set.  A state is a combination of
	   "PTH_CANCEL_ENABLE" or "PTH_CANCEL_DISABLE" and
	   "PTH_CANCEL_DEFERRED" or "PTH_CANCEL_ASYNCHRONOUS".
	   "PTH_CANCEL_ENABLE|PTH_CANCEL_DEFERRED" (or "PTH_CANCEL_DEFAULT")
	   is the default state where cancellation is possible but only at
	   cancellation points.	 Use "PTH_CANCEL_DISABLE" to complete disable
	   cancellation for a thread and "PTH_CANCEL_ASYNCHRONOUS" for
	   allowing asynchronous cancellations, i.e., cancellations which can
	   happen at any time.

       void pth_cancel_point(void);
	   This explicitly enter a cancellation point. When the current
	   cancellation state is "PTH_CANCEL_DISABLE" or no cancellation
	   request is pending, this has no side-effect and returns
	   immediately. Else it calls `"pth_exit(PTH_CANCELED)"'.

   Event Handling
       Pth has a very flexible event facility which is linked into the
       scheduler through the pth_wait(3) function. The following functions
       provide the handling of event rings.

       pth_event_t pth_event(unsigned long spec, ...);
	   This creates a new event ring consisting of a single initial event.
	   The type of the generated event is specified by spec. The following
	   types are available:

	   "PTH_EVENT_FD"
	       This is a file descriptor event. One or more of
	       "PTH_UNTIL_FD_READABLE", "PTH_UNTIL_FD_WRITEABLE" or
	       "PTH_UNTIL_FD_EXCEPTION" have to be OR-ed into spec to specify
	       on which state of the file descriptor you want to wait.	The
	       file descriptor itself has to be given as an additional
	       argument.  Example:
	       `"pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_READABLE, fd)"'.

	   "PTH_EVENT_SELECT"
	       This is a multiple file descriptor event modeled directly after
	       the select(2) call (actually it is also used to implement
	       pth_select(3) internally).  It's a convenient way to wait for a
	       large set of file descriptors at once and at each file
	       descriptor for a different type of state. Additionally as a
	       nice side-effect one receives the number of file descriptors
	       which causes the event to be occurred (using BSD semantics,
	       i.e., when a file descriptor occurred in two sets it's counted
	       twice). The arguments correspond directly to the select(2)
	       function arguments except that there is no timeout argument
	       (because timeouts already can be handled via "PTH_EVENT_TIME"
	       events).

	       Example: `"pth_event(PTH_EVENT_SELECT, &rc, nfd, rfds, wfds,
	       efds)"' where "rc" has to be of type `"int *"', "nfd" has to be
	       of type `"int"' and "rfds", "wfds" and "efds" have to be of
	       type `"fd_set *"' (see select(2)). The number of occurred file
	       descriptors are stored in "rc".

	   "PTH_EVENT_SIGS"
	       This is a signal set event. The two additional arguments have
	       to be a pointer to a signal set (type `"sigset_t *"') and a
	       pointer to a signal number variable (type `"int *"').  This
	       event waits until one of the signals in the signal set
	       occurred.  As a result the occurred signal number is stored in
	       the second additional argument. Keep in mind that the Pth
	       scheduler doesn't block signals automatically.  So when you
	       want to wait for a signal with this event you've to block it
	       via sigprocmask(2) or it will be delivered without your notice.
	       Example: `"sigemptyset(&set); sigaddset(&set, SIGINT);
	       pth_event(PTH_EVENT_SIG, &set, &sig);"'.

	   "PTH_EVENT_TIME"
	       This is a time point event. The additional argument has to be
	       of type "pth_time_t" (usually on-the-fly generated via
	       pth_time(3)). This events waits until the specified time point
	       has elapsed. Keep in mind that the value is an absolute time
	       point and not an offset. When you want to wait for a specified
	       amount of time, you've to add the current time to the offset
	       (usually on-the-fly achieved via pth_timeout(3)).  Example:
	       `"pth_event(PTH_EVENT_TIME, pth_timeout(2,0))"'.

	   "PTH_EVENT_RTIME"
	       This is a time interval event. The additional argument has to
	       be of type "pth_time_t" (usually on-the-fly generated via
	       pth_time(3)), containing a time interval. During creation, it
	       is converted into PTH_EVENT_TIME. It has the advantage, that it
	       only uses the pthsem internal clock. Example:
	       `"pth_event(PTH_EVENT_TIME, pth_timeout(2,0))"'.	 is equal to
	       `"pth_event(PTH_EVENT_RTIME, pth_time(2,0))"'.

	   "PTH_EVENT_MSG"
	       This is a message port event. The additional argument has to be
	       of type "pth_msgport_t". This events waits until one or more
	       messages were received on the specified message port.  Example:
	       `"pth_event(PTH_EVENT_MSG, mp)"'.

	   "PTH_EVENT_TID"
	       This is a thread event. The additional argument has to be of
	       type "pth_t".  One of "PTH_UNTIL_TID_NEW",
	       "PTH_UNTIL_TID_READY", "PTH_UNTIL_TID_WAITING" or
	       "PTH_UNTIL_TID_DEAD" has to be OR-ed into spec to specify on
	       which state of the thread you want to wait.  Example:
	       `"pth_event(PTH_EVENT_TID|PTH_UNTIL_TID_DEAD, tid)"'.

	   "PTH_EVENT_FUNC"
	       This is a custom callback function event. Three additional
	       arguments have to be given with the following types: `"int
	       (*)(void *)"', `"void *"' and `"pth_time_t"'. The first is a
	       function pointer to a check function and the second argument is
	       a user-supplied context value which is passed to this function.
	       The scheduler calls this function on a regular basis (on his
	       own scheduler stack, so be very careful!) and the thread is
	       kept sleeping while the function returns "FALSE". Once it
	       returned "TRUE" the thread will be awakened. The check interval
	       is defined by the third argument, i.e., the check function is
	       polled again not until this amount of time elapsed. Example:
	       `"pth_event(PTH_EVENT_FUNC, func, arg, pth_time(0,500000))"'.

	   "PTH_EVENT_SEM"
	       This is a semaphore event. It waits for a semaphore, until it
	       can be decremented. By default 1 is used for this, with the
	       flag "PTH_UNTIL_COUNT" other values can be used. If the flag
	       "PTH_UNTIL_DECREMENT" is used, the semaphore value is
	       decremented (so the lock is obtained), else the event is
	       signaled, if it would be possible. Examples:

	       * pth_event(PTH_EVENT_SEM|PTH_UNTIL_DECREMENT|PTH_UNTIL_COUNT,
	       &sem,2): event waits, utils the value of the semaphore is >= 2
	       and subtracts then two from it

	       * pth_event(PTH_EVENT_SEM|PTH_UNTIL_COUNT, &sem,2): event
	       waits, util the value of the semaphore is >= 2

	       * pth_event(PTH_EVENT_SEM|PTH_UNTIL_DECREMENT, &sem): event
	       waits, util the value of the semaphore is >= 1 and subtracts
	       then 1 from it

	       * pth_event(PTH_EVENT_SEM, &sem): event waits, util the value
	       of the semaphore is >= 1

       unsigned long pth_event_typeof(pth_event_t ev);
	   This returns the type of event ev. It's a combination of the
	   describing "PTH_EVENT_XX" and "PTH_UNTIL_XX" value. This is
	   especially useful to know which arguments have to be supplied to
	   the pth_event_extract(3) function.

       int pth_event_extract(pth_event_t ev, ...);
	   When pth_event(3) is treated like sprintf(3), then this function is
	   sscanf(3), i.e., it is the inverse operation of pth_event(3). This
	   means that it can be used to extract the ingredients of an event.
	   The ingredients are stored into variables which are given as
	   pointers on the variable argument list.  Which pointers have to be
	   present depends on the event type and has to be determined by the
	   caller before via pth_event_typeof(3).

	   To make it clear, when you constructed ev via `"ev =
	   pth_event(PTH_EVENT_FD, fd);"' you have to extract it via
	   `"pth_event_extract(ev, &fd)"', etc. For multiple arguments of an
	   event the order of the pointer arguments is the same as for
	   pth_event(3). But always keep in mind that you have to always
	   supply pointers to variables and these variables have to be of the
	   same type as the argument of pth_event(3) required.

       pth_event_t pth_event_concat(pth_event_t ev, ...);
	   This concatenates one or more additional event rings to the event
	   ring ev and returns ev. The end of the argument list has to be
	   marked with a "NULL" argument. Use this function to create real
	   events rings out of the single-event rings created by pth_event(3).

       pth_event_t pth_event_isolate(pth_event_t ev);
	   This isolates the event ev from possibly appended events in the
	   event ring.	When in ev only one event exists, this returns "NULL".
	   When remaining events exists, they form a new event ring which is
	   returned.

       pth_event_t pth_event_walk(pth_event_t ev, int direction);
	   This walks to the next (when direction is "PTH_WALK_NEXT") or
	   previews (when direction is "PTH_WALK_PREV") event in the event
	   ring ev and returns this new reached event. Additionally
	   "PTH_UNTIL_OCCURRED" can be OR-ed into direction to walk to the
	   next/previous occurred event in the ring ev.

       pth_status_t pth_event_status(pth_event_t ev);
	   This returns the status of event ev. This is a fast operation
	   because only a tag on ev is checked which was either set or still
	   not set by the scheduler. In other words: This doesn't check the
	   event itself, it just checks the last knowledge of the scheduler.
	   The possible returned status codes are: "PTH_STATUS_PENDING" (event
	   is still pending), "PTH_STATUS_OCCURRED" (event successfully
	   occurred), "PTH_STATUS_FAILED" (event failed).

       int pth_event_free(pth_event_t ev, int mode);
	   This deallocates the event ev (when mode is "PTH_FREE_THIS") or all
	   events appended to the event ring under ev (when mode is
	   "PTH_FREE_ALL").

   Key-Based Storage
       The following functions provide thread-local storage through unique
       keys similar to the POSIX Pthread API. Use this for thread specific
       global data.

       int pth_key_create(pth_key_t *key, void (*func)(void *));
	   This created a new unique key and stores it in key.	Additionally
	   func can specify a destructor function which is called on the
	   current threads termination with the key.

       int pth_key_delete(pth_key_t key);
	   This explicitly destroys a key key.

       int pth_key_setdata(pth_key_t key, const void *value);
	   This stores value under key.

       void *pth_key_getdata(pth_key_t key);
	   This retrieves the value under key.

   Message Port Communication
       The following functions provide message ports which can be used for
       efficient and flexible inter-thread communication.

       pth_msgport_t pth_msgport_create(const char *name);
	   This returns a pointer to a new message port. If name name is not
	   "NULL", the name can be used by other threads via
	   pth_msgport_find(3) to find the message port in case they do not
	   know directly the pointer to the message port.

       void pth_msgport_destroy(pth_msgport_t mp);
	   This destroys a message port mp. Before all pending messages on it
	   are replied to their origin message port.

       pth_msgport_t pth_msgport_find(const char *name);
	   This finds a message port in the system by name and returns the
	   pointer to it.

       int pth_msgport_pending(pth_msgport_t mp);
	   This returns the number of pending messages on message port mp.

       int pth_msgport_put(pth_msgport_t mp, pth_message_t *m);
	   This puts (or sends) a message m to message port mp.

       pth_message_t *pth_msgport_get(pth_msgport_t mp);
	   This gets (or receives) the top message from message port mp.
	   Incoming messages are always kept in a queue, so there can be more
	   pending messages, of course.

       int pth_msgport_reply(pth_message_t *m);
	   This replies a message m to the message port of the sender.

   Thread Cleanups
       Per-thread cleanup functions.

       int pth_cleanup_push(void (*handler)(void *), void *arg);
	   This pushes the routine handler onto the stack of cleanup routines
	   for the current thread.  These routines are called in LIFO order
	   when the thread terminates.

       int pth_cleanup_pop(int execute);
	   This pops the top-most routine from the stack of cleanup routines
	   for the current thread. When execute is "TRUE" the routine is
	   additionally called.

   Process Forking
       The following functions provide some special support for process
       forking situations inside the threading environment.

       int pth_atfork_push(void (*prepare)(void *), void (*)(void *parent),
       void (*)(void *child), void *arg);
	   This function declares forking handlers to be called before and
	   after pth_fork(3), in the context of the thread that called
	   pth_fork(3). The prepare handler is called before fork(2)
	   processing commences. The parent handler is called	after fork(2)
	   processing completes in the parent process.	The child handler is
	   called after fork(2) processing completed in the child process. If
	   no handling is desired at one or more of these three points, the
	   corresponding handler can be given as "NULL".  Each handler is
	   called with arg as the argument.

	   The order of calls to pth_atfork_push(3) is significant. The parent
	   and child handlers are called in the order in which they were
	   established by calls to pth_atfork_push(3), i.e., FIFO. The prepare
	   fork handlers are called in the opposite order, i.e., LIFO.

       int pth_atfork_pop(void);
	   This removes the top-most handlers on the forking handler stack
	   which were established with the last pth_atfork_push(3) call. It
	   returns "FALSE" when no more handlers couldn't be removed from the
	   stack.

       pid_t pth_fork(void);
	   This is a variant of fork(2) with the difference that the current
	   thread only is forked into a separate process, i.e., in the parent
	   process nothing changes while in the child process all threads are
	   gone except for the scheduler and the calling thread. When you
	   really want to duplicate all threads in the current process you
	   should use fork(2) directly. But this is usually not reasonable.
	   Additionally this function takes care of forking handlers as
	   established by pth_fork_push(3).

   Synchronization
       The following functions provide synchronization support via mutual
       exclusion locks (mutex), read-write locks (rwlock), condition variables
       (cond) and barriers (barrier). Keep in mind that in a non-preemptive
       threading system like Pth this might sound unnecessary at the first
       look, because a thread isn't interrupted by the system. Actually when
       you have a critical code section which doesn't contain any pth_xxx()
       functions, you don't need any mutex to protect it, of course.

       But when your critical code section contains any pth_xxx() function the
       chance is high that these temporarily switch to the scheduler. And this
       way other threads can make progress and enter your critical code
       section, too.  This is especially true for critical code sections which
       implicitly or explicitly use the event mechanism.

       int pth_mutex_init(pth_mutex_t *mutex);
	   This dynamically initializes a mutex variable of type
	   `"pth_mutex_t"'.  Alternatively one can also use static
	   initialization via `"pth_mutex_t mutex = PTH_MUTEX_INIT"'.

       int pth_mutex_acquire(pth_mutex_t *mutex, int try, pth_event_t ev);
	   This acquires a mutex mutex.	 If the mutex is already locked by
	   another thread, the current threads execution is suspended until
	   the mutex is unlocked again or additionally the extra events in ev
	   occurred (when ev is not "NULL").  Recursive locking is explicitly
	   supported, i.e., a thread is allowed to acquire a mutex more than
	   once before its released. But it then also has be released the same
	   number of times until the mutex is again lockable by others.	 When
	   try is "TRUE" this function never suspends execution. Instead it
	   returns "FALSE" with "errno" set to "EBUSY".

       int pth_mutex_release(pth_mutex_t *mutex);
	   This decrements the recursion locking count on mutex and when it is
	   zero it releases the mutex mutex.

       int pth_rwlock_init(pth_rwlock_t *rwlock);
	   This dynamically initializes a read-write lock variable of type
	   `"pth_rwlock_t"'.  Alternatively one can also use static
	   initialization via `"pth_rwlock_t rwlock = PTH_RWLOCK_INIT"'.

       int pth_rwlock_acquire(pth_rwlock_t *rwlock, int op, int try,
       pth_event_t ev);
	   This acquires a read-only (when op is "PTH_RWLOCK_RD") or a read-
	   write (when op is "PTH_RWLOCK_RW") lock rwlock. When the lock is
	   only locked by other threads in read-only mode, the lock succeeds.
	   But when one thread holds a read-write lock, all locking attempts
	   suspend the current thread until this lock is released again.
	   Additionally in ev events can be given to let the locking timeout,
	   etc. When try is "TRUE" this function never suspends execution.
	   Instead it returns "FALSE" with "errno" set to "EBUSY".

       int pth_rwlock_release(pth_rwlock_t *rwlock);
	   This releases a previously acquired (read-only or read-write) lock.

       int pth_cond_init(pth_cond_t *cond);
	   This dynamically initializes a condition variable variable of type
	   `"pth_cond_t"'.  Alternatively one can also use static
	   initialization via `"pth_cond_t cond = PTH_COND_INIT"'.

       int pth_cond_await(pth_cond_t *cond, pth_mutex_t *mutex, pth_event_t
       ev);
	   This awaits a condition situation. The caller has to follow the
	   semantics of the POSIX condition variables: mutex has to be
	   acquired before this function is called. The execution of the
	   current thread is then suspended either until the events in ev
	   occurred (when ev is not "NULL") or cond was notified by another
	   thread via pth_cond_notify(3).  While the thread is waiting, mutex
	   is released. Before it returns mutex is reacquired.

       int pth_cond_notify(pth_cond_t *cond, int broadcast);
	   This notified one or all threads which are waiting on cond.	When
	   broadcast is "TRUE" all thread are notified, else only a single
	   (unspecified) one.

       int pth_barrier_init(pth_barrier_t *barrier, int threshold);
	   This dynamically initializes a barrier variable of type
	   `"pth_barrier_t"'.  Alternatively one can also use static
	   initialization via `"pth_barrier_t barrier =
	   PTH_BARRIER_INIT("threadhold")"'.

       int pth_barrier_reach(pth_barrier_t *barrier);
	   This function reaches a barrier barrier. If this is the last thread
	   (as specified by threshold on init of barrier) all threads are
	   awakened.  Else the current thread is suspended until the last
	   thread reached the barrier and this way awakes all threads. The
	   function returns (beside "FALSE" on error) the value "TRUE" for any
	   thread which neither reached the barrier as the first nor the last
	   thread; "PTH_BARRIER_HEADLIGHT" for the thread which reached the
	   barrier as the first thread and "PTH_BARRIER_TAILLIGHT" for the
	   thread which reached the barrier as the last thread.

   Semaphore support
       The interface provides functions to set/get the value of a semaphore,
       increment it with arbitrary values, wait, until the value becomes
       bigger than a given value (without or with decrementing, if the
       condition becomes true.

       The data-type for the semaphore is names "pth_sem_t" and it has an
       initializer like "pth_cond_t".

       int pth_sem_init(pth_sem_t *sem);
	   This dynamically initializes a semaphore variable of type
	   `"pth_sem_t"'.  Alternatively one can also use static
	   initialization via `"pth_sem_t semaphore = PTH_SEM_INIT"'.

       int pth_sem_dec(pth_sem_t *sem);
	   waits, until the value of "sem" is >= 1 and decrement it.

       int pth_sem_dec_value(pth_sem_t *sem, unsigned value);
	   waits, until the value of "sem" is >= "value" and subtracts
	   "value".

       int pth_sem_inc(pth_sem_t *sem, int notify);
	   increments "sem". The scheduler is started, if "notify" is not
	   null.

       int pth_sem_inc_value(pth_sem_t *sem, unsigned value, int notify);
	   adds value to "sem". The scheduler is started, if "notify" is not
	   null.

       int pth_sem_set_value(pth_sem_t *sem, unsigned value);
	   sets the value of "sem" to "value".

       int pth_sem_get_value(pth_sem_t *sem, unsigned *value);
	   stores the value of "sem" in *"value".

   User-Space Context
       The following functions provide a stand-alone sub-API for user-space
       context switching. It internally is based on the same underlying
       machine context switching mechanism the threads in GNU Pth are based
       on.  Hence these functions you can use for implementing your own simple
       user-space threads. The "pth_uctx_t" context is somewhat modeled after
       POSIX ucontext(3).

       The time required to create (via pth_uctx_make(3)) a user-space context
       can range from just a few microseconds up to a more dramatical time
       (depending on the machine context switching method which is available
       on the platform). On the other hand, the raw performance in switching
       the user-space contexts is always very good (nearly independent of the
       used machine context switching method). For instance, on an Intel
       Pentium-III CPU with 800Mhz running under FreeBSD 4 one usually
       achieves about 260,000 user-space context switches (via
       pth_uctx_switch(3)) per second.

       int pth_uctx_create(pth_uctx_t *uctx);
	   This function creates a user-space context and stores it into uctx.
	   There is still no underlying user-space context configured. You
	   still have to do this with pth_uctx_make(3). On success, this
	   function returns "TRUE", else "FALSE".

       int pth_uctx_make(pth_uctx_t uctx, char *sk_addr, size_t sk_size, const
       sigset_t *sigmask, void (*start_func)(void *), void *start_arg,
       pth_uctx_t uctx_after);
	   This function makes a new user-space context in uctx which will
	   operate on the run-time stack sk_addr (which is of maximum size
	   sk_size), with the signals in sigmask blocked (if sigmask is not
	   "NULL") and starting to execute with the call
	   start_func(start_arg). If sk_addr is "NULL", a stack is dynamically
	   allocated. The stack size sk_size has to be at least 16384 (16KB).
	   If the start function start_func returns and uctx_after is not
	   "NULL", an implicit user-space context switch to this context is
	   performed. Else (if uctx_after is "NULL") the process is terminated
	   with exit(3). This function is somewhat modeled after POSIX
	   makecontext(3). On success, this function returns "TRUE", else
	   "FALSE".

       int pth_uctx_switch(pth_uctx_t uctx_from, pth_uctx_t uctx_to);
	   This function saves the current user-space context in uctx_from for
	   later restoring by another call to pth_uctx_switch(3) and restores
	   the new user-space context from uctx_to, which previously had to be
	   set with either a previous call to pth_uctx_switch(3) or initially
	   by pth_uctx_make(3). This function is somewhat modeled after POSIX
	   swapcontext(3). If uctx_from or uctx_to are "NULL" or if uctx_to
	   contains no valid user-space context, "FALSE" is returned instead
	   of "TRUE". These are the only errors possible.

       int pth_uctx_destroy(pth_uctx_t uctx);
	   This function destroys the user-space context in uctx. The run-time
	   stack associated with the user-space context is deallocated only if
	   it was not given by the application (see sk_addr of
	   pth_uctx_create(3)).	 If uctx is "NULL", "FALSE" is returned
	   instead of "TRUE". This is the only error possible.

   Generalized POSIX Replacement API
       The following functions are generalized replacements functions for the
       POSIX API, i.e., they are similar to the functions under `Standard
       POSIX Replacement API' but all have an additional event argument which
       can be used for timeouts, etc.

       int pth_sigwait_ev(const sigset_t *set, int *sig, pth_event_t ev);
	   This is equal to pth_sigwait(3) (see below), but has an additional
	   event argument ev. When pth_sigwait(3) suspends the current threads
	   execution it usually only uses the signal event on set to awake.
	   With this function any number of extra events can be used to awake
	   the current thread (remember that ev actually is an event ring).

       int pth_connect_ev(int s, const struct sockaddr *addr, socklen_t
       addrlen, pth_event_t ev);
	   This is equal to pth_connect(3) (see below), but has an additional
	   event argument ev. When pth_connect(3) suspends the current threads
	   execution it usually only uses the I/O event on s to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       int pth_accept_ev(int s, struct sockaddr *addr, socklen_t *addrlen,
       pth_event_t ev);
	   This is equal to pth_accept(3) (see below), but has an additional
	   event argument ev. When pth_accept(3) suspends the current threads
	   execution it usually only uses the I/O event on s to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       int pth_select_ev(int nfd, fd_set *rfds, fd_set *wfds, fd_set *efds,
       struct timeval *timeout, pth_event_t ev);
	   This is equal to pth_select(3) (see below), but has an additional
	   event argument ev. When pth_select(3) suspends the current threads
	   execution it usually only uses the I/O event on rfds, wfds and efds
	   to awake. With this function any number of extra events can be used
	   to awake the current thread (remember that ev actually is an event
	   ring).

       int pth_poll_ev(struct pollfd *fds, unsigned int nfd, int timeout,
       pth_event_t ev);
	   This is equal to pth_poll(3) (see below), but has an additional
	   event argument ev. When pth_poll(3) suspends the current threads
	   execution it usually only uses the I/O event on fds to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_read_ev(int fd, void *buf, size_t nbytes, pth_event_t ev);
	   This is equal to pth_read(3) (see below), but has an additional
	   event argument ev. When pth_read(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_readv_ev(int fd, const struct iovec *iovec, int iovcnt,
       pth_event_t ev);
	   This is equal to pth_readv(3) (see below), but has an additional
	   event argument ev. When pth_readv(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_write_ev(int fd, const void *buf, size_t nbytes,
       pth_event_t ev);
	   This is equal to pth_write(3) (see below), but has an additional
	   event argument ev. When pth_write(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_writev_ev(int fd, const struct iovec *iovec, int iovcnt,
       pth_event_t ev);
	   This is equal to pth_writev(3) (see below), but has an additional
	   event argument ev. When pth_writev(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_recv_ev(int fd, void *buf, size_t nbytes, int flags,
       pth_event_t ev);
	   This is equal to pth_recv(3) (see below), but has an additional
	   event argument ev. When pth_recv(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_recvfrom_ev(int fd, void *buf, size_t nbytes, int flags,
       struct sockaddr *from, socklen_t *fromlen, pth_event_t ev);
	   This is equal to pth_recvfrom(3) (see below), but has an additional
	   event argument ev. When pth_recvfrom(3) suspends the current
	   threads execution it usually only uses the I/O event on fd to
	   awake. With this function any number of extra events can be used to
	   awake the current thread (remember that ev actually is an event
	   ring).

       ssize_t pth_send_ev(int fd, const void *buf, size_t nbytes, int flags,
       pth_event_t ev);
	   This is equal to pth_send(3) (see below), but has an additional
	   event argument ev. When pth_send(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

       ssize_t pth_sendto_ev(int fd, const void *buf, size_t nbytes, int
       flags, const struct sockaddr *to, socklen_t tolen, pth_event_t ev);
	   This is equal to pth_sendto(3) (see below), but has an additional
	   event argument ev. When pth_sendto(3) suspends the current threads
	   execution it usually only uses the I/O event on fd to awake. With
	   this function any number of extra events can be used to awake the
	   current thread (remember that ev actually is an event ring).

   Standard POSIX Replacement API
       The following functions are standard replacements functions for the
       POSIX API.  The difference is mainly that they suspend the current
       thread only instead of the whole process in case the file descriptors
       will block.

       int pth_nanosleep(const struct timespec *rqtp, struct timespec *rmtp);
	   This is a variant of the POSIX nanosleep(3) function. It suspends
	   the current threads execution until the amount of time in rqtp
	   elapsed.  The thread is guaranteed to not wake up before this time,
	   but because of the non-preemptive scheduling nature of Pth, it can
	   be awakened later, of course. If rmtp is not "NULL", the "timespec"
	   structure it references is updated to contain the unslept amount
	   (the request time minus the time actually slept time). The
	   difference between nanosleep(3) and pth_nanosleep(3) is that that
	   pth_nanosleep(3) suspends only the execution of the current thread
	   and not the whole process.

       int pth_usleep(unsigned int usec);
	   This is a variant of the 4.3BSD usleep(3) function. It suspends the
	   current threads execution until usec microseconds (= usec*1/1000000
	   sec) elapsed.  The thread is guaranteed to not wake up before this
	   time, but because of the non-preemptive scheduling nature of Pth,
	   it can be awakened later, of course.	 The difference between
	   usleep(3) and pth_usleep(3) is that that pth_usleep(3) suspends
	   only the execution of the current thread and not the whole process.

       unsigned int pth_sleep(unsigned int sec);
	   This is a variant of the POSIX sleep(3) function. It suspends the
	   current threads execution until sec seconds elapsed.	 The thread is
	   guaranteed to not wake up before this time, but because of the non-
	   preemptive scheduling nature of Pth, it can be awakened later, of
	   course.  The difference between sleep(3) and pth_sleep(3) is that
	   pth_sleep(3) suspends only the execution of the current thread and
	   not the whole process.

       pid_t pth_waitpid(pid_t pid, int *status, int options);
	   This is a variant of the POSIX waitpid(2) function. It suspends the
	   current threads execution until status information is available for
	   a terminated child process pid.  The difference between waitpid(2)
	   and pth_waitpid(3) is that pth_waitpid(3) suspends only the
	   execution of the current thread and not the whole process.  For
	   more details about the arguments and return code semantics see
	   waitpid(2).

       int pth_system(const char *cmd);
	   This is a variant of the POSIX system(3) function. It executes the
	   shell command cmd with Bourne Shell ("sh") and suspends the current
	   threads execution until this command terminates. The difference
	   between system(3) and pth_system(3) is that pth_system(3) suspends
	   only the execution of the current thread and not the whole process.
	   For more details about the arguments and return code semantics see
	   system(3).

       int pth_sigmask(int how, const sigset_t *set, sigset_t *oset)
	   This is the Pth thread-related equivalent of POSIX sigprocmask(2)
	   respectively pthread_sigmask(3). The arguments how, set and oset
	   directly relate to sigprocmask(2), because Pth internally just uses
	   sigprocmask(2) here. So alternatively you can also directly call
	   sigprocmask(2), but for consistency reasons you should use this
	   function pth_sigmask(3).

       int pth_sigwait(const sigset_t *set, int *sig);
	   This is a variant of the POSIX.1c sigwait(3) function. It suspends
	   the current threads execution until a signal in set occurred and
	   stores the signal number in sig. The important point is that the
	   signal is not delivered to a signal handler. Instead it's caught by
	   the scheduler only in order to awake the pth_sigwait() call. The
	   trick and noticeable point here is that this way you get an
	   asynchronous aware application that is written completely
	   synchronously. When you think about the problem of asynchronous
	   safe functions you should recognize that this is a great benefit.

       int pth_connect(int s, const struct sockaddr *addr, socklen_t addrlen);
	   This is a variant of the 4.2BSD connect(2) function. It establishes
	   a connection on a socket s to target specified in addr and addrlen.
	   The difference between connect(2) and pth_connect(3) is that
	   pth_connect(3) suspends only the execution of the current thread
	   and not the whole process.  For more details about the arguments
	   and return code semantics see connect(2).

       int pth_accept(int s, struct sockaddr *addr, socklen_t *addrlen);
	   This is a variant of the 4.2BSD accept(2) function. It accepts a
	   connection on a socket by extracting the first connection request
	   on the queue of pending connections, creating a new socket with the
	   same properties of s and allocates a new file descriptor for the
	   socket (which is returned).	The difference between accept(2) and
	   pth_accept(3) is that pth_accept(3) suspends only the execution of
	   the current thread and not the whole process.  For more details
	   about the arguments and return code semantics see accept(2).

       int pth_select(int nfd, fd_set *rfds, fd_set *wfds, fd_set *efds,
       struct timeval *timeout);
	   This is a variant of the 4.2BSD select(2) function.	It examines
	   the I/O descriptor sets whose addresses are passed in rfds, wfds,
	   and efds to see if some of their descriptors are ready for reading,
	   are ready for writing, or have an exceptional condition pending,
	   respectively.  For more details about the arguments and return code
	   semantics see select(2).

       int pth_pselect(int nfd, fd_set *rfds, fd_set *wfds, fd_set *efds,
       const struct timespec *timeout, const sigset_t *sigmask);
	   This is a variant of the POSIX pselect(2) function, which in turn
	   is a stronger variant of 4.2BSD select(2). The difference is that
	   the higher-resolution "struct timespec" is passed instead of the
	   lower-resolution "struct timeval" and that a signal mask is
	   specified which is temporarily set while waiting for input. For
	   more details about the arguments and return code semantics see
	   pselect(2) and select(2).

       int pth_poll(struct pollfd *fds, unsigned int nfd, int timeout);
	   This is a variant of the SysV poll(2) function. It examines the I/O
	   descriptors which are passed in the array fds to see if some of
	   them are ready for reading, are ready for writing, or have an
	   exceptional condition pending, respectively. For more details about
	   the arguments and return code semantics see poll(2).

       ssize_t pth_read(int fd, void *buf, size_t nbytes);
	   This is a variant of the POSIX read(2) function. It reads up to
	   nbytes bytes into buf from file descriptor fd.  The difference
	   between read(2) and pth_read(2) is that pth_read(2) suspends
	   execution of the current thread until the file descriptor is ready
	   for reading. For more details about the arguments and return code
	   semantics see read(2).

       ssize_t pth_readv(int fd, const struct iovec *iovec, int iovcnt);
	   This is a variant of the POSIX readv(2) function. It reads data
	   from file descriptor fd into the first iovcnt rows of the iov
	   vector.  The difference between readv(2) and pth_readv(2) is that
	   pth_readv(2) suspends execution of the current thread until the
	   file descriptor is ready for reading. For more details about the
	   arguments and return code semantics see readv(2).

       ssize_t pth_write(int fd, const void *buf, size_t nbytes);
	   This is a variant of the POSIX write(2) function. It writes nbytes
	   bytes from buf to file descriptor fd.  The difference between
	   write(2) and pth_write(2) is that pth_write(2) suspends execution
	   of the current thread until the file descriptor is ready for
	   writing.  For more details about the arguments and return code
	   semantics see write(2).

       ssize_t pth_writev(int fd, const struct iovec *iovec, int iovcnt);
	   This is a variant of the POSIX writev(2) function. It writes data
	   to file descriptor fd from the first iovcnt rows of the iov vector.
	   The difference between writev(2) and pth_writev(2) is that
	   pth_writev(2) suspends execution of the current thread until the
	   file descriptor is ready for reading. For more details about the
	   arguments and return code semantics see writev(2).

       ssize_t pth_pread(int fd, void *buf, size_t nbytes, off_t offset);
	   This is a variant of the POSIX pread(3) function.  It performs the
	   same action as a regular read(2), except that it reads from a given
	   position in the file without changing the file pointer.  The first
	   three arguments are the same as for pth_read(3) with the addition
	   of a fourth argument offset for the desired position inside the
	   file.

       ssize_t pth_pwrite(int fd, const void *buf, size_t nbytes, off_t
       offset);
	   This is a variant of the POSIX pwrite(3) function.  It performs the
	   same action as a regular write(2), except that it writes to a given
	   position in the file without changing the file pointer. The first
	   three arguments are the same as for pth_write(3) with the addition
	   of a fourth argument offset for the desired position inside the
	   file.

       ssize_t pth_recv(int fd, void *buf, size_t nbytes, int flags);
	   This is a variant of the SUSv2 recv(2) function and equal to
	   ``pth_recvfrom(fd, buf, nbytes, flags, NULL, 0)''.

       ssize_t pth_recvfrom(int fd, void *buf, size_t nbytes, int flags,
       struct sockaddr *from, socklen_t *fromlen);
	   This is a variant of the SUSv2 recvfrom(2) function. It reads up to
	   nbytes bytes into buf from file descriptor fd while using flags and
	   from/fromlen. The difference between recvfrom(2) and
	   pth_recvfrom(2) is that pth_recvfrom(2) suspends execution of the
	   current thread until the file descriptor is ready for reading. For
	   more details about the arguments and return code semantics see
	   recvfrom(2).

       ssize_t pth_send(int fd, const void *buf, size_t nbytes, int flags);
	   This is a variant of the SUSv2 send(2) function and equal to
	   ``pth_sendto(fd, buf, nbytes, flags, NULL, 0)''.

       ssize_t pth_sendto(int fd, const void *buf, size_t nbytes, int flags,
       const struct sockaddr *to, socklen_t tolen);
	   This is a variant of the SUSv2 sendto(2) function. It writes nbytes
	   bytes from buf to file descriptor fd while using flags and
	   to/tolen. The difference between sendto(2) and pth_sendto(2) is
	   that pth_sendto(2) suspends execution of the current thread until
	   the file descriptor is ready for writing. For more details about
	   the arguments and return code semantics see sendto(2).

EXAMPLE
       The following example is a useless server which does nothing more than
       listening on TCP port 12345 and displaying the current time to the
       socket when a connection was established. For each incoming connection
       a thread is spawned. Additionally, to see more multithreading, a
       useless ticker thread runs simultaneously which outputs the current
       time to "stderr" every 5 seconds. The example contains no error
       checking and is only intended to show you the look and feel of Pth.

	#include <stdio.h>
	#include <stdlib.h>
	#include <errno.h>
	#include <sys/types.h>
	#include <sys/socket.h>
	#include <netinet/in.h>
	#include <arpa/inet.h>
	#include <signal.h>
	#include <netdb.h>
	#include <unistd.h>
	#include "pth.h"

	#define PORT 12345

	/* the socket connection handler thread */
	static void *handler(void *_arg)
	{
	    int fd = (int)_arg;
	    time_t now;
	    char *ct;

	    now = time(NULL);
	    ct = ctime(&now);
	    pth_write(fd, ct, strlen(ct));
	    close(fd);
	    return NULL;
	}

	/* the stderr time ticker thread */
	static void *ticker(void *_arg)
	{
	    time_t now;
	    char *ct;
	    float load;

	    for (;;) {
		pth_sleep(5);
		now = time(NULL);
		ct = ctime(&now);
		ct[strlen(ct)-1] = '\0';
		pth_ctrl(PTH_CTRL_GETAVLOAD, &load);
		printf("ticker: time: %s, average load: %.2f\n", ct, load);
	    }
	}

	/* the main thread/procedure */
	int main(int argc, char *argv[])
	{
	    pth_attr_t attr;
	    struct sockaddr_in sar;
	    struct protoent *pe;
	    struct sockaddr_in peer_addr;
	    int peer_len;
	    int sa, sw;
	    int port;

	    pth_init();
	    signal(SIGPIPE, SIG_IGN);

	    attr = pth_attr_new();
	    pth_attr_set(attr, PTH_ATTR_NAME, "ticker");
	    pth_attr_set(attr, PTH_ATTR_STACK_SIZE, 64*1024);
	    pth_attr_set(attr, PTH_ATTR_JOINABLE, FALSE);
	    pth_spawn(attr, ticker, NULL);

	    pe = getprotobyname("tcp");
	    sa = socket(AF_INET, SOCK_STREAM, pe->p_proto);
	    sar.sin_family = AF_INET;
	    sar.sin_addr.s_addr = INADDR_ANY;
	    sar.sin_port = htons(PORT);
	    bind(sa, (struct sockaddr *)&sar, sizeof(struct sockaddr_in));
	    listen(sa, 10);

	    pth_attr_set(attr, PTH_ATTR_NAME, "handler");
	    for (;;) {
		peer_len = sizeof(peer_addr);
		sw = pth_accept(sa, (struct sockaddr *)&peer_addr, &peer_len);
		pth_spawn(attr, handler, (void *)sw);
	    }
	}

BUILD ENVIRONMENTS
       In this section we will discuss the canonical ways to establish the
       build environment for a Pth based program. The possibilities supported
       by Pth range from very simple environments to rather complex ones.

   Manual Build Environment (Novice)
       As a first example, assume we have the above test program staying in
       the source file "foo.c". Then we can create a very simple build
       environment by just adding the following "Makefile":

	$ vi Makefile
	| CC	  = cc
	| CFLAGS  = `pth-config --cflags`
	| LDFLAGS = `pth-config --ldflags`
	| LIBS	  = `pth-config --libs`
	|
	| all: foo
	| foo: foo.o
	|     $(CC) $(LDFLAGS) -o foo foo.o $(LIBS)
	| foo.o: foo.c
	|     $(CC) $(CFLAGS) -c foo.c
	| clean:
	|     rm -f foo foo.o

       This imports the necessary compiler and linker flags on-the-fly from
       the Pth installation via its "pth-config" program. This approach is
       straight-forward and works fine for small projects.

   Autoconf Build Environment (Advanced)
       The previous approach is simple but inflexible. First, to speed up
       building, it would be nice to not expand the compiler and linker flags
       every time the compiler is started. Second, it would be useful to also
       be able to build against uninstalled Pth, that is, against a Pth source
       tree which was just configured and built, but not installed. Third, it
       would be also useful to allow checking of the Pth version to make sure
       it is at least a minimum required version.  And finally, it would be
       also great to make sure Pth works correctly by first performing some
       sanity compile and run-time checks. All this can be done if we use GNU
       autoconf and the "AC_CHECK_PTH" macro provided by Pth. For this, we
       establish the following three files:

       First we again need the "Makefile", but this time it contains autoconf
       placeholders and additional cleanup targets. And we create it under the
       name "Makefile.in", because it is now an input file for autoconf:

	$ vi Makefile.in
	| CC	  = @CC@
	| CFLAGS  = @CFLAGS@
	| LDFLAGS = @LDFLAGS@
	| LIBS	  = @LIBS@
	|
	| all: foo
	| foo: foo.o
	|     $(CC) $(LDFLAGS) -o foo foo.o $(LIBS)
	| foo.o: foo.c
	|     $(CC) $(CFLAGS) -c foo.c
	| clean:
	|     rm -f foo foo.o
	| distclean:
	|     rm -f foo foo.o
	|     rm -f config.log config.status config.cache
	|     rm -f Makefile

       Because autoconf generates additional files, we added a canonical
       "distclean" target which cleans this up. Secondly, we wrote
       "configure.ac", a (minimal) autoconf script specification:

	$ vi configure.ac
	| AC_INIT(Makefile.in)
	| AC_CHECK_PTH(1.3.0)
	| AC_OUTPUT(Makefile)

       Then we let autoconf's "aclocal" program generate for us an
       "aclocal.m4" file containing Pth's "AC_CHECK_PTH" macro. Then we
       generate the final "configure" script out of this "aclocal.m4" file and
       the "configure.ac" file:

	$ aclocal --acdir=`pth-config --acdir`
	$ autoconf

       After these steps, the working directory should look similar to this:

	$ ls -l
	-rw-r--r--  1 rse  users    176 Nov  3 11:11 Makefile.in
	-rw-r--r--  1 rse  users  15314 Nov  3 11:16 aclocal.m4
	-rwxr-xr-x  1 rse  users  52045 Nov  3 11:16 configure
	-rw-r--r--  1 rse  users     63 Nov  3 11:11 configure.ac
	-rw-r--r--  1 rse  users   4227 Nov  3 11:11 foo.c

       If we now run "configure" we get a correct "Makefile" which immediately
       can be used to build "foo" (assuming that Pth is already installed
       somewhere, so that "pth-config" is in $PATH):

	$ ./configure
	creating cache ./config.cache
	checking for gcc... gcc
	checking whether the C compiler (gcc   ) works... yes
	checking whether the C compiler (gcc   ) is a cross-compiler... no
	checking whether we are using GNU C... yes
	checking whether gcc accepts -g... yes
	checking how to run the C preprocessor... gcc -E
	checking for GNU Pth... version 1.3.0, installed under /usr/local
	updating cache ./config.cache
	creating ./config.status
	creating Makefile
	rse@en1:/e/gnu/pth/ac
	$ make
	gcc -g -O2 -I/usr/local/include -c foo.c
	gcc -L/usr/local/lib -o foo foo.o -lpth

       If Pth is installed in non-standard locations or "pth-config" is not in
       $PATH, one just has to drop the "configure" script a note about the
       location by running "configure" with the option "--with-pth="dir (where
       dir is the argument which was used with the "--prefix" option when Pth
       was installed).

   Autoconf Build Environment with Local Copy of Pth (Expert)
       Finally let us assume the "foo" program stays under either a GPL or
       LGPL distribution license and we want to make it a stand-alone package
       for easier distribution and installation.  That is, we don't want to
       oblige the end-user to install Pth just to allow our "foo" package to
       compile. For this, it is a convenient practice to include the required
       libraries (here Pth) into the source tree of the package (here "foo").
       Pth ships with all necessary support to allow us to easily achieve this
       approach. Say, we want Pth in a subdirectory named "pth/" and this
       directory should be seamlessly integrated into the configuration and
       build process of "foo".

       First we again start with the "Makefile.in", but this time it is a more
       advanced version which supports subdirectory movement:

	$ vi Makefile.in
	| CC	  = @CC@
	| CFLAGS  = @CFLAGS@
	| LDFLAGS = @LDFLAGS@
	| LIBS	  = @LIBS@
	|
	| SUBDIRS = pth
	|
	| all: subdirs_all foo
	|
	| subdirs_all:
	|     @$(MAKE) $(MFLAGS) subdirs TARGET=all
	| subdirs_clean:
	|     @$(MAKE) $(MFLAGS) subdirs TARGET=clean
	| subdirs_distclean:
	|     @$(MAKE) $(MFLAGS) subdirs TARGET=distclean
	| subdirs:
	|     @for subdir in $(SUBDIRS); do \
	|	  echo "===> $$subdir ($(TARGET))"; \
	|	  (cd $$subdir; $(MAKE) $(MFLAGS) $(TARGET) || exit 1) || exit 1; \
	|	  echo "<=== $$subdir"; \
	|     done
	|
	| foo: foo.o
	|     $(CC) $(LDFLAGS) -o foo foo.o $(LIBS)
	| foo.o: foo.c
	|     $(CC) $(CFLAGS) -c foo.c
	|
	| clean: subdirs_clean
	|     rm -f foo foo.o
	| distclean: subdirs_distclean
	|     rm -f foo foo.o
	|     rm -f config.log config.status config.cache
	|     rm -f Makefile

       Then we create a slightly different autoconf script "configure.ac":

	$ vi configure.ac
	| AC_INIT(Makefile.in)
	| AC_CONFIG_AUX_DIR(pth)
	| AC_CHECK_PTH(1.3.0, subdir:pth --disable-tests)
	| AC_CONFIG_SUBDIRS(pth)
	| AC_OUTPUT(Makefile)

       Here we provided a default value for "foo"'s "--with-pth" option as the
       second argument to "AC_CHECK_PTH" which indicates that Pth can be found
       in the subdirectory named "pth/". Additionally we specified that the
       "--disable-tests" option of Pth should be passed to the "pth/"
       subdirectory, because we need only to build the Pth library itself. And
       we added a "AC_CONFIG_SUBDIR" call which indicates to autoconf that it
       should configure the "pth/" subdirectory, too. The "AC_CONFIG_AUX_DIR"
       directive was added just to make autoconf happy, because it wants to
       find a "install.sh" or "shtool" script if "AC_CONFIG_SUBDIRS" is used.

       Now we let autoconf's "aclocal" program again generate for us an
       "aclocal.m4" file with the contents of Pth's "AC_CHECK_PTH" macro.
       Finally we generate the "configure" script out of this "aclocal.m4"
       file and the "configure.ac" file.

	$ aclocal --acdir=`pth-config --acdir`
	$ autoconf

       Now we have to create the "pth/" subdirectory itself. For this, we
       extract the Pth distribution to the "foo" source tree and just rename
       it to "pth/":

	$ gunzip <pth-X.Y.Z.tar.gz | tar xvf -
	$ mv pth-X.Y.Z pth

SYSTEM CALL WRAPPER FACILITY
       Pth per default uses an explicit API, including the system calls. For
       instance you've to explicitly use pth_read(3) when you need a thread-
       aware read(3) and cannot expect that by just calling read(3) only the
       current thread is blocked. Instead with the standard read(3) call the
       whole process will be blocked. But because for some applications
       (mainly those consisting of lots of third-party stuff) this can be
       inconvenient.  Here it's required that a call to read(3) `magically'
       means pth_read(3). The problem here is that such magic Pth cannot
       provide per default because it's not really portable.  Nevertheless Pth
       provides a two step approach to solve this problem:

   Soft System Call Mapping
       This variant is available on all platforms and can always be enabled by
       building Pth with "--enable-syscall-soft". This then triggers some
       "#define"'s in the "pth.h" header which map for instance read(3) to
       pth_read(3), etc. Currently the following functions are mapped:
       fork(2), nanosleep(3), usleep(3), sleep(3), sigwait(3), waitpid(2),
       system(3), select(2), poll(2), connect(2), accept(2), read(2),
       write(2), recv(2), send(2), recvfrom(2), sendto(2).

       The drawback of this approach is just that really all source files of
       the application where these function calls occur have to include
       "pth.h", of course. And this also means that existing libraries,
       including the vendor's stdio, usually will still block the whole
       process if one of its I/O functions block.

   Hard System Call Mapping
       This variant is available only on those platforms where the syscall(2)
       function exists and there it can be enabled by building Pth with
       "--enable-syscall-hard". This then builds wrapper functions (for
       instances read(3)) into the Pth library which internally call the real
       Pth replacement functions (pth_read(3)). Currently the following
       functions are mapped: fork(2), nanosleep(3), usleep(3), sleep(3),
       waitpid(2), system(3), select(2), poll(2), connect(2), accept(2),
       read(2), write(2).

       The drawback of this approach is that it depends on syscall(2)
       interface and prototype conflicts can occur while building the wrapper
       functions due to different function signatures in the vendor C header
       files.  But the advantage of this mapping variant is that the source
       files of the application where these function calls occur have not to
       include "pth.h" and that existing libraries, including the vendor's
       stdio, magically become thread-aware (and then block only the current
       thread).

IMPLEMENTATION NOTES
       Pth is very portable because it has only one part which perhaps has to
       be ported to new platforms (the machine context initialization). But it
       is written in a way which works on mostly all Unix platforms which
       support makecontext(2) or at least sigstack(2) or sigaltstack(2) [see
       "pth_mctx.c" for details]. Any other Pth code is POSIX and ANSI C based
       only.

       The context switching is done via either SUSv2 makecontext(2) or POSIX
       make[sig]setjmp(3) and [sig]longjmp(3). Here all CPU registers, the
       program counter and the stack pointer are switched. Additionally the
       Pth dispatcher switches also the global Unix "errno" variable [see
       "pth_mctx.c" for details] and the signal mask (either implicitly via
       sigsetjmp(3) or in an emulated way via explicit setprocmask(2) calls).

       The Pth event manager is mainly select(2) and gettimeofday(2) based,
       i.e., the current time is fetched via gettimeofday(2) once per context
       switch for time calculations and all I/O events are implemented via a
       single central select(2) call [see "pth_sched.c" for details].

       The thread control block management is done via virtual priority queues
       without any additional data structure overhead. For this, the queue
       linkage attributes are part of the thread control blocks and the queues
       are actually implemented as rings with a selected element as the entry
       point [see "pth_tcb.h" and "pth_pqueue.c" for details].

       Most time critical code sections (especially the dispatcher and event
       manager) are speeded up by inline functions (implemented as ANSI C pre-
       processor macros). Additionally any debugging code is completely
       removed from the source when not built with "-DPTH_DEBUG" (see Autoconf
       "--enable-debug" option), i.e., not only stub functions remain [see
       "pth_debug.c" for details].

RESTRICTIONS
       Pth (intentionally) provides no replacements for non-thread-safe
       functions (like strtok(3) which uses a static internal buffer) or
       synchronous system functions (like gethostbyname(3) which doesn't
       provide an asynchronous mode where it doesn't block). When you want to
       use those functions in your server application together with threads,
       you've to either link the application against special third-party
       libraries (or for thread-safe/reentrant functions possibly against an
       existing "libc_r" of the platform vendor). For an asynchronous DNS
       resolver library use the GNU adns package from Ian Jackson ( see
       http://www.gnu.org/software/adns/adns.html ).

HISTORY
       The Pth library was designed and implemented between February and July
       1999 by Ralf S. Engelschall after evaluating numerous (mostly
       preemptive) thread libraries and after intensive discussions with Peter
       Simons, Martin Kraemer, Lars Eilebrecht and Ralph Babel related to an
       experimental (matrix based) non-preemptive C++ scheduler class written
       by Peter Simons.

       Pth was then implemented in order to combine the non-preemptive
       approach of multithreading (which provides better portability and
       performance) with an API similar to the popular one found in Pthread
       libraries (which provides easy programming).

       So the essential idea of the non-preemptive approach was taken over
       from Peter Simons scheduler. The priority based scheduling algorithm
       was suggested by Martin Kraemer. Some code inspiration also came from
       an experimental threading library (rsthreads) written by Robert S. Thau
       for an ancient internal test version of the Apache webserver.  The
       concept and API of message ports was borrowed from AmigaOS' Exec
       subsystem. The concept and idea for the flexible event mechanism came
       from Paul Vixie's eventlib (which can be found as a part of BIND v8).

BUG REPORTS AND SUPPORT
       If you think you have found a bug in Pth, you should send a report as
       complete as possible to bug-pth@gnu.org. If you can, please try to fix
       the problem and include a patch, made with '"diff -u3"', in your
       report. Always, at least, include a reasonable amount of description in
       your report to allow the author to deterministically reproduce the bug.

       For further support you additionally can subscribe to the
       pth-users@gnu.org mailing list by sending an Email to
       pth-users-request@gnu.org with `"subscribe pth-users"' (or `"subscribe
       pth-users" address' if you want to subscribe from a particular Email
       address) in the body. Then you can discuss your issues with other Pth
       users by sending messages to pth-users@gnu.org. Currently (as of August
       2000) you can reach about 110 Pth users on this mailing list. Old
       postings you can find at
       http://www.mail-archive.com/pth-users@gnu.org/.

SEE ALSO
   Related Web Locations
       `comp.programming.threads Newsgroup Archive',
       http://www.deja.com/topics_if.xp?
       search=topic&group=comp.programming.threads

       `comp.programming.threads Frequently Asked Questions (F.A.Q.)',
       http://www.lambdacs.com/newsgroup/FAQ.html

       `Multithreading - Definitions and Guidelines', Numeric Quest Inc 1998;
       http://www.numeric-quest.com/lang/multi-frame.html

       `The Single UNIX Specification, Version 2 - Threads', The Open Group
       1997; http://www.opengroup.org/onlinepubs /007908799/xsh/threads.html

       SMI Thread Resources, Sun Microsystems Inc;
       http://www.sun.com/workshop/threads/

       Bibliography on threads and multithreading, Torsten Amundsen;
       http://liinwww.ira.uka.de/bibliography/Os/threads.html

   Related Books
       B. Nichols, D. Buttlar, J.P. Farrel: `Pthreads Programming - A POSIX
       Standard for Better Multiprocessing', O'Reilly 1996; ISBN 1-56592-115-1

       B. Lewis, D. J. Berg: `Multithreaded Programming with Pthreads', Sun
       Microsystems Press, Prentice Hall 1998; ISBN 0-13-680729-1

       B. Lewis, D. J. Berg: `Threads Primer - A Guide To Multithreaded
       Programming', Prentice Hall 1996; ISBN 0-13-443698-9

       S. J. Norton, M. D. Dipasquale: `Thread Time - The Multithreaded
       Programming Guide', Prentice Hall 1997; ISBN 0-13-190067-6

       D. R. Butenhof: `Programming with POSIX Threads', Addison Wesley 1997;
       ISBN 0-201-63392-2

   Related Manpages
       pth-config(1), pthread(3).

       getcontext(2), setcontext(2), makecontext(2), swapcontext(2),
       sigstack(2), sigaltstack(2), sigaction(2), sigemptyset(2),
       sigaddset(2), sigprocmask(2), sigsuspend(2), sigsetjmp(3),
       siglongjmp(3), setjmp(3), longjmp(3), select(2), gettimeofday(2).

AUTHOR
	Ralf S. Engelschall
	rse@engelschall.com
	www.engelschall.com

2.0.8				 pthsem 2.0.8			     .::pth(3)
[top]

List of man pages available for DragonFly

Copyright (c) for man pages and the logo by the respective OS vendor.

For those who want to learn more, the polarhome community provides shell access and support.

[legal] [privacy] [GNU] [policy] [cookies] [netiquette] [sponsors] [FAQ]
Tweet
Polarhome, production since 1999.
Member of Polarhome portal.
Based on Fawad Halim's script.
....................................................................
Vote for polarhome
Free Shell Accounts :: the biggest list on the net