Document number: N2043=06-0113

Ion Gaztañaga (igaztanaga at gmail dot com)
2006-06-18

Simplifying And Extending Mutex and Scoped Lock Types For C++ Multi-Threading Library

Contents

Introduction

Recent "A Multi-threading Library for Standard C++, Revision 1" (N1907) paper proposes mutex and scoped lock types based on Boost.Threads library. This paper is an effort to analyze the 6 mutex types and 3 scoped lock types N1907 proposes to suggest a simplification. The paper also proposes new mutex types to complete a C++ thread synchronization library.

Simplifying mutex types

N1907 divides mutex types depending on:

Do we need so many mutex types? The Boost.Threads implementation is probably conditioned by the fact that POSIX mutex operations had no timed operations. Boost.Threads' timed mutexes are implemented combining mutex and condition variables. Due to this limitation, timed_mutex implementation has a performance penalty comparing to mutex and try_mutex, since mutex locking and try locking are natively implemented in POSIX threads.

However, times are changing and POSIX offers optional timed functions (_POSIX_TIMEOUTS option) from IEEE Std 1003.1-2001 that are currently present in most popular operating systems. The Boost.Threads timed mutexes implementation is rather old (2001-2003) and we can suppose that for C++ TR2 and future C++ standard, timeout functions will be present in all systems. Note that the lack of timed functions was a UNIX issue, since Windows supports timed functions.

Based on this argumentation, this paper proposes unifying all Boost.Threads mutex types into 2 types, mutex and recursive_mutex, that would have the same capabilities as N1907's timed_mutex and recursive_timed_mutex.

It's possible that C++ could get more mutex types like process-shared mutexes (recursive and non-recursive) or read-write mutexes. Following current Boost.Threads' interface, 6 mutex types would become then 12 mutex types with their process-shared counterparts. This paper proposes reducing 6 mutex types to 2 mutex types and the addition of a null_mutex.

N1907's mutex types have no public locking interface. The entire locking interface is private, giving friendship access to some scoped lock types, so mutex locking is only possible using scoped locks. Many users find private locking functions too limiting and this paper proposes making locking functions public. Scoped locks will still be the preferred way to lock mutexes but the user should have the possibility to use mutexes directly to build advanced synchronization utilities based on mutexes, where scoped locking pattern is not used.

Common mutex operations

This paper proposes standardizing the syntax and effects of common mutex operations so that a user can implement his own mutex and use synchronization classes like scoped_lock. This is similar to STL containers: the standard implements some containers that follow an interface, and the user can write his own containers and use STL algorithms and utilities. This is the proposed interface for common operations:

void lock()

bool try_lock()

bool timed_lock(const absolute_time& abs_time)

void unlock()

Like Boost.Threads mutexes, we can specify 3 mutex models:

This paper proposes that standardized mutex types should be implementations of a TimedMutex model.

Should Mutex Operations Throw?

We can standardize also if proposed mutex operations throw on failure. The question is, can mutex operations fail? If we revise Boost.Threads implementation, we'll see that no locking/unlocking operation throws. If we revise the most popular mutex implementations we can see the following failure errors:

In Windows platforms, using CRITICAL_SECTION feature can't return any error:

In POSIX platforms int pthread_mutex_lock(pthread_mutex_t *mutex), int pthread_mutex_trylock(pthread_mutex_t *mutex) and int pthread_mutex_unlock(pthread_mutex_t *mutex) functions can return the following errors:

EBUSY and EPERM are logic errors that can be perfectly avoided using ScopedLocks, and even if the user obtains that error because of bad coding, what can the user do to solve it? The only way is to solve this error to recompile the application.

EINVAL error can't happen in C++ since all mutexes are initialized in the constructor, and there is no way to call locking functions from another object type, since locking functions are member functions. If this error occurs it can show a memory corruption problem, so the only safe bet is to kill the application.

EAGAIN can happen if recursive locking exceeds its limits. This is the only possible runtime error. However, it's possible to implement locking mutexes without locking limitation. Even if the underlying implementation returns this error, Boost.Threads asserts on this error in debug mode and ignores it release mode.

In conclusion, although we can obtain error return values for mutex operations, it's not clear if those are useful to force a throwing interface in the standatd C++ mutex interface. Non-throwing interface can simplify code and improve space-time in resource constricted environments like embedded systems. When ScopedLock pattern is used, unlocking is executed in the destructor, and since throwing exceptions in destructors is discouraged the unlocking exception will be ignored. This paper makes no formal proposal about throwing semantics of locking operations, but suggests that we need a deep analysis to know if mutex errors are possible and if they can be handled by the user.

Simplified Mutex Interface

This paper proposes the simplification of Boost.Threads types into two mutex types following the TimedMutex model:

class mutex
{
   //Non-copyable
   mutex(const mutex &m);

   //Non-assignable
   mutex &operator=(const mutex &m);

   public:

   //Constructs the mutex in unlocked state
   mutex();

   //Destroys the mutex
   ~mutex();

   //Obtains exclusive ownership of the mutex
   void lock();

   //Will attempt to obtain exclusive ownership of the mutex
   bool try_lock();

   //Will attempt to obtain exclusive ownership of the mutex if it can do
   //so within the specified time interval
   bool timed_lock(const absolute_time &abs_time);

   //Will release exclusive ownership of the mutex
   void unlock();
};
class recursive_mutex
{
   //Non-copyable
   recursive_mutex(const recursive_mutex &m);

   //Non-assignable
   recursive_mutex &operator=(const recursive_mutex &m);

   public:

   //Constructs the mutex in unlocked state
   recursive_mutex();

   //Destroys the mutex
   ~recursive_mutex();

   //Obtains exclusive ownership of the mutex
   void lock();

   //Will attempt to obtain exclusive ownership of the mutex
   bool try_lock();

   //Will attempt to obtain exclusive ownership of the mutex if it can do
   //so within the specified time interval
   bool timed_lock(const absolute_time &abs_time);

   //Will release exclusive ownership of the mutex
   void unlock();
};

Proposed mutex types are non-copyable and non-assignable. Since pthread_mutex_t type is not defined as copyable (unlike for example, file handles), this paper also suggests making mutex types non-movable.

Null Mutex

When programming generic code, many times we search the configurability of locking policies. When configuring a class to avoid locking, a common pattern is to develop a class that simulates a mutex (let's call it Null Mutex), but returns immediately simulating success in locking operations. These operations are commonly inline functions that can be optimized away by the compiler. This paper suggests including a Null Mutex class in C++ multithreading library, with the following operations:

class null_mutex
{
   //Non-copyable
   null_mutex(const recursive_mutex &m);

   //Non-assignable
   null_mutex &operator=(const recursive_mutex &m);

   public:

   //Constructor (empty)
   null_mutex()
   {}

   //Destructor (empty)
   ~null_mutex()
   {}

   //Obtains exclusive ownership of the mutex (empty)
   void lock()
   {}

   //Will attempt to obtain exclusive ownership of
   //the mutex (always succeeds)
   bool try_lock()
   {  return true;   }

   //Will attempt to obtain exclusive ownership of the mutex if it
   //can do so within the specified time interval. (always succeeds)
   bool timed_lock(const absolute_time &abs_time)
   {  return true;   }

   //Will release exclusive ownership of the mutex (empty)
   void unlock()
   {}
};

Simplifying scoped lock types

Like mutexes, N1907 offers 3 scoped_lock types. To know which scoped locks are suitable for a given mutex, each mutex defines via typedef the allowed scoped locks. In N1907 a mutex only grants locking capability (via friend declaration) to those scoped_locks. For example, Boost.Threads' timed_mutex defines 3 scoped lock types internally:

  typedef implementation-defined scoped_lock;      
  typedef implementation-defined scoped_try_lock;  
  typedef implementation-defined scoped_timed_lock;

This paper proposes an independent scoped_lock that can lock any object that implements a BlockingMutex, TryMutex or TimedMutex model. A single templatized locker can have all locking operations. A mutex does not need to implement all operations (for example, timed operations) to be used with the locker. If the user tries to use a non-implemented functionality a compile-time error will be generated saying that the mutex type has not the requested operation. The proposed locker would have overloaded constructors to initialize the locker without locking the mutex or executing a lock(), try_lock(), or timed_lock() operation.

Additionally, there are situations where we want to leave the mutex locked (implementing more complex synchronization classes, for example) but we want to reverse locking operations if an exception occurs. This is similar to a memory allocation: we use an auto_ptr object to avoid leaks if an exception is thrown after obtaining the memory, but we want to release() that auto_ptr when all steps have been successfully completed.

Scoped Lock interface

The interface of the proposed scoped_lock is presented here with a possible implementation The operations are easy to understand and scoped_lock can lock any object that implements a BlockingMutex, TryMutex or TimedMutex interface. scoped_lock can be perfectly used with a mutex that has no timed_lock or try_lock capabilities, as long as the functions requiring those functionalities are not called.

//Types to specify scoped_lock's constructors behaviour
static dont_lock_t   dont_lock;
static try_to_lock_t try_to_lock;

//The scoped lock class
template <class Mutex>
class scoped_lock
{
   private:
   typedef scoped_lock<Mutex> this_type;
   typedef bool this_type::*unspecified_bool_type;

   //Non-copyable
   scoped_lock(scoped_lock const&);
   //Non-assignable
   scoped_lock& operator=  (scoped_lock const&);

   public:
   typedef Mutex mutex_type;

   //Constructs the scoped_lock and locks the mutex using lock()
   explicit scoped_lock(mutex_type& m)
      : mp_mutex(&m), m_locked(false)
   {  mp_mutex->lock();   m_locked = true;  }

   //Constructs the scoped_lock but doesn't lock the mutex 
   scoped_lock(mutex_type& m, dont_lock_t)
      : mp_mutex(&m), m_locked(false)
   {}

   //Constructs the scoped_lock and tries to lock the mutex 
   //using try_lock()
   scoped_lock(mutex_type& m, try_to_lock_t)
      : mp_mutex(&m), m_locked(mp_mutex->try_lock())
   {}

   //Constructs the scoped_lock and tries to lock the mutex 
   //using timed_lock()
   scoped_lock(mutex_type& m, const absolute_time &abs_time)
      : mp_mutex(&m), m_locked(mp_mutex->timed_lock(abs_time))
   {}

   //Destroys the scoped_lock. If the mutex was not released and
   //the mutex is locked, unlocks it using unlock() function
   ~scoped_lock()
   {
      try{  if(m_locked &&amp; mp_mutex)   mp_mutex->unlock();  }
      catch(...){}
   }

   //Locks the mutex using. If it was locked throws an exception
   void lock()
   {  
      if(!mp_mutex || m_locked)
         throw lock_exception();
      mp_mutex->lock();
      m_locked = true;
   }

   //Try to lock the mutex. If it was locked throw an exception
   bool try_lock()
   {  
      if(!mp_mutex || m_locked)
         throw lock_exception();
      m_locked = mp_mutex->try_lock();
      return m_locked;
   }

   //Timed lock the mutex. If it was locked throw an exception
   bool timed_lock(const absolute_time &abs_time)
   {  
      if(!mp_mutex || m_locked)
         throw lock_exception();
      m_locked = mp_mutex->timed_lock(abs_time);
      return m_locked;
   }

   //Unlock the mutex. If it was already unlocked throw an exception
   void unlock()
   {
      if(!mp_mutex || !m_locked)
         throw lock_exception();
      mp_mutex->unlock();
      m_locked = false;
   }

   //Return true if the mutex is locked
   bool locked() const
   {  return m_locked;  }

   //If the mutex is locked return a true bool type, otherwise false
   operator unspecified_bool_type() const
   {  return m_locked? &this_type::m_locked : 0;   }

   //Returns a pointer to the contained mutex
   mutex_type* mutex() const
   {  return  mp_mutex;  }

   //Releases the contained mutex. The lock has lost ownership of
   //the mutex and will not try to unlock it in the destructor
   mutex_type* release()
   {
      mutex_type *mut = mp_mutex;
      mp_mutex = 0;
      return mut;   
   }
 
   private:
   mutex_type *mp_mutex; 
   bool        m_locked;
};

It's About Time

N1907 and Boost.Threads use a xtime class representing absolute time. xtime was supposed to be a temporary solution until a general time library was accepted in Boost. Recent Boost.Threads discussions have considered Boost.Date_Time as a better solution and this library is being proposed for TR2 (N1900) However, some might argue that locking operations should use an elapsed time instead of an absolute time.

Both approaches have strong and weak points. An elapsed time is more intuitive for the user, that does not need to obtain the current time, add some elapsed time and pass it to the timed operation. However, elapsed times are not appropriate for realtime systems, since the scheduler can put the thread on hold at any moment so it's difficult to obtain a predictable sleep time. We can have also spurious wakeups in condition variables, so we would need to recompute the elapsed time each wakeup, and we might not know how much time we have already waited. Windows uses elapsed times.

However, absolute times based on the realtime clock can have also problems if the date is changed. Time synchronization is common in real-time systems and changing the system's time while waiting can provoke many problems since POSIX states that relative time waits based on the realtime clock shouldn't modify their behaviour, whereas absolute-time based services should follow the new absolute time. POSIX mutexes use absolute times based only in the real-time clock.

So there is no portable solution nowadays. The ideal would be to have all possible approaches, having elapsed time and absolute time waiting functions and encapsulating the timer type in the time structure so that one can use relative times based on the realtime clock or absolute times based on the realtime clock or a monotonic system time that can't be changed. In general, it seems that absolute times are better to simplify the coding of timed complex functions that call several timed operations, since the user does not need to recompute the elapsed time after each timed operation. We should also see if a monitonic clock is better suited than the realtime clock.

Any portable solution will lead to changes in all operating systems (POSIX or Windows). This paper proposes absolute times as a better approach, but time management is an unresolved core issue in C++.

Both elapsed and absolute times will need high resolution time representation. Although microsecond representation might be enough for current applications, this paper, following the POSIX timespec structure proposes nanosecond resolution for standard C++ time representations.

Condition Variables And Mutex Types

Proposed Boost.Threads condition variables wait operations are defined taking a ScopedLock as argument. The wait operation is described like this:

template <class Lock>
    void wait(Lock& lock);

The first member function atomically unlocks lock and blocks until the condition object is signalled by a call to notify_one or to notify_all. When the calling thread becomes unblocked it locks the lock object lock before it returns.

Although the lock is a template argument, the truth is that only boost::mutex can be used as it's internal real mutex type. Boost.Threads condition variables don't work with boost::timed_mutex, so that the following non-blocking pattern can't be implemented in Boost:

int func(int expected_state, const xtime &timeout)
{
   //Try to lock until timeout reaches
   boost::timed_mutex::scoped_timed_lock lock(m_mutex, timeout);

   if (lock.locked() == true){
      while (m_expected_state != expected_state){

         //Try to wait until timeout reaches
         if (!m_condition.timed_wait(lock, timeout)){
            // Timed out on condition
            return TimedOut;
         }
      }
      // Change the progress state
   }
   else{
      // mutex timed out
      return TimedOut;
   }
   return Success;
}

This paper proposes that TryLock and TimedLock mutexes should be compatible with condition variables. Since this paper proposes unifying Mutex and ScopedLock types, and the proposed simplified mutex follows the TimedMutex model, we can implement this pattern replacing the boost timed scoped lock with:

std::scoped_lock<std::mutex> lock(m_mutex, timeout);

Embedded/Real Time Limitations

Embedded World/Realtime systems are increasingly using C++ and threads and mutexes are common operations in those systems. The problem is that in those systems exceptions are disabled for performance or predictability reasons. As "Technical Report on C++ Performance" states, exceptions can be unsuitable for these systems. Filesystem Proposal (N1934) offers overloaded functions taking an extra system_error_code& ec argument.

Basic synchronization utilities like mutexes and condition variables could also offer this overload with functions that can throw, to extend C++ synchronization classes to environments without exceptions.

Interprocess Synchronization

Along with multithreading synchronization classes, this paper also proposes adopting synchronization classes that can be shared between processes through shared memory or memory mapped files, to synchronize threads from different processes. These synchronization classes are essential to develop more complex interprocess communication channels, an important missing feature in standard C++. This paper proposes that C++ interprocess synchronization classes should mirror C++ thread synchronization classes. Since they implement the same interface, utilities like scoped_lock can be used both with thread and process synchronization classes. Lifetime requirements of shared memory and memory mapped files will derive in special construction and destruction requirements, as we will see in the next chapter.

interprocess_mutex and interprocess_recursive_mutex are the interprocess equivalents of mutex and recursive_mutex classes:

class interprocess_mutex
{
   //Non-copyable
   interprocess_mutex(const interprocess_mutex &m);

   //Non-assignable
   interprocess_mutex &operator=(const interprocess_mutex &m);

   public:

   //Constructs the interprocess_mutex in unlocked state
   interprocess_mutex();

   //Destroys the interprocess_mutex
   ~interprocess_mutex();

   //Obtains exclusive ownership of the interprocess_mutex
   void lock();

   //Will attempt to obtain exclusive ownership of the interprocess_mutex
   bool try_lock();

   //Will attempt to obtain exclusive ownership of the interprocess_mutex if it can do
   //so within the specified time interval
   bool timed_lock(const absolute_time &abs_time);

   //Will release exclusive ownership of the interprocess_mutex
   void unlock();
};
class interprocess_recursive_mutex
{
   //Non-copyable
   interprocess_recursive_mutex(const interprocess_recursive_mutex &m);

   //Non-assignable
   interprocess_recursive_mutex &operator=(const interprocess_recursive_mutex &m);

   public:

   //Constructs the interprocess_recursive_mutex in unlocked state
   interprocess_recursive_mutex();

   //Destroys the interprocess_recursive_mutex
   ~interprocess_recursive_mutex();

   //Obtains exclusive ownership of the interprocess_mutex
   void lock();

   //Will attempt to obtain exclusive ownership of the interprocess_recursive_mutex
   bool try_lock();

   //Will attempt to obtain exclusive ownership of the interprocess_recursive_mutex 
   //if it can do so within the specified time interval
   bool timed_lock(const absolute_time &abs_time);

   //Will release exclusive ownership of the interprocess_recursive_mutex
   void unlock();
};

interprocess_condition is the interprocess equivalent of condition and can only be used with locks locking interprocess_mutex mutex type:

class interprocess_condition
{
   //Non-copyable
   interprocess_condition(const interprocess_recursive_mutex &m);

   //Non-assignable
   interprocess_condition &operator=(const interprocess_recursive_mutex &m);

   public:

   //Constructs the interprocess_condition
   interprocess_condition();

   //Destroys the interprocess_condition
   ~interprocess_condition();

   //Atomically unlocks lock and blocks until the condition object is signalled by a 
   //call to notify_one or to notify_all. When the calling thread becomes unblocked 
   //it locks the lock object lock before it returns.
   template <class Lock>
       bool wait(Lock& lock);

   //The second member function atomically unlocks lock and blocks until the condition 
   //object is signalled by a call to notify_one or to notify_all and the predicate pred() 
   //returns true (that is, the function incorporates the loop that is needed to avoid 
   //spurious wakeups). When the calling thread becomes unblocked it locks the lock object 
   //lock before it returns.
   template <class Lock, class Pred>>
       bool wait(Lock& lock, Pred pred);

   //Atomically unlocks lock and blocks until the condition object is signalled by a call 
   //to notify_one or to notify_all, or until after the time specified by the time abs_time. 
   //When the calling thread becomes unblocked it locks the lock object lock before it returns. 
   //The function returns false if it returned because the time had passed, otherwise it returns true.
   template <class Lock>
       bool timed_wait(Lock& lock, const absolute_time& abs_time);

   //Atomically unlocks lock and blocks until the condition object is signalled by a call to 
   //notify_one or to notify_all and the predicate pred() returns true (that is, the 
   //function incorporates the loop that is needed to avoid spurious wakeups), or until after 
   //the time specified by abs_time. When the calling thread becomes unblocked it 
   //locks the lock object lock before it returns. The function returns false if it returned 
   //because the abs_time had passed, otherwise it returns true.
   template <class Lock, class Pred>>
       bool timed_wait(Lock& lock, const absolute_time& abs_time, Pred pred);

   //Unblocks one of the threads of a process that is blocked on the 
   //condition object at the time of the call. If no threads are blocked 
   //on the object at the time of the call the function does nothing.
   void notify_one();

   //Unblocks all of the threads of all processes that are blocked on the
   //condition object at the time of the call. If no threads are blocked
   //on the object at the time of the call the function does nothing.
   void notify_all();
};

Interprocess Synchronization With Memory Mapped Files

The lifetime requirements for synchronization objects to be constructed in shared memory or memory mapped files are harder than those for their thread counterparts. Specially, if we accept in standard C++ the possibility to construct synchronization objects in persistent memory-backends, like memory mapped files.

With thread synchronization objects, we can apply normal C++ constructor and destructor semantics. The constructor can allocate synchronization resources from the operating system and the destructor can release them. If a process crashes, the OS can also free them safely. So far, so good.

With shared memory, although it's more difficult, we can apply the same logic. The destruction semantics, however, are trickier. When a process crashes, the operating system can't free the resources, because other processes might be using the synchronization object. Process crashing happens frecuently in our systems, because of programming errors or due to explicit user "kill" requests. So interprocess mechanisms should be freed by the operating system when the memory where the interprocess synchronization object is placed is destroyed. This forces synchronization object designers to have (quasi) trivial destructors in process-shared synchronization objects and pass destruction guarantees to the OS support.

If we implement an interprocess_mutex that can leak resources when the shared memory is destroyed, we have quite a big problem there. Even an ordered destruction without process crashing would require quite a big job for the users: make sure that this process is the only process attached to the memory, destroy the synchronization resources, and destroy the shared memory, all atomically. It's clear that it's easier just to destroy the shared memory. So a trivial destructor (or quasi trivial, just with some debugging code, for example) is a basic need for robust interprocess synchronization objects.

Now let's analyze memory mapped files (and some shared memory implementations that might have filesystem persistence). Like shared memory, the OS should destroy the objects when the memory mapped file is unmapped by all the processes. But memory mapped files have filesystem persistence so let's imagine that we want to implement a persistent message queue using interprocess synchronization objects constructed in a fixed size file:

This is a complicated problem to solve. So as a consequence this paper proposes that when the file is remapped, all the synchronization objects should be ready to be used. Since the operating system does not know the contents of a file, and where the synchronization objects were constructed, this will require a handle-free implementation of synchronization objects.

A handle-free implementation means that instead of handling an opaque kernel handle, requiring initialization and cross-process global handles, operating system objects should be allocated on demand, when the first synchronization operation is requested. This will transform interprocess synchronization objects' constructors to a fairly simple POD-filling code (initializing atomic variables values, and setting constant data such marking the object as recursive or requesting a scheduling algorithm).

Although this might seem complicated, systems that implement process-shared mechanisms (for example POSIX system supporting the PTHREAD_PROCESS_SHARED atribute) have all the needed operating system functions to implement this. For example, the Linux kernel has a basic synchronization objects called futex that is essential to implement this filesystem persistence behaviour.

To sum up, this paper proposes the following lifetime requirements for interprocess synchronization objects:

These two requirements are not hard to achieve, since systems implementing POSIX process-shared synchronization mechanism implement this behaviour. To emulate this behaviour in systems without process-shared synchronization objects, atomic operations and spin-locks can be used, until those operating systems implement the needed operating system resources.

Do We Need Semaphore Types?

The original Boost.Threads library had a semaphore class but it was removed because it was considered error prone. As the FAQ of the library states:

11. Why has class semaphore disappeared?

Semaphore was removed as too error prone. The same effect can be achieved with greater safety by the combination of a mutex and a condition variable.

Certainly, the semaphore interface can be emulated with a condition and a mutex, but this emulation can't be used for important system programming tasks, like device driver implementations, signal handling or interrupt handlers.

Condition variables can't be used in interrupt handlers because interrupt handlers should never block and thus, they shouldn't acquire locks. The post() semaphore operation turns out to be the only synchronization operation that can be safely used in an interrupt handler. Shared variables protected by locks cannot be safely accessed in interrupt handlers. Typically, a dedicated device thread gets woken up by the interrupt handler:

semaphore sem(/*count*/0);
mutex     mut;

void interrupt_handler()
{
   //...
   sem.post();
   return from interrupt handler
}

void dedicated_device_thread
{
   while (1) {
      sem.wait();
      scoped_lock(mut);
      //Deal with interrupt...
   }
}

Event if a semaphore can be simulated with a mutex/condition combination, a simulated post() operation can block and a semaphore is much more efficient than a condition variable, since it requires far less operations and it's usually natively supported by the operating systems. For several real-time and time critical patterns a semaphore offers the needed synchronization with the best performance.

This paper acknowledges that semaphores should be only used in situations where mutex and condition variable combination can't offer enough performance or the function notifying an event can't be blocked. Because of this, this paper proposes standardizing a semaphore interface that can be optionally implemented by vendors (this is likely to happen in embedded C++ implementations). All embedded and realtime operating systems offer semaphore operations, so this shows the importance of semaphore classes for programming non-blocking critical tasks. The proposed interface is:

void post()

void wait()

bool try_wait()

bool timed_wait(const absolute_time &abs_time)

The standard semaphore class would have the following interface:

class semaphore
{
   semaphore(const semaphore &);
   semaphore &operator =(const semaphore &);

   public:

   semaphore(int initial_count);

   ~semaphore();

   void post();

   void wait();

   bool try_wait();

   bool timed_wait(const absolute_time &abs_time);
};

The semaphore class can also have its interprocess_semaphore counterpart.

Prior Art

The interprocess synchronization classes have been developed under the Shmem library and they are being improved for the Boost.Interprocess library. Process-shared mutex, condition variables and semaphores are native in POSIX operating systems supporting process-shared synchronization objects and spin-locks are used in Windows operating systems. These basic interprocess synchronization classes can be used to implement more complex interprocess communication channels, like message queues.

Acknowledgments

Thanks to Howard Hinnant for his explanations, support and improvements. Thanks to Boost mailing list members participating in Shmem/Interprocess review process, where mutex and lock simplification was suggested.

References