CSC/ECE 506 Spring 2010/ch 3 yl: Difference between revisions

From Expertiza_Wiki
Jump to navigation Jump to search
 
(246 intermediate revisions by 3 users not shown)
Line 1: Line 1:
Supplement to Chapter 3: Support for parallel-programming models. Discuss how DOACROSS, DOPIPE, DOALL, etc. are implemented in packages such as Posix threads, Intel Thread Building Blocks, OpenMP 2.0 and 3.0.
Supplement to Chapter 3: Support for parallel-programming models. Discuss how DOACROSS, DOPIPE, DOALL, etc. are implemented in packages such as Posix threads, Intel Thread Building Blocks, OpenMP 2.0 and 3.0.


==Parallel-programming models==
==Overview==


===Loop-independent vs. loop-carried dependences===
In this wiki supplement, we will discuss how the three kinds of parallelisms, i.e. DOALL, DOACROSS and DOPIPE implemented in the threads packages - OpenMP, Intel Threading Building Block, POSIX Threads. We discuss the each packages from the respects of variable scopes & Reduction/DOALL/DOACROSS/DOPIPE implementations.
Before performing the three kinds of parallelism analysis, we need to discuss about loop-dependence analysis first.


====Statement dependences====
==Implementation==
To discuss code analysis, let's define the framework as follows.  
 
Let S denote a statement in the source code.  
===OpenMP===
Let [] denote loop iteration space.  
The OpenMP Application Program Interface (API) supports multi-platform shared-memory parallel programming in C/C++ and Fortran on all architectures, including Unix platforms and Windows NT platforms. Jointly defined by a group of major computer hardware and software vendors, OpenMP is a portable, scalable model that gives shared-memory parallel programmers a simple and flexible interface for developing parallel applications for platforms ranging from the desktop to the supercomputer.
Let S1 -> S2 denote statement S1 executes before S2.  
 
Then, we further define three statement dependences:
====Variable Clauses ====
*S1 ->T S2 denotes true dependence: S1 -> S2, and S1 writes to a location that is read by S2.  
There are many different types of clauses in OpenMP and each of them has various characteristics. Here we introduce data sharing attribute clauses, Synchronization clauses, Scheduling clauses, Initialization and Reduction.
*S1 ->A S2 denotes anti dependence: S1 -> S2, and S1 reads a location written by S2.
=====Data sharing attribute clauses=====
*S1 ->O S2 denotes output dependence: S1 -> S2, and S1 writes to the same location written by S2.
* ''shared'': the data within a parallel region is shared, which means visible and accessible by all threads simultaneously. By default, all variables in the work sharing region are shared except the loop iteration counter.
To give a better understanding, we illustrate the dependences using the code section below:
  Format: shared ''(list)''
  '''for''' (i=1; i<N; i++)
 
    '''S1:''' a[i] = i;
  SHARED variables behave as follows:
    '''S2:''' b[i] = a[i-1];
  1. Existing in only one memory location and all threads can read or write to that address
    '''S3:''' b[i] = a[i]+c[i+1];
 
    '''S4:''' c[i] = 2*i;
* ''private'': the data within a parallel region is private to each thread, which means each thread will have a local copy and use it as a temporary variable. A private variable is not initialized and the value is not maintained for use outside the parallel region. By default, the loop iteration counters in the OpenMP loop constructs are private.
The dependences corresponding to the code are:  
  Format: private ''(list)''
*S1[i-1] ->T S2[i]: a[i-1] is written in statement S1 in the (i-1)th iteration, and read in S2 in the ith iteration.
 
*S1[i] ->T S3[i]: a[i] is written in statement S1 in the ith iteration, and read in S3 in the ith iteration.
  PRIVATE variables behave as follows:
*S3[i] ->A S4[i+1]: c[i] is read in statement S3 in the ith iteration, and written in S4 in the (i+1)th iteration.
    1. A new object of the same type is declared once for each thread in the team
*S2[i] ->O S3[i]: b[i] is written in statement S2 in the ithe iteration, and also written in S3 in the ith itertation.
    2. All references to the original object are replaced with references to the new object
    3. Variables declared PRIVATE should be assumed to be uninitialized for each thread
 
* ''default'': allows the programmer to state that the default data scoping within a parallel region will be either ''shared'', or ''none'' for C/C++, or ''shared'', ''firstprivate'', ''private'', or ''none'' for Fortran.  The ''none'' option forces the programmer to declare each variable in the parallel region using the data sharing attribute clauses.
  Format: default (shared | none)
 
  DEFAULT variables behave as follows:
    1. Specific variables can be exempted from the default using the PRIVATE, SHARED, FIRSTPRIVATE, LASTPRIVATE, and REDUCTION clauses.
    2. Using NONE as a default requires that the programmer explicitly scope all variables.
 
=====Synchronization clauses=====
* ''critical section'': the enclosed code block will be executed by only one thread at a time, and not simultaneously executed by multiple threads. It is often used to protect shared data from race conditions.
  Format: #pragma omp critical ''[ name ] newline''
          ''structured_block''
 
  CRITICAL SECTION behaves as follows:
    1.f a thread is currently executing inside a CRITICAL region and another thread reaches that CRITICAL region and attempts to execute it, it will block until the first thread exits that CRITICAL region.
    2. It is illegal to branch into or out of a CRITICAL block.
 
* ''atomic'': similar to ''critical section'', but advise the compiler to use special hardware instructions for better performance. Compilers may choose to ignore this suggestion from users and use ''critical section'' instead.
  Format: #pragma omp atomic  ''newline''
          ''statement_expression''
 
  ATOMIC behaves as follows:
    1. Only to a single, immediately following statement.
    2. An atomic statement must follow a specific syntax.
 
* ''ordered'': the structured block is executed in the order in which iterations would be executed in a sequential loop
  Format: #pragma omp for ordered ''[clauses...]''
          ''(loop region)''
          #pragma omp ordered  ''newline''
          ''structured_block
          (endo of loop region)''
 
  ORDERED behaves as follows:
    1. only appear in the dynamic extent of ''for'' or ''parallel for (C/C++)''.
    2. Only one thread is allowed in an ordered section at any time.
    3. It is illegal to branch into or out of an ORDERED block.
    4. A loop which contains an ORDERED directive, must be a loop with an ORDERED clause.
 
* ''barrier'': each thread waits until all of the other threads of a team have reached this point. A work-sharing construct has an implicit barrier synchronization at the end.
  Format: #pragma omp barrier  ''newline''
 
  BARRIER behaves as follows:
    1. All threads in a team (or none) must execute the BARRIER region.
    2. The sequence of work-sharing regions and barrier regions encountered must be the same for every thread in a team.
 
*''taskwait'': specifies that threads completing assigned work can proceed without waiting for all threads in the team to finish. In the absence of this clause, threads encounter a barrier synchronization at the end of the work sharing construct.
  Format: #pragma omp taskwait  ''newline''
 
  TASKWAIT behaves as follows:
    1. Placed only at a point where a base language statement is allowed.
    2. Not be used in place of the statement following an if, while, do, switch, or label.
 
*''flush'': The FLUSH directive identifies a synchronization point at which the implementation must provide a consistent view of memory. Thread-visible variables are written back to memory at this point.
  Format: #pragma omp flush ''(list)  newline''
 
  FLUSH behaves as follows:
    1. The optional list contains a list of named variables that will be flushed in order to avoid flushing all variables.
    2. Implementations must ensure any prior modifications to thread-visible variables are visible to all threads after this point.
 
=====Scheduling clauses=====
*''schedule(type, chunk)'': This is useful if the work sharing construct is a do-loop or for-loop. The iteration(s) in the work sharing construct are assigned to threads according to the scheduling method defined by this clause. The three types of scheduling are:
#''static'': Here, all the threads are allocated iterations before they execute the loop iterations. The iterations are divided among threads equally by default. However, specifying an integer for the parameter "chunk" will allocate "chunk" number of contiguous iterations to a particular thread.
#''dynamic'': Here, some of the iterations are allocated to a smaller number of threads. Once a particular thread finishes its allocated iteration, it returns to get another one from the iterations that are left. The parameter "chunk" defines the number of contiguous iterations that are allocated to a thread at a time.
#''guided'': A large chunk of contiguous iterations are allocated to each thread dynamically (as above). The chunk size decreases exponentially with each successive allocation to a minimum size specified in the parameter "chunk"
=====Initialization=====
* ''firstprivate'': the data is private to each thread, but initialized using the value of the variable using the same name from the master thread.
  Format: firstprivate ''(list)''
 
  FIRSTPRIVATE variables behave as follows:
    1. Listed variables are initialized according to the value of their original objects prior to entry into the parallel or work-sharing construct.
 
* ''lastprivate'': the data is private to each thread. The value of this private data will be copied to a global variable using the same name outside the parallel region if current iteration is the last iteration in the parallelized loop.  A variable can be both ''firstprivate'' and ''lastprivate''.
  Format: lastprivate ''(list)''
 
* ''threadprivate'': The data is a global data, but it is private in each parallel region during the runtime. The difference between ''threadprivate'' and ''private'' is the global scope associated with threadprivate and the preserved value across parallel regions.
  Format: #pragma omp threadprivate ''(list)''
 
  THREADPRIVATE variables behave as follows:  
    1. On first entry to a parallel region, data in THREADPRIVATE variables and common blocks should be assumed undefined.
    2. The THREADPRIVATE directive must appear after every declaration of a thread private variable/common block.
 
=====Reduction=====
* ''reduction'': the variable has a local copy in each thread, but the values of the local copies will be summarized (reduced) into a global shared variable. This is very useful if a particular operation (specified in "operator" for this particular clause) on a datatype that runs iteratively so that its value at a particular iteration depends on its value at a previous iteration. Basically, the steps that lead up to the operational increment are parallelized, but the threads gather up and wait before updating the datatype, then increments the datatype in order so as to avoid racing condition.
  Format: reduction ''(operator: list)''
 
  REDUTION variables behave as follows:
    1. Variables in the list must be named scalar variables. They can not be array or structure type variables. They must also be declared SHARED in the enclosing context.
    2. Reduction operations may not be associative for real numbers.
 
====DOALL====
 
In code 3.20, first it must include the header file ''omp.h'' which contains OpenMP function clarations. Next, A paralel region is started by #pragma omp parallel and we enclose this program bu curly brackets. We can use (setenv OMP_NUM_THREADS n) to specify the number of threads. Another way to determine the number of threads is directly calling a function (omp_set_numtheads (n)).  
Code 3.20 only has one loop to execute and we want to execute in parallel, so we combine the start of the parallel loop and the start of the parallel region with one directive ''#pragma omp parallel for''.
'''Code 3.20 A DOALL parallelism example in OpenMP
'''#include''' <omp.h>
'''...'''
'''#pragma''' omp parallel //start of parallel region
  '''{'''
  '''...'''
  '''#pragma''' omp parallel for default (shared)
  '''for''' ( i = 0; i < n ; i++)
    '''A[i]''' = A[i] + A[i] - 3.0;
'''}'''//end for parallel region
 
Apparently, there is no loop-carried dependence in ''i'' loop. With OpenMP, we only need to insert the ''pragma'' directive ''parallel for''. The ''dafault(shared)'' clauses states that all variables within the scope of the loop are shared  unless otherwise specified.
 
====DOACROSS====
We will introduce how to implement DOACROSS in OpenMP. Here is an example code which has not been paralleled yet.
'''Sample Code'''
01: for(i=1; i< N; i++) {
02: for(j=1; j<N; j++){
03: a[i][j]=a[i-1][j]+a[i][j-1];
04: }
05: }
 
From this sample code, obviously, there is dependence existing here.
a[i,j] -> T a[i+1, j+1]
 
In OpenMP, DOALL parallel can be implemented by insert a “#pragma omp for” before the “for” structure in the source code. But there is not a pragma corresponding to DOACROSS parallel.
 
When we implement DOACROSS, we use a shared array "_mylocks[threadid]" which is defined to store events of each thread. Besides, a private variable _counter0 is defined to indicate the event which current thread is waiting for. "mylock" indicates the total number of threads.
The number of threads is got by function "omp_get_num_threads()" and current thread's id is got by function "omp_get_thread_num()".
 
*omp_get_num_threads(): Returns the number of threads that are currently in the team executing the parallel region from which it is called.
Format: #include <omp.h>
        int omp_get_num_threads(void)
 
OMP_GET_NUM_THREADS behaves as following:
  1. If this call is made from a serial portion of the program, or a nested parallel region that is serialized, it will return 1.
  2. The default number of threads is implementation dependent.
 
*omp_get_thread_num(): Returns the thread number of the thread, within the team, making this call. This number will be between 0 and OMP_GET_NUM_THREADS-1. The master thread of the team is thread 0
Format: #include <omp.h>
        int omp_get_thread_num(void)
 
OMP_GET_THREAD_NUM behaves as followings:
  1. If called from a nested parallel region, or a serial region, this function will return 0.
 
Now, let's see the code which has been paralleled and explanation.
 
01: int _mylocks[256]; //thread’s synchronized array
02: #pragma omp parallel
03: {
04: int _counter0 = 1;
05: int _my_id = omp_get_thread_num();
06: int _my_nprocs= omp_get_num_threads();
07: _mylocks[my_id] = 0;
08: for(j_tile = 0; j_tile<N-1; j_tile+=M){
09: if(_my_id>0) {
10: do{
11: #pragma omp flush(_mylock)
12: } while(_mylock[myid-1]<_counter0);
13: #pragma omp flush(a, _mylock)
14: _counter0 += 1;
15: }
16: #pragma omp for nowait
17: for(i=1; i< N; i++) {
18: for(j=j_tile;j<j_tile+M;j++){
19: a[i][j]=a[i-1][j]+a[i][j-1];
20: }
21: }
22: _mylock[myid] += 1;
23: #pragma omp flush(a, _mylock)
24: }
25: }
 
We paralleled the original program in two steps.
 
*First step: We divide i loop among the other four processors by inserting an OpenMP to construct “#programa omp for nowait” (line 16). Afterwards, each processor will take four interations of the loop i. The same to j loop. Assume the size of each block is 4. Each processor will execute four iterations of loop j. In order to let the total iterations be equal to the original program, j has to be enclosed in loop i. So, the new loop will be looked like ''for (j_tile = 2; j_tile <= 15; j_tile += 4)'', line 18.
The lower bound of loop j is set to j_tile and the upper bound will be j)tile+3. We will keep other statements remained.
 
*Second step: We are going to Synchronize the neighbor threads. After first step, the four processor will finish computing a block 4x4. If we parallel all these four processors, the dependence will be violated. So, we have to synchronized them by neighbors.
We set 4 variables as followings:
1. A private variable: _my_nprocs = omp_get_num_threads(), which indicates the total number of threads that run corresponding parallel region.
2. A private variable : _my_id = omp_get_thread_num(),which indicates  the unique ID for current thread.
3. A shared array:_mylocks[proc], is initialize by 0 for each element, which is used to indicate whether the thread of proc-1 has finish computing the current block.
4. A private variable :_counter0, is initialize by 1, which indicate the block that current thread is waiting for.
 
With the four variable, threads are synchronized:
The first thread continue to run with out waiting (line 9), because its thread ID is 0. Then all other thread can not go down after line 12. If the value in ''_mylocks[_my_id-1]'' is smaller than ''_counter0''.
 
Otherwise, the block that current thread is waiting for must have be completed, and current thread can go down line 12, and mark the next block it will wait for by add 1 to ''_counter0'' (line 14).
 
When current thread finish its block, it will set that it
has finish a block by ''mylocks[proc]++''. Once the neighbor thread finds the value has been changed, it will continue running and so on. The below figure presents it to us.
[[Image:Synchorization.jpg]]
 
====DOPIPE====
Here is another example code and we are going to parallel it in DOPIPE parallelism. There is a dependence, which is S2 -> T S1, existing in the sample code.
'''Sample Code'''
01: for(i=1; i< N; i++) {
02: S1: a[i]=b[i];
03: S2: c[i]=c[i-1]+a[i];
04:  
05: }
Now, let's see how to parallel the sample code to DOPIPE parallelism.
we still use a shared array "_mylocks[threadid]" which is defined to store events of each thread. Besides, a private variable _counter0 is defined to indicate the event which current thread is waiting for. "mylock" indicates the total number of threads.
The number of threads is got by function "omp_get_num_threads()" and current thread's id is got by function "omp_get_thread_num()".
 
01: int _mylocks[256]; //thread’s synchronized array
02: #pragma omp parallel
03: {
04: int _counter0 = 1;
05: int _my_id = omp_get_thread_num();
06: int _my_nprocs= omp_get_num_threads();
07: _mylocks[my_id] = 0;
08: for(i_tile = 0; i_tile<N-1; i_tile+=M){
09: if(_my_id>0) {
10: do{
11: #pragma omp flush(_mylock)
12: } while(_mylock[myid-1]<_counter0);
13: #pragma omp flush(a, _mylock)
14: _counter0 += 1;
15: }
16: #pragma omp for nowait
17: for(i=1; i< N; i++) {
18: a[i]=b[i];
19: }
20: for(i=1; i< N; i++) {
21: c[i]=c[i-1]+a[i];
22: }
23: _mylock[myid] += 1;
24: #pragma omp flush(a, _mylock)
25: }
26: }
 
Ideally, We parallelized the original program into two steps.
 
*First step: We divide i loop among the other processors by inserting an OpenMP to construct “#programa omp for nowait” (line 16). Afterwards, each processor will take interations of the loop i. Now, there are two loop i existing and each loop i contains different statements. Also, we will keep other statements remained.
 
*Second step: We are going to Synchronize the threads. After first step, processors will finish computing
a[i]=b[i]. If we parallel all the processors to do the second loop i, the dependence will be violated. So, we have to synchronized them by neighbors.
Still, we set 4 variables as followings:
1. A private variable: _my_nprocs = omp_get_num_threads(), which indicates the total number of threads that run corresponding parallel region.
2. A private variable : _my_id = omp_get_thread_num(),which indicates  the unique ID for current thread.
3. A shared array:_mylocks[proc], is initialize by 0 for each element, which is used to indicate whether the thread of proc-1 has finish computing the current block.
4. A private variable :_counter0, is initialize by 1, which indicate the block that current thread is waiting for.
 
When current thread finish its block, it will set that it has finish a block by ''mylocks[proc]++''. Once the processors finish their own block, the other processors will be able to get the value to use that value to execute in its statement and process that.
 
====Functional Parallelism====
 
In order to introduce function parallelism, we want to execute some code section in parallel with another code section. We use code 3.21 to show two loops execute in parallel with respect to one another, although each loop is sequentially executed.
 
'''Code''' 3.21 A function parallelism example in OpenMP
'''pragma''' omp parallel shared(A, B)private(i)
'''{'''
  '''#pragma''' omp sections nowait
  '''{'''
      '''pragma''' omp section
      '''for'''( i = 0; i < n ; i++)
        '''A[i]''' = A[i]*A[i] - 4.0;
      '''pragma''' omp section
      '''for'''( i = 0; i < n ; i++)
        '''B[i]''' = B[i]*B[i] - 9.0;
  '''}'''//end omp sections
'''}'''//end omp parallel
 
In code 3.21, there are two loops needed to be executed in parallel. We just need to insert two ''pragma omp section'' statements. Since we insert these two statements, those two loops will execute sequentially.
 
===Intel Thread Building Blocks===
 
Intel Threading Building Blocks (Intel TBB) is a library that supports scalable
parallel programming using standard ISO C++ code. It does not require special
languages or compilers. It is designed to promote scalable data parallel programming.
The library consists of data structures and algorithms that allow a programmer to avoid some complications arising from the use of native threading packages such as POSIX threads, Windows threads, or the portable Boost Threads in which individual threads of execution are created, synchronized, and terminated manually. Instead the library abstracts access to the multiple processors by allowing the operations to be treated as "tasks," which are allocated to individual cores dynamically by the library's run-time engine, and by automating efficient use of the cache. This approach groups TBB in a family of solutions for parallel programming aiming to decouple the programming from the particulars of the underlying machine. Also, Intel Threading Building Blocks provides net results, which enables you to specify
parallelism more conveniently than using raw threads, and at the same time can
improve performance.
 
====Variables Scope====
 
Intel TBB is a collection of components for parallel programming, here is the overview of the library contents:
 
* Basic algorithms: parallel_for, parallel_reduce, parallel_scan
* Advanced algorithms: parallel_while, parallel_do,pipeline, parallel_sort
* Containers: concurrent_queue, concurrent_vector, concurrent_hash_map
* Scalable memory allocation: scalable_malloc, scalable_free, scalable_realloc, scalable_calloc, scalable_allocator, cache_aligned_allocator
* Mutual exclusion: mutex, spin_mutex, queuing_mutex, spin_rw_mutex, queuing_rw_mutex, recursive mutex
* Atomic operations: fetch_and_add, fetch_and_increment, fetch_and_decrement, compare_and_swap, fetch_and_store
* Timing: portable fine grained global time stamp
* Task Scheduler: direct access to control the creation and activation of tasks
 
Then we will focus on some specific TBB variables.
 
=====parallel_for=====


====Loop-independent dependences====
Parallel_for is the template function that performs parallel iteration over a range of values. In Intel TBB, a lot of DOALL cases could be implemented by using this function. The syntax is as follows:
template<typename Index, typename Function>
Function parallel_for(Index first, Index_type last, Index step, Function f);
template<typename Range, typename Body>
void parallel_for( const Range& range, const Body& body, [, partitioner] );


Loop-independent dependence is defined as a dependence that exists between statements within a loop iteration.
A parallel_for(first, last, step, f) represents parallel execution of the loop: "for( auto i=first; i<last; i+=step ) f(i);".


====Loop-carried dependences====
=====parallel_reduce=====


Loop-carried dependence is defined as a dependence that exists between a statement in one iteration with another statement in a different iteration.
Function parallel_reduce computes reduction over a range. Syntax is as follows:
template<typename Range, typename Value, typename Func, typename Reduction>
Value parallel_reduce( const Range& range, const Value& identity, const Func& func, const Reduction& reduction );


===DOALL===
The functional form parallel_reduce(range,identity,func,reduction) performs a
'''for'''    i:=2:N-1 '''do''' A(i):=[A(i-1) + A(i) + A(i+1)]/3; '''next''' i;
parallel reduction by applying func to subranges in range and reducing the results
'''forall''' i:=2:N-1 '''do''' A(i):=[A(i-1) + A(i) + A(i+1)]/3;
using binary operator reduction. It returns the result of the reduction. Parameter func
and reduction can be lambda expressions.


'''for''' (i=2; i<=n; i+=2)
=====parallel_scan=====
  '''s''': a[i] = a[i-2];
'''for''' (i=3; i<=n; i+=2)
  '''s''': a[i] = a[i-2];


===DOACROSS===
This template function computes parallel prefix. Syntax is as follows:
In code 3.7, apparently there is a loop-carried dependence existing in the code.
template<typename Range, typename Body>
void parallel_scan( const Range& range, Body& body );
template<typename Range, typename Body>
void parallel_scan( const Range& range, Body& body, const auto_partitioner& );
   
   
  '''S[i] -> T S[i+1]'''
  template<typename Range, typename Body>
void parallel_scan( const Range& range, Body& body, const simple_partitioner& );
 
A parallel_scan(range,body) computes a parallel prefix, also known as parallel
scan. This computation is an advanced concept in parallel computing that is
sometimes useful in scenarios that appear to have inherently serial dependences. A
further explanation will be given in the DOACROSS example.
 
=====pipeline=====
 
This class performs pipelined execution. Members as follows:
namespace tbb {
    class pipeline {
    public:
        pipeline();
        ~pipeline();
        void add_filter( filter& f );
        void run( size_t max_number_of_live_tokens );
        void clear();
  };
}
 
A pipeline represents pipelined application of a series of filters to a stream of items.
Each filter operates in a particular mode: parallel, serial in order, or serial out of order. With a parallel filter,
we could implement DOPIPE parallelism.
 
====Reduction====
 
The reduction in Intel TBB is implemented using parallel_reduce function. A parallel_reduce recursively splits the range into subranges and uses the splitting constructor to make one or more copies of the body for each thread. We use an example to illustrate this:
 
#include "tbb/parallel_reduce.h"
#include "tbb/blocked_range.h"
using namespace tbb;
struct Sum {
    float value;
    Sum() : value(0) {}
    Sum( Sum& s, split ) {value = 0;}
    void operator()( const blocked_range<float*>& r ) {
        float temp = value;
        for( float* a=r.begin(); a!=r.end(); ++a ) {
            temp += *a;
        }
        value = temp;
    }
    void join( Sum& rhs ) {value += rhs.value;}
};
float ParallelSum( float array[], size_t n ) {
    Sum total;
    parallel_reduce( blocked_range<float*>( array, array+n ), total );
    return total.value;
}
 
The above example sums the values in the array. The parallel_reduce will do the reduction within the range of (array, array+n), to split the working body, and then join them by the return value for each split.
 
====DOALL====
The implementation of DOALL parallelism in Intel TBB will involve Parallel_for function.
To better illustrate the usage, here we discuss a simple example. The following is the original code:
   
   
Obviously, there is no way to implement it as DOALL parallelism.  
void SerialApplyFoo( float a[], size_t n ) {
    for( size_t i=0; i<n; ++i )
        Foo(a[i]);
}
 
After using Intel TBB, it could be switched to the following:
 
#include "tbb/blocked_range.h"
#include "tbb/parallel_for.h"
   
   
  '''Code 3.7 A loop with loop-carried dependence'''
  class ApplyFoo {
  '''for''' (i=1; i<=N; i++) {
    float *const my_a;
  '''S:''' a[i] = a[i-1] + b[i] * c[i];}
  public:
    void operator( )( const blocked_range<size_t>& r ) const {
        float *a = my_a;
        for( size_t i=r.begin(); i!=r.end( ); ++i )
            Foo(a[i]);
    }
    ApplyFoo( float a[] ) :
        my_a(a)
    {}
};
   
   
If we split code 3.7 as two loops, then the fist loop can be implement in DOALL parallism.  
void ParallelApplyFoo( float a[], size_t n ) {
In code 3.8, first of all, we create a new array named temp[i]. Second, we put temp[i] in a loop which is individual and loop-independence.  
    parallel_for(blocked_range<size_t>(0,n,The_grain_size_You_Pick), ApplyFoo(a) );
However, this solution generated a disadvantage: high storage. Due to increasing the array size of temp by i++, the size of temp depends on the number of iterations instead of threads.  
}
If N ( # of iteration) is bigger, then the size of temp will be larger.  
 
The example is the simplest DOALL parallelism, similar as the one in textbook, and execution graph will be just similar as the one in DOALL section above. But with the help with this simple illustration, the TBB code just gives you a flavor of how it would be implemented in Intel Threading Building Blocks.
 
A little more to say, parallel_for takes an optional third argument to specify a partitioner, which I used "The_grain_size_You_Pick" to represent. If you want to manually divide the grain and assign the work to processors, you could specify that in the function. Or, you could use automatic grain provided TBB. The auto_partitioner provides an alternative that heuristically chooses the grain size so that you do not have to specify one. The heuristic attempts to limit overhead while still providing ample opportunities for load balancing. Then, the last three line of the TBB code above will be:
 
void ParallelApplyFoo( float a[], size_t n ) {
    parallel_for(blocked_range<size_t>(0,n), ApplyFoo(a), auto_partitioner( ) );
}
 
====DOACROSS====
 
We could find a good example in Intel TBB to implement a DOACROSS with the help of parallel_scan. As stated in the parallel_scan section, this function computes a parallel prefix, also known as parallel
scan. This computation is an advanced concept in parallel computing which
could be helpful in scenarios that appear to have inherently serial dependences, which could be loop-carried dependences.
 
Let's consider this scenario (which is actually the mathematical definition of parallel prefix): 
T temp = id⊕;
for( int i=1; i<=n; ++i ) {
    temp = temp ⊕ x[i];
    y[i] = temp;
}
 
When we implement this in TBB using parallel_scan, it becomes:
 
using namespace tbb;
class Body {
    T sum;
    T* const y;
    const T* const x;
public:
    Body( T y_[], const T x_[] ) : sum(id⊕), x(x_), y(y_) {}
    T get_sum() const {return sum;}
    template<typename Tag>
    void operator()( const blocked_range<int>& r, Tag ) {
        T temp = sum;
        for( int i=r.begin(); i<r.end(); ++i ) {
            temp = temp ⊕ x[i];
            if( Tag::is_final_scan() )
                y[i] = temp;
        }
        sum = temp;
    }
    Body( Body& b, split ) : x(b.x), y(b.y), sum(id⊕) {}
    void reverse_join( Body& a ) { sum = a.sum ⊕ sum;}
    void assign( Body& b ) {sum = b.sum;}
};
 
float DoParallelScan( T y[], const T x[], int n ) {
    Body body(y,x);
    parallel_scan( blocked_range<int>(0,n), body );
    return body.get_sum();
}
 
It is the second part (function DoParallelScan) that we have to focus on.  
 
Actually, this example is just the scenario mentioned above that could take advantages of parallel_scan. The "inherently serial dependences" is taken care of by the functionality of parallel_scan. By computing the prefix, the serial code could be implemented in parallel with just one function.
 
====DOPIPE====
 
Pipeline class is the Intel TBB that performs pipelined execution. A pipeline represents pipelined application of a series of filters to a stream of items. Each filter operates in a particular mode: parallel, serial in order, or serial out of order. So this class can be used to implement a DOPIPE parallelism.
 
Here is a comparatively complex example about pipeline implementation. Also, if we look carefully, this is an example with both DOPIPE and DOACROSS:
 
#include <iostream>
#include "tbb/pipeline.h"
#include "tbb/tbb_thread.h"
#include "tbb/task_scheduler_init.h"
using namespace tbb;
char InputString[] = "abcdefg\n";
class InputFilter: public filter {
    char* my_ptr;
public:
    void* operator()(void*) {
        if (*my_ptr)
            return my_ptr++;
        else
            return NULL;
    }
    InputFilter() :
        filter( serial_in_order ), my_ptr(InputString) {}
};
class OutputFilter: public thread_bound_filter {
public:
    void* operator()(void* item) {
        std::cout << *(char*)item;
        return NULL;
    }
    OutputFilter() : thread_bound_filter(serial_in_order) {}
};
void RunPipeline(pipeline* p) {
    p->run(8);
}
int main() {
    // Construct the pipeline
    InputFilter f;
    OutputFilter g;
    pipeline p;
    p.add_filter(f);
    p.add_filter(g);
    // Another thread initiates execution of the pipeline
    tbb_thread t(RunPipeline,&p);
    // Process the thread_bound_filter with the current thread.
    while (g.process_item()!=thread_bound_filter::end_of_stream)
        continue;
    // Wait for pipeline to finish on the other thread.
    t.join();
    return 0;
}
 
The example above shows a pipeline with two filters where the second filter is a thread_bound_filter serviced by the main thread. The main thread does the following after constructing the pipeline:
1. Start the pipeline on another thread.
2. Service the thread_bound_filter until it reaches end_of_stream.
3. Wait for the other thread to finish.
 
===POSIX Threads===


'''Code 3.8 A split version of the loop in Code 3.7'''
POSIX Threads, or Pthreads, is a POSIX standard for threads. The standard, POSIX.1c, Threads extensions (IEEE Std 1003.1c-1995), defines an API for creating and manipulating threads.
'''for''' (i=1; i<=N; i++) {  //this loop has DOALL parallelism
  '''S1:''' temp[i] = b[i] * c[i];} 
'''for''' (i=1; i<=N; i++) {
  '''S2''': a[i] = a[i-1] + temp[i];}'''
Let's see code 3.9, We still create a new array temp[i]. There are two differences between code 3.9 (DOACROSS) and code 3.8 (DOALL). First, the size of array is not increasing by the # of iterations but the # of threads, because temp[i] is not a shared array but a private scalar. Second, there is only one loop which contains S1 and S2.
In DOACROSS parallelism, we insert two primitives, ''wait(name)'' and ''post(name)''. The principle of DOACROSS parallelism is to insert ''point-to-point synchronization''. Notice the statement ''post(i)'' and and ''wait(i-1)''. ''post(i)'' is a signal produced by producer a[i]. Once a[i] has been produced, it will be post immediately. Meanwhile, the consumer is waiting i-1 by using ''wait(i-1)''. The reason makes consumer has to wait the previous a[i] is because S2 has to use a[i-1] to calculate to generate a[i]. Next, when S2 produce a[i], it will post it, post(i), to let the next consumer consume it. We will discuss how DOACROSS parallelism works in figure later.


====Variable Scopes====
Pthreads defines a set of C programming language types, functions and constants. It is implemented with a pthread.h header and a thread library.


'''Code 3.9 Exploiting DOACROSS parallelism in the loop from code 3.7'''
There are around 100 Pthreads procedures, all prefixed "pthread_". The subroutines which comprise the Pthreads API can be informally grouped into four major groups:
'''post(0);'''
'''for''' (i=1; i<=N; i++) {
  '''S1:''' temp[i] = b[i] * c[i];}
  '''wait(i-1);'''
  '''S2''': a[i] = a[i-1] + temp[i];'''
  '''post(i);}'''


In this figure, we will introduce how DOACROSS parallelism executes by using ''post'' and ''wait''.
* '''Thread management:''' Routines that work directly on threads - creating, detaching, joining, etc. They also include functions to set/query thread attributes (joinable, scheduling etc.) E.g.pthread_create(), pthread_join().
In task 1 (i=1), S2 post post(1) as it finishes its own statement. Next, in task 2 (i=2), when S1 finished its own statement, task 2 has to wait until task 1 finishes and post (1). The statement of S2 in task 2 is ''a[2] = a[2-1] + temp[2];'' and it means S2 can not produce value without a[1]. The same to task 3, S2 in task 3 is ''a[3] = a[3-1] + temp[3];'' and has to wait until task 2 finishes its task and post a[2].
* '''Mutexes:''' Routines that deal with synchronization, called a "mutex", which is an abbreviation for "mutual exclusion". Mutex functions provide for creating, destroying, locking and unlocking mutexes. These are supplemented by mutex attribute functions that set or modify attributes associated with mutexes. E.g. pthread_mutex_lock(); pthread_mutex_trylock(); pthread_mutex_unlock().
* '''Condition variables:''' Routines that address communications between threads that share a mutex. Based upon programmer specified conditions. This group includes functions to create, destroy, wait and signal based upon specified variable values. Functions to set/query condition variable attributes are also included. E.g. pthread_cond_signal(); pthread_cond_broadcast(); pthread_cond_wait(); pthread_cond_timedwait();pthread_cond_reltimedwait_np().
* '''Synchronization:''' Routines that manage read/write locks and barriers. E.g. pthread_rwlock_rdlock(); pthread_rwlock_tryrdlock(); pthread_rwlock_wrlock();pthread_rwlock_trywrlock(); pthread_rwlock_unlock();pthread_barrier_init(); pthread_barrier_wait()


[[Image:DOACROSS.jpg]]
====DOALL====


===DOPIPE===
The following is a simple code example in C, as DOALL parallelism, to print out each threads' ID#.


In code 3.12, there is a loop-carried dependence existing in the loop.
#include <pthread.h>
#include <stdio.h>
#include <stdlib.h>
#define NUM_THREADS    5
 
void *PrintHello(void *threadid)
{
    long tid;
 
    tid = (long)threadid;
    printf("Hello World! It's me, thread #%ld!\n", tid);
    pthread_exit(NULL);
}
 
int main (int argc, char *argv[])
{
    pthread_t threads[NUM_THREADS];
 
    int rc;
    long t;
    for(t=0; t<NUM_THREADS; t++){
      printf("In main: creating thread %ld\n", t);
      rc = pthread_create(&threads[t], NULL, PrintHello, (void *)t);
 
      if (rc){
          printf("ERROR; return code from pthread_create() is %d\n", rc);
          exit(-1);
      }
    }
    pthread_exit(NULL);
}


'''S1[i]''' -> T S1[i+1]
This loop contains only single statement which doesn't cross the iterations, so each iteration could be considered as a parallel task.
====DOACROSS====


When it comes to using Pthreads to implement DOACROSS, it could express functional parallelism easily, but make the parallelism unnecessarily complicated. See an example below: from '''POSIX Threads Programming''' by Blaise Barney
#include <pthread.h>
#include <stdio.h>
#include <stdlib.h>
#define NUM_THREADS
   
   
  '''Code 3.12 A loop amenable to pipelined parallelism'''
  void *BusyWork(void *t)
'''for''' (i=1; i<=N; i++) {
{
  '''S1:''' a[i] = a[i-1] + b[i];
  int i;
  '''S2:''' c[i] = c[i] + a[i];}
  long tid;
  double result=0.0;
  tid = (long)t;
  printf("Thread %ld starting...\n",tid);
  for (i=0; i<1000000; i++)
  {
      result = result + sin(i) * tan(i);
  }
  printf("Thread %ld done. Result = %e\n",tid, result);
  pthread_exit((void*) t);
}
   
   
Here we split code 3.12 in 2 loops and make it parallelized in DOALL parallelism. For the first loop, there is no way to use DOALL parallelism because of loop-carried dependence. Otherwise, for the second loop, it is available to use DOALL parallelism since there is no loop carried dependence but loop-independence. Howevr, DOALL is not the only solution to parallelize code 3.12 and it still has an disadvantage. If there are only a few processors to run this program, then the speedup will not be too much. We will introduce another solution, DOPIPE parallelism, in code 3.13.
int main (int argc, char *argv[])
{
  pthread_t thread[NUM_THREADS];
  pthread_attr_t attr;
  int rc;
  long t;
  void *status;
   
   
'''Code DOALL parallelism(changed from 3.12) '''
  /* Initialize and set thread detached attribute */
  '''for''' (i=1; i<=N; i++) {
  pthread_attr_init(&attr);
  '''S1:''' a[i] = a[i-1] + b[i];}
  pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_JOINABLE);
  '''for''' (i=1; i<=N; i++) { //this loop has DOALL parallelism
   
  '''S2''': c[i] = c[i] + a[i];}
  for(t=0; t<NUM_THREADS; t++) {
      printf("Main: creating thread %ld\n", t);
      rc = pthread_create(&thread[t], &attr, BusyWork, (void *)t);
      if (rc) {
        printf("ERROR; return code from pthread_create()
                is %d\n", rc);
        exit(-1);
        }
      }
   
  /* Free attribute and wait for the other threads */
  pthread_attr_destroy(&attr);
  for(t=0; t<NUM_THREADS; t++) {
      rc = pthread_join(thread[t], &status);
      if (rc) {
        printf("ERROR; return code from pthread_join()
                is %d\n", rc);
        exit(-1);
        }
      printf("Main: completed join with thread %ld having a status 
            of %ld\n",t,(long)status);
      }
printf("Main: program completed. Exiting.\n");
pthread_exit(NULL);
}
 
This example demonstrates how to "wait" for thread completions by using the Pthread join routine. Since some implementations of Pthreads may not create threads in a joinable state, the threads in this example are explicitly created in a joinable state so that they can be joined later.
 
====DOPIPE====
There is examples of using Posix Threads to implement DOPIPE parallelism, but unnecessarily complex. Due to the long length, we won't provide it here. If the reader is interested, it could be found in <li>[http://homepage.mac.com/dbutenhof/Threads/code/pipe.c Pthreads DOPIPE example]</li>
 
===Comparison among the three===
 
====A unified example====
 
We use a simple parallel example from [http://sourceforge.net Sourceforge.net] to show how it will be implemented in the three packages, namely, POSIX Threads, Intel TBB, OpenMP, to show some common and differences among them.
 
Following is the original code:
 
Grid1 *g = new Grid1(0, n+1);
Grid1IteratorSub it(1, n, g);
DistArray x(g), y(g);
...
float e = 0;
ForEach(int i, it,
    x(i) += ( y(i+1) + y(i-1) )*.5;
    e += sqr( y(i) ); )
...
 
Then we are going to show the implementations in different packages, and also make a brief summary of the three packages.
 
=====In POSIX Thread=====
 
POSIX Thread: Symmetric multi processing, e.g. SMP multi-processor computers, multi-core processors, virtual shared memory computer.
 
Data layout: A single global memory. Each thread reads global shared data and writes to a private fraction of global data.
 
A simplified translation of the example parallel-for loop is given below.


Global declaration:


Here, we use DOPIPE parallelism to solve code 3.12. We still split code 3.12 as twp loops and insert primitive synchronization, ''post'' and ''wait''.
#include <pthread.h>
In order to calculate c[i], S2 has to get the value of a[i]. However, there is no a[i] in the second loop. S2 has to wait until S1 finished processing a[i] and post it. Once S1 produces a[i], S2 consumes a[i] immediately and starts process its own statement. We will explain this principle in figure later.
float *x, *y;
float vec[8];
int nn, pp;


thread code:


  '''Code 3.13 DOPIPE parallel version of the loop from code 3.12'''
  void *sub1(void *arg) {
'''for''' (i=1; i<=N; i++) { //this loop has DOALL parallelism
    int p = (int)arg;
  '''S1:''' a[i] = a[i-1] + b[i];
    float e_local = 0;
  '''post(i);'''}
    for (int i=1+(nn*p)/pp; i<1+(nn*(p+1))/pp; ++i) {
'''for''' (i=1; i<=N; i++) {
      x[i] += ( y[i+1] + y[i-1] )*.5;
  '''wait(i);'''
      e_local += y[i] * y[i];
  '''S2''': c[i] = c[i] + a[i];}
    }
    vec[p] = e_local;
    return (void*) 0;
}


This figure present us how DOPIPE parallelism works. Each task processes different statement. For example, task1 processes S1 and task2 processes S2 in this figure. On the other hands, S1 is a producer to produce a[i]. Once S1 produces a[i], it will post a[i], ''post(i)'', to let S2 consume. S2 has to wait for a[i] by using ''wait(i)'', because S2 does not know the value of a[i] until S1 produces and posts it.
main code:


[[Image:DOPIPE.jpg]]
x = new float[n+1];
y = new float[n+1];
...
float e = 0;
int p_threads = 8;
nn = n-1;
pp = p_threads;
pthread_t threads[8];
pthread_attr_t attr;
pthread_attr_init(&attr);
for (int p=0; p<p_threads; ++p)
    pthread_create(&threads[p], &attr,
      sub1, (void *)p);
for (int p=0; p<p_threads; ++p) {
    pthread_join(threads[p], NULL);
    e += vec[p];
}
...
delete[] x, y;


==Implementation==
=====In Intel Threading Building Blocks=====
 
Intel TBB: A C++ library for thread programming, e.g. SMP multi-processor computers, multi-core processors, virtual shared memory computer.
 
Data layout: A single global memory. Each thread reads global shared data and writes to a private fraction of global data.
 
Translation of the example parallel-for loop is given below.
 
Global:
#include "tbb/task_scheduler_init.h"
#include "tbb/blocked_range.h"
#include "tbb/parallel_reduce.h"
#include "tbb/cache_aligned_allocator.h"
using namespace tbb;
 
thread code:
struct sub1 {
    float ee;
    float *x, *y;
    sub1(float *xx, float *yy) : ee(0), x(xx), y(yy) {}
    sub1(sub1& s, split) { ee = 0; x = s.x; y = s.y; }
    void operator() (const blocked_range<int> & r){
      float e = ee;
      for (int i = r.begin(); i!= r.end(); ++i) {
        x[i] += ( y[i+1] + y[i-1] )*.5;
        e += y[i] * y[i];
      }
      ee = e;
    }
    void join(sub1& s) { ee += s.ee; }
};
 
main code:
task_scheduler_init init;
...
float e;
float *x = cache_aligned_allocator<float>().allocate(n+1);
float *y = cache_aligned_allocator<float>().allocate(n+1);
...
sub1 s(x, y);
parallel_reduce(blocked_range<int>(1, n, 1000), s);
e = s.ee;
...
cache_aligned_allocator<float>().deallocate(x, n+1);
cache_aligned_allocator<float>().deallocate(y, n+1);
 
=====In OpenMP shared memory parallel code annotations=====
 
OpenMP: Usually automatic paralleization with a run-time system based on a thread library.
 
A simplified translation of the example parallel-for loop is given below.
 
Global:
float e;
 
main code:
float *x = new float[n+1];
float *y = new float[n+1];
...
e = 0;
#pragma omp for reduction(+:e)
for (int i=1; i<n; ++i) {
    x[i] += ( y[i+1] + y[i-1] )*.5;
    e += y[i] * y[i];
}
...
delete[] x, y;
 
====Summary: Difference among them====
 
*Pthreads works for all the parallelism and could express functional parallelism easily, but it needs to build specialized synchronization primitives and explicitly privatize variables, makes it more effort needed to switch a serial program in to parallel mode.
 
*OpenMP can provide many performance enhancing features, such as atomic, barrier and flush synchronization primitives. It is very simple to use OpenMP to exploit DOALL parallelism, but the syntax for expressing functional parallelism is awkward.
 
*Intel TBB relies on generic programming, it performs better with custom iteration spaces or complex reduction operations. Also, it provides generic parallel patterns for parallel while-loops, data-flow pipeline models, parallel sorts and prefixes, so it's better in cases go beyond loop-based parallelism.


==References==
==References==
<ol>
<ol>
<li>[http://en.wikipedia.org/wiki/Parallel_computing wikipedia: Parallel Computing]</li>
<li>[http://openmp.org/wp/about-openmp/ OpenMP.org]</li>
<li>[http://www.cesr.ncsu.edu/solihin/Main.html FUNDAMENTALS OF PARALLEL COMPUTER ARCHITECTURE, Yan Solihin, Aug 2009]</li>
<li>[https://docs.google.com/viewer?a=v&pid=gmail&attid=0.1&thid=126f8a391c11262c&mt=application%2Fpdf&url=https%3A%2F%2Fmail.google.com%2Fmail%2F%3Fui%3D2%26ik%3Dd38b56c94f%26view%3Datt%26th%3D126f8a391c11262c%26attid%3D0.1%26disp%3Dattd%26realattid%3Df_g602ojwk0%26zw&sig=AHIEtbTeQDhK98IswmnVSfrPBMfmPLH5Nw An Optimal Abtraction Model for Hardware Multithreading in Modern Processor Architectures]</li>
<li>[http://www.threadingbuildingblocks.org/uploads/81/91/Latest%20Open%20Source%20Documentation/Reference.pdf Intel Threading Building Blocks 2.2 for Open Source Reference Manual]</li>
<li>[http://www.csc.ncsu.edu/faculty/efg/506/s10/ NCSU CSC 506 Parallel Computing Systems]</li>
<li>[http://parallel-for.sourceforge.net/tbb.html Sourceforge.net]</li>
<li>[https://computing.llnl.gov/tutorials/openMP/ OpenMP]</li>
<li>[http://www.computer.org/portal/web/csdl/doi/10.1109/SNPD.2009.16 Barrier Optimization for OpenMP Program]</li>
<li>[http://cs.anu.edu.au/~Alistair.Rendell/sc02/module3.pdf Performance Programming: Theory, Practice and Case Studies]</li>
<li>[http://software.intel.com/en-us/articles/intel-threading-building-blocks-openmp-or-native-threads/ Intel® Threading Building Blocks, OpenMP, or native threads?]</li>
<li>[https://computing.llnl.gov/tutorials/pthreads/#Joining POSIX Threads Programming by Blaise Barney, Lawrence Livermore National Laboratory]</li>
<li>[http://homepage.mac.com/dbutenhof/Threads/source.html Programing with POSIX Threads source code]</li>
</ol>
</ol>

Latest revision as of 04:02, 4 May 2010

Supplement to Chapter 3: Support for parallel-programming models. Discuss how DOACROSS, DOPIPE, DOALL, etc. are implemented in packages such as Posix threads, Intel Thread Building Blocks, OpenMP 2.0 and 3.0.

Overview

In this wiki supplement, we will discuss how the three kinds of parallelisms, i.e. DOALL, DOACROSS and DOPIPE implemented in the threads packages - OpenMP, Intel Threading Building Block, POSIX Threads. We discuss the each packages from the respects of variable scopes & Reduction/DOALL/DOACROSS/DOPIPE implementations.

Implementation

OpenMP

The OpenMP Application Program Interface (API) supports multi-platform shared-memory parallel programming in C/C++ and Fortran on all architectures, including Unix platforms and Windows NT platforms. Jointly defined by a group of major computer hardware and software vendors, OpenMP is a portable, scalable model that gives shared-memory parallel programmers a simple and flexible interface for developing parallel applications for platforms ranging from the desktop to the supercomputer.

Variable Clauses

There are many different types of clauses in OpenMP and each of them has various characteristics. Here we introduce data sharing attribute clauses, Synchronization clauses, Scheduling clauses, Initialization and Reduction.

Data sharing attribute clauses
  • shared: the data within a parallel region is shared, which means visible and accessible by all threads simultaneously. By default, all variables in the work sharing region are shared except the loop iteration counter.
 Format: shared (list)
 SHARED variables behave as follows:
 1. Existing in only one memory location and all threads can read or write to that address 
  • private: the data within a parallel region is private to each thread, which means each thread will have a local copy and use it as a temporary variable. A private variable is not initialized and the value is not maintained for use outside the parallel region. By default, the loop iteration counters in the OpenMP loop constructs are private.
 Format: private (list)
 PRIVATE variables behave as follows: 
   1. A new object of the same type is declared once for each thread in the team
   2. All references to the original object are replaced with references to the new object
   3. Variables declared PRIVATE should be assumed to be uninitialized for each thread 
  • default: allows the programmer to state that the default data scoping within a parallel region will be either shared, or none for C/C++, or shared, firstprivate, private, or none for Fortran. The none option forces the programmer to declare each variable in the parallel region using the data sharing attribute clauses.
 Format: default (shared | none)
 DEFAULT variables behave as follows: 
   1. Specific variables can be exempted from the default using the PRIVATE, SHARED, FIRSTPRIVATE, LASTPRIVATE, and REDUCTION clauses. 
   2. Using NONE as a default requires that the programmer explicitly scope all variables.
Synchronization clauses
  • critical section: the enclosed code block will be executed by only one thread at a time, and not simultaneously executed by multiple threads. It is often used to protect shared data from race conditions.
 Format: #pragma omp critical [ name ]  newline
          structured_block
 CRITICAL SECTION behaves as follows:
   1.f a thread is currently executing inside a CRITICAL region and another thread reaches that CRITICAL region and attempts to execute it, it will block until the first thread exits that CRITICAL region.
   2. It is illegal to branch into or out of a CRITICAL block. 
  • atomic: similar to critical section, but advise the compiler to use special hardware instructions for better performance. Compilers may choose to ignore this suggestion from users and use critical section instead.
 Format: #pragma omp atomic  newline
          statement_expression
 ATOMIC behaves as follows:
   1. Only to a single, immediately following statement.
   2. An atomic statement must follow a specific syntax. 
  • ordered: the structured block is executed in the order in which iterations would be executed in a sequential loop
 Format: #pragma omp for ordered [clauses...]
         (loop region)
         #pragma omp ordered  newline
         structured_block
         (endo of loop region)
 ORDERED behaves as follows:
   1. only appear in the dynamic extent of for or parallel for (C/C++).
   2. Only one thread is allowed in an ordered section at any time.
   3. It is illegal to branch into or out of an ORDERED block. 
   4. A loop which contains an ORDERED directive, must be a loop with an ORDERED clause. 
  • barrier: each thread waits until all of the other threads of a team have reached this point. A work-sharing construct has an implicit barrier synchronization at the end.
  Format: #pragma omp barrier  newline
  BARRIER behaves as follows:
   1. All threads in a team (or none) must execute the BARRIER region.
   2. The sequence of work-sharing regions and barrier regions encountered must be the same for every thread in a team.
  • taskwait: specifies that threads completing assigned work can proceed without waiting for all threads in the team to finish. In the absence of this clause, threads encounter a barrier synchronization at the end of the work sharing construct.
  Format: #pragma omp taskwait  newline
  TASKWAIT behaves as follows:
   1. Placed only at a point where a base language statement is allowed.
   2. Not be used in place of the statement following an if, while, do, switch, or label.
  • flush: The FLUSH directive identifies a synchronization point at which the implementation must provide a consistent view of memory. Thread-visible variables are written back to memory at this point.
  Format: #pragma omp flush (list)  newline
  FLUSH behaves as follows:
   1. The optional list contains a list of named variables that will be flushed in order to avoid flushing all variables.
   2. Implementations must ensure any prior modifications to thread-visible variables are visible to all threads after this point.
Scheduling clauses
  • schedule(type, chunk): This is useful if the work sharing construct is a do-loop or for-loop. The iteration(s) in the work sharing construct are assigned to threads according to the scheduling method defined by this clause. The three types of scheduling are:
  1. static: Here, all the threads are allocated iterations before they execute the loop iterations. The iterations are divided among threads equally by default. However, specifying an integer for the parameter "chunk" will allocate "chunk" number of contiguous iterations to a particular thread.
  2. dynamic: Here, some of the iterations are allocated to a smaller number of threads. Once a particular thread finishes its allocated iteration, it returns to get another one from the iterations that are left. The parameter "chunk" defines the number of contiguous iterations that are allocated to a thread at a time.
  3. guided: A large chunk of contiguous iterations are allocated to each thread dynamically (as above). The chunk size decreases exponentially with each successive allocation to a minimum size specified in the parameter "chunk"
Initialization
  • firstprivate: the data is private to each thread, but initialized using the value of the variable using the same name from the master thread.
 Format: firstprivate (list)
 FIRSTPRIVATE variables behave as follows: 
   1. Listed variables are initialized according to the value of their original objects prior to entry into the parallel or work-sharing construct. 
  • lastprivate: the data is private to each thread. The value of this private data will be copied to a global variable using the same name outside the parallel region if current iteration is the last iteration in the parallelized loop. A variable can be both firstprivate and lastprivate.
 Format: lastprivate (list)
  • threadprivate: The data is a global data, but it is private in each parallel region during the runtime. The difference between threadprivate and private is the global scope associated with threadprivate and the preserved value across parallel regions.
 Format: #pragma omp threadprivate (list)
 THREADPRIVATE variables behave as follows: 
   1. On first entry to a parallel region, data in THREADPRIVATE variables and common blocks should be assumed undefined. 
   2. The THREADPRIVATE directive must appear after every declaration of a thread private variable/common block.
Reduction
  • reduction: the variable has a local copy in each thread, but the values of the local copies will be summarized (reduced) into a global shared variable. This is very useful if a particular operation (specified in "operator" for this particular clause) on a datatype that runs iteratively so that its value at a particular iteration depends on its value at a previous iteration. Basically, the steps that lead up to the operational increment are parallelized, but the threads gather up and wait before updating the datatype, then increments the datatype in order so as to avoid racing condition.
 Format: reduction (operator: list)
 REDUTION variables behave as follows: 
   1. Variables in the list must be named scalar variables. They can not be array or structure type variables. They must also be declared SHARED in the enclosing context.
   2. Reduction operations may not be associative for real numbers.

DOALL

In code 3.20, first it must include the header file omp.h which contains OpenMP function clarations. Next, A paralel region is started by #pragma omp parallel and we enclose this program bu curly brackets. We can use (setenv OMP_NUM_THREADS n) to specify the number of threads. Another way to determine the number of threads is directly calling a function (omp_set_numtheads (n)). Code 3.20 only has one loop to execute and we want to execute in parallel, so we combine the start of the parallel loop and the start of the parallel region with one directive #pragma omp parallel for.

Code 3.20 A DOALL parallelism example in OpenMP
#include <omp.h>
...
#pragma omp parallel //start of parallel region
{
 ...
 #pragma omp parallel for default (shared)
 for ( i = 0; i < n ; i++)
   A[i] = A[i] + A[i] - 3.0;
}//end for parallel region

Apparently, there is no loop-carried dependence in i loop. With OpenMP, we only need to insert the pragma directive parallel for. The dafault(shared) clauses states that all variables within the scope of the loop are shared unless otherwise specified.

DOACROSS

We will introduce how to implement DOACROSS in OpenMP. Here is an example code which has not been paralleled yet.

Sample Code
01: for(i=1; i< N; i++) {
02: for(j=1; j<N; j++){
03: a[i][j]=a[i-1][j]+a[i][j-1];
04: }
05: }

From this sample code, obviously, there is dependence existing here.

a[i,j] -> T a[i+1, j+1]

In OpenMP, DOALL parallel can be implemented by insert a “#pragma omp for” before the “for” structure in the source code. But there is not a pragma corresponding to DOACROSS parallel.

When we implement DOACROSS, we use a shared array "_mylocks[threadid]" which is defined to store events of each thread. Besides, a private variable _counter0 is defined to indicate the event which current thread is waiting for. "mylock" indicates the total number of threads. The number of threads is got by function "omp_get_num_threads()" and current thread's id is got by function "omp_get_thread_num()".

  • omp_get_num_threads(): Returns the number of threads that are currently in the team executing the parallel region from which it is called.
Format: #include <omp.h>
        int omp_get_num_threads(void)
OMP_GET_NUM_THREADS behaves as following:
 1. If this call is made from a serial portion of the program, or a nested parallel region that is serialized, it will return 1. 
 2. The default number of threads is implementation dependent. 
  • omp_get_thread_num(): Returns the thread number of the thread, within the team, making this call. This number will be between 0 and OMP_GET_NUM_THREADS-1. The master thread of the team is thread 0
Format: #include <omp.h>
        int omp_get_thread_num(void)
OMP_GET_THREAD_NUM behaves as followings:
 1. If called from a nested parallel region, or a serial region, this function will return 0. 

Now, let's see the code which has been paralleled and explanation.

01: int _mylocks[256]; //thread’s synchronized array
02: #pragma omp parallel
03: {
04: int _counter0 = 1;
05: int _my_id = omp_get_thread_num();
06: int _my_nprocs= omp_get_num_threads();
07: _mylocks[my_id] = 0;
08: for(j_tile = 0; j_tile<N-1; j_tile+=M){
09: if(_my_id>0) {
10: do{
11: #pragma omp flush(_mylock)
12: } while(_mylock[myid-1]<_counter0);
13: #pragma omp flush(a, _mylock)
14: _counter0 += 1;
15: }
16: #pragma omp for nowait
17: for(i=1; i< N; i++) {
18: for(j=j_tile;j<j_tile+M;j++){
19: a[i][j]=a[i-1][j]+a[i][j-1];
20: }
21: }
22: _mylock[myid] += 1;
23: #pragma omp flush(a, _mylock)
24: }
25: }

We paralleled the original program in two steps.

  • First step: We divide i loop among the other four processors by inserting an OpenMP to construct “#programa omp for nowait” (line 16). Afterwards, each processor will take four interations of the loop i. The same to j loop. Assume the size of each block is 4. Each processor will execute four iterations of loop j. In order to let the total iterations be equal to the original program, j has to be enclosed in loop i. So, the new loop will be looked like for (j_tile = 2; j_tile <= 15; j_tile += 4), line 18.

The lower bound of loop j is set to j_tile and the upper bound will be j)tile+3. We will keep other statements remained.

  • Second step: We are going to Synchronize the neighbor threads. After first step, the four processor will finish computing a block 4x4. If we parallel all these four processors, the dependence will be violated. So, we have to synchronized them by neighbors.

We set 4 variables as followings: 1. A private variable: _my_nprocs = omp_get_num_threads(), which indicates the total number of threads that run corresponding parallel region. 2. A private variable : _my_id = omp_get_thread_num(),which indicates the unique ID for current thread. 3. A shared array:_mylocks[proc], is initialize by 0 for each element, which is used to indicate whether the thread of proc-1 has finish computing the current block. 4. A private variable :_counter0, is initialize by 1, which indicate the block that current thread is waiting for.

With the four variable, threads are synchronized: The first thread continue to run with out waiting (line 9), because its thread ID is 0. Then all other thread can not go down after line 12. If the value in _mylocks[_my_id-1] is smaller than _counter0.

Otherwise, the block that current thread is waiting for must have be completed, and current thread can go down line 12, and mark the next block it will wait for by add 1 to _counter0 (line 14).

When current thread finish its block, it will set that it has finish a block by mylocks[proc]++. Once the neighbor thread finds the value has been changed, it will continue running and so on. The below figure presents it to us.

DOPIPE

Here is another example code and we are going to parallel it in DOPIPE parallelism. There is a dependence, which is S2 -> T S1, existing in the sample code.

Sample Code
01: for(i=1; i< N; i++) {
02: S1: a[i]=b[i];
03: S2: c[i]=c[i-1]+a[i];
04: 
05: }

Now, let's see how to parallel the sample code to DOPIPE parallelism. we still use a shared array "_mylocks[threadid]" which is defined to store events of each thread. Besides, a private variable _counter0 is defined to indicate the event which current thread is waiting for. "mylock" indicates the total number of threads. The number of threads is got by function "omp_get_num_threads()" and current thread's id is got by function "omp_get_thread_num()".

01: int _mylocks[256]; //thread’s synchronized array
02: #pragma omp parallel
03: {
04: int _counter0 = 1;
05: int _my_id = omp_get_thread_num();
06: int _my_nprocs= omp_get_num_threads();
07: _mylocks[my_id] = 0;
08: for(i_tile = 0; i_tile<N-1; i_tile+=M){
09: if(_my_id>0) {
10: do{
11: #pragma omp flush(_mylock)
12: } while(_mylock[myid-1]<_counter0);
13: #pragma omp flush(a, _mylock)
14: _counter0 += 1;
15: }
16: #pragma omp for nowait
17: for(i=1; i< N; i++) {
18: a[i]=b[i];
19: }
20: for(i=1; i< N; i++) {
21: c[i]=c[i-1]+a[i];
22: }
23: _mylock[myid] += 1;
24: #pragma omp flush(a, _mylock)
25: }
26: }

Ideally, We parallelized the original program into two steps.

  • First step: We divide i loop among the other processors by inserting an OpenMP to construct “#programa omp for nowait” (line 16). Afterwards, each processor will take interations of the loop i. Now, there are two loop i existing and each loop i contains different statements. Also, we will keep other statements remained.
  • Second step: We are going to Synchronize the threads. After first step, processors will finish computing

a[i]=b[i]. If we parallel all the processors to do the second loop i, the dependence will be violated. So, we have to synchronized them by neighbors. Still, we set 4 variables as followings: 1. A private variable: _my_nprocs = omp_get_num_threads(), which indicates the total number of threads that run corresponding parallel region. 2. A private variable : _my_id = omp_get_thread_num(),which indicates the unique ID for current thread. 3. A shared array:_mylocks[proc], is initialize by 0 for each element, which is used to indicate whether the thread of proc-1 has finish computing the current block. 4. A private variable :_counter0, is initialize by 1, which indicate the block that current thread is waiting for.

When current thread finish its block, it will set that it has finish a block by mylocks[proc]++. Once the processors finish their own block, the other processors will be able to get the value to use that value to execute in its statement and process that.

Functional Parallelism

In order to introduce function parallelism, we want to execute some code section in parallel with another code section. We use code 3.21 to show two loops execute in parallel with respect to one another, although each loop is sequentially executed.

Code 3.21 A function parallelism example in OpenMP
pragma omp parallel shared(A, B)private(i)
{
 #pragma omp sections nowait
 {
     pragma omp section
     for( i = 0; i < n ; i++)
        A[i] = A[i]*A[i] - 4.0;
     pragma omp section
     for( i = 0; i < n ; i++)
        B[i] = B[i]*B[i] - 9.0;
 }//end omp sections
}//end omp parallel

In code 3.21, there are two loops needed to be executed in parallel. We just need to insert two pragma omp section statements. Since we insert these two statements, those two loops will execute sequentially.

Intel Thread Building Blocks

Intel Threading Building Blocks (Intel TBB) is a library that supports scalable parallel programming using standard ISO C++ code. It does not require special languages or compilers. It is designed to promote scalable data parallel programming. The library consists of data structures and algorithms that allow a programmer to avoid some complications arising from the use of native threading packages such as POSIX threads, Windows threads, or the portable Boost Threads in which individual threads of execution are created, synchronized, and terminated manually. Instead the library abstracts access to the multiple processors by allowing the operations to be treated as "tasks," which are allocated to individual cores dynamically by the library's run-time engine, and by automating efficient use of the cache. This approach groups TBB in a family of solutions for parallel programming aiming to decouple the programming from the particulars of the underlying machine. Also, Intel Threading Building Blocks provides net results, which enables you to specify parallelism more conveniently than using raw threads, and at the same time can improve performance.

Variables Scope

Intel TBB is a collection of components for parallel programming, here is the overview of the library contents:

  • Basic algorithms: parallel_for, parallel_reduce, parallel_scan
  • Advanced algorithms: parallel_while, parallel_do,pipeline, parallel_sort
  • Containers: concurrent_queue, concurrent_vector, concurrent_hash_map
  • Scalable memory allocation: scalable_malloc, scalable_free, scalable_realloc, scalable_calloc, scalable_allocator, cache_aligned_allocator
  • Mutual exclusion: mutex, spin_mutex, queuing_mutex, spin_rw_mutex, queuing_rw_mutex, recursive mutex
  • Atomic operations: fetch_and_add, fetch_and_increment, fetch_and_decrement, compare_and_swap, fetch_and_store
  • Timing: portable fine grained global time stamp
  • Task Scheduler: direct access to control the creation and activation of tasks

Then we will focus on some specific TBB variables.

parallel_for

Parallel_for is the template function that performs parallel iteration over a range of values. In Intel TBB, a lot of DOALL cases could be implemented by using this function. The syntax is as follows:

template<typename Index, typename Function>
Function parallel_for(Index first, Index_type last, Index step, Function f);
template<typename Range, typename Body>
void parallel_for( const Range& range, const Body& body, [, partitioner] );

A parallel_for(first, last, step, f) represents parallel execution of the loop: "for( auto i=first; i<last; i+=step ) f(i);".

parallel_reduce

Function parallel_reduce computes reduction over a range. Syntax is as follows:

template<typename Range, typename Value, typename Func, typename Reduction>
Value parallel_reduce( const Range& range, const Value& identity, const Func& func, const Reduction& reduction );

The functional form parallel_reduce(range,identity,func,reduction) performs a parallel reduction by applying func to subranges in range and reducing the results using binary operator reduction. It returns the result of the reduction. Parameter func and reduction can be lambda expressions.

parallel_scan

This template function computes parallel prefix. Syntax is as follows:

template<typename Range, typename Body>
void parallel_scan( const Range& range, Body& body );

template<typename Range, typename Body>
void parallel_scan( const Range& range, Body& body, const auto_partitioner& );

template<typename Range, typename Body>
void parallel_scan( const Range& range, Body& body, const simple_partitioner& );

A parallel_scan(range,body) computes a parallel prefix, also known as parallel scan. This computation is an advanced concept in parallel computing that is sometimes useful in scenarios that appear to have inherently serial dependences. A further explanation will be given in the DOACROSS example.

pipeline

This class performs pipelined execution. Members as follows:

namespace tbb {
    class pipeline {
    public:
       pipeline();
       ~pipeline(); 
       void add_filter( filter& f );
       void run( size_t max_number_of_live_tokens );
       void clear();
  };
}

A pipeline represents pipelined application of a series of filters to a stream of items. Each filter operates in a particular mode: parallel, serial in order, or serial out of order. With a parallel filter, we could implement DOPIPE parallelism.

Reduction

The reduction in Intel TBB is implemented using parallel_reduce function. A parallel_reduce recursively splits the range into subranges and uses the splitting constructor to make one or more copies of the body for each thread. We use an example to illustrate this:

#include "tbb/parallel_reduce.h"
#include "tbb/blocked_range.h"
using namespace tbb;
struct Sum {
    float value;
    Sum() : value(0) {}
    Sum( Sum& s, split ) {value = 0;}
    void operator()( const blocked_range<float*>& r ) {
        float temp = value;
        for( float* a=r.begin(); a!=r.end(); ++a ) {
            temp += *a;
        }
        value = temp;
    }
    void join( Sum& rhs ) {value += rhs.value;}
};
float ParallelSum( float array[], size_t n ) {
    Sum total;
    parallel_reduce( blocked_range<float*>( array, array+n ), total );
    return total.value;
}

The above example sums the values in the array. The parallel_reduce will do the reduction within the range of (array, array+n), to split the working body, and then join them by the return value for each split.

DOALL

The implementation of DOALL parallelism in Intel TBB will involve Parallel_for function. To better illustrate the usage, here we discuss a simple example. The following is the original code:

void SerialApplyFoo( float a[], size_t n ) {
    for( size_t i=0; i<n; ++i )
        Foo(a[i]);
}

After using Intel TBB, it could be switched to the following:

#include "tbb/blocked_range.h"
#include "tbb/parallel_for.h"

class ApplyFoo {
    float *const my_a;
public:
    void operator( )( const blocked_range<size_t>& r ) const {
        float *a = my_a;
        for( size_t i=r.begin(); i!=r.end( ); ++i )
            Foo(a[i]);
    }
    ApplyFoo( float a[] ) :
        my_a(a)
    {}
};

void ParallelApplyFoo( float a[], size_t n ) {
    parallel_for(blocked_range<size_t>(0,n,The_grain_size_You_Pick), ApplyFoo(a) );
}

The example is the simplest DOALL parallelism, similar as the one in textbook, and execution graph will be just similar as the one in DOALL section above. But with the help with this simple illustration, the TBB code just gives you a flavor of how it would be implemented in Intel Threading Building Blocks.

A little more to say, parallel_for takes an optional third argument to specify a partitioner, which I used "The_grain_size_You_Pick" to represent. If you want to manually divide the grain and assign the work to processors, you could specify that in the function. Or, you could use automatic grain provided TBB. The auto_partitioner provides an alternative that heuristically chooses the grain size so that you do not have to specify one. The heuristic attempts to limit overhead while still providing ample opportunities for load balancing. Then, the last three line of the TBB code above will be:

void ParallelApplyFoo( float a[], size_t n ) {
    parallel_for(blocked_range<size_t>(0,n), ApplyFoo(a), auto_partitioner( ) );
}

DOACROSS

We could find a good example in Intel TBB to implement a DOACROSS with the help of parallel_scan. As stated in the parallel_scan section, this function computes a parallel prefix, also known as parallel scan. This computation is an advanced concept in parallel computing which could be helpful in scenarios that appear to have inherently serial dependences, which could be loop-carried dependences.

Let's consider this scenario (which is actually the mathematical definition of parallel prefix):

T temp = id⊕;
for( int i=1; i<=n; ++i ) {
    temp = temp ⊕ x[i];
    y[i] = temp;
}

When we implement this in TBB using parallel_scan, it becomes:

using namespace tbb;
class Body {
    T sum;
    T* const y;
    const T* const x;
public:
    Body( T y_[], const T x_[] ) : sum(id⊕), x(x_), y(y_) {}
    T get_sum() const {return sum;}
    template<typename Tag>
    void operator()( const blocked_range<int>& r, Tag ) {
        T temp = sum;
        for( int i=r.begin(); i<r.end(); ++i ) {
            temp = temp ⊕ x[i];
            if( Tag::is_final_scan() )
                y[i] = temp;
        } 
        sum = temp;
    }
    Body( Body& b, split ) : x(b.x), y(b.y), sum(id⊕) {}
    void reverse_join( Body& a ) { sum = a.sum ⊕ sum;}
    void assign( Body& b ) {sum = b.sum;}
};
float DoParallelScan( T y[], const T x[], int n ) {
    Body body(y,x);
    parallel_scan( blocked_range<int>(0,n), body );
    return body.get_sum();
}

It is the second part (function DoParallelScan) that we have to focus on.

Actually, this example is just the scenario mentioned above that could take advantages of parallel_scan. The "inherently serial dependences" is taken care of by the functionality of parallel_scan. By computing the prefix, the serial code could be implemented in parallel with just one function.

DOPIPE

Pipeline class is the Intel TBB that performs pipelined execution. A pipeline represents pipelined application of a series of filters to a stream of items. Each filter operates in a particular mode: parallel, serial in order, or serial out of order. So this class can be used to implement a DOPIPE parallelism.

Here is a comparatively complex example about pipeline implementation. Also, if we look carefully, this is an example with both DOPIPE and DOACROSS:

#include <iostream>
#include "tbb/pipeline.h"
#include "tbb/tbb_thread.h"
#include "tbb/task_scheduler_init.h"
using namespace tbb;
char InputString[] = "abcdefg\n";
class InputFilter: public filter {
    char* my_ptr;
public:
    void* operator()(void*) {
        if (*my_ptr)
            return my_ptr++;
        else
            return NULL;
    }
    InputFilter() :
        filter( serial_in_order ), my_ptr(InputString) {}
};
class OutputFilter: public thread_bound_filter {
public:
    void* operator()(void* item) {
        std::cout << *(char*)item;
        return NULL;
    }
    OutputFilter() : thread_bound_filter(serial_in_order) {}
};
void RunPipeline(pipeline* p) {
    p->run(8);
}
int main() {
    // Construct the pipeline
    InputFilter f;
    OutputFilter g;
    pipeline p;
    p.add_filter(f);
    p.add_filter(g);
    // Another thread initiates execution of the pipeline
    tbb_thread t(RunPipeline,&p);
    // Process the thread_bound_filter with the current thread.
    while (g.process_item()!=thread_bound_filter::end_of_stream)
        continue;
    // Wait for pipeline to finish on the other thread.
    t.join();
    return 0;
}

The example above shows a pipeline with two filters where the second filter is a thread_bound_filter serviced by the main thread. The main thread does the following after constructing the pipeline: 1. Start the pipeline on another thread. 2. Service the thread_bound_filter until it reaches end_of_stream. 3. Wait for the other thread to finish.

POSIX Threads

POSIX Threads, or Pthreads, is a POSIX standard for threads. The standard, POSIX.1c, Threads extensions (IEEE Std 1003.1c-1995), defines an API for creating and manipulating threads.

Variable Scopes

Pthreads defines a set of C programming language types, functions and constants. It is implemented with a pthread.h header and a thread library.

There are around 100 Pthreads procedures, all prefixed "pthread_". The subroutines which comprise the Pthreads API can be informally grouped into four major groups:

  • Thread management: Routines that work directly on threads - creating, detaching, joining, etc. They also include functions to set/query thread attributes (joinable, scheduling etc.) E.g.pthread_create(), pthread_join().
  • Mutexes: Routines that deal with synchronization, called a "mutex", which is an abbreviation for "mutual exclusion". Mutex functions provide for creating, destroying, locking and unlocking mutexes. These are supplemented by mutex attribute functions that set or modify attributes associated with mutexes. E.g. pthread_mutex_lock(); pthread_mutex_trylock(); pthread_mutex_unlock().
  • Condition variables: Routines that address communications between threads that share a mutex. Based upon programmer specified conditions. This group includes functions to create, destroy, wait and signal based upon specified variable values. Functions to set/query condition variable attributes are also included. E.g. pthread_cond_signal(); pthread_cond_broadcast(); pthread_cond_wait(); pthread_cond_timedwait();pthread_cond_reltimedwait_np().
  • Synchronization: Routines that manage read/write locks and barriers. E.g. pthread_rwlock_rdlock(); pthread_rwlock_tryrdlock(); pthread_rwlock_wrlock();pthread_rwlock_trywrlock(); pthread_rwlock_unlock();pthread_barrier_init(); pthread_barrier_wait()

DOALL

The following is a simple code example in C, as DOALL parallelism, to print out each threads' ID#.

#include <pthread.h>
#include <stdio.h>
#include <stdlib.h>
#define NUM_THREADS     5
 
void *PrintHello(void *threadid)
{
   long tid;
 
   tid = (long)threadid;
   printf("Hello World! It's me, thread #%ld!\n", tid);
   pthread_exit(NULL);
}
 
int main (int argc, char *argv[])
{
   pthread_t threads[NUM_THREADS];
 
   int rc;
   long t;
   for(t=0; t<NUM_THREADS; t++){
      printf("In main: creating thread %ld\n", t);
      rc = pthread_create(&threads[t], NULL, PrintHello, (void *)t);
 
      if (rc){
         printf("ERROR; return code from pthread_create() is %d\n", rc);
         exit(-1);
      }
   }
   pthread_exit(NULL);
}

This loop contains only single statement which doesn't cross the iterations, so each iteration could be considered as a parallel task.

DOACROSS

When it comes to using Pthreads to implement DOACROSS, it could express functional parallelism easily, but make the parallelism unnecessarily complicated. See an example below: from POSIX Threads Programming by Blaise Barney

#include <pthread.h>
#include <stdio.h>
#include <stdlib.h>
#define NUM_THREADS	

void *BusyWork(void *t)
{
  int i;
  long tid;
  double result=0.0;
  tid = (long)t;
  printf("Thread %ld starting...\n",tid);
  for (i=0; i<1000000; i++)
  {
     result = result + sin(i) * tan(i);
  }
  printf("Thread %ld done. Result = %e\n",tid, result);
  pthread_exit((void*) t);
}

int main (int argc, char *argv[])
{
  pthread_t thread[NUM_THREADS];
  pthread_attr_t attr;
  int rc;
  long t;
  void *status;

  /* Initialize and set thread detached attribute */
  pthread_attr_init(&attr);
  pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_JOINABLE);

  for(t=0; t<NUM_THREADS; t++) {
     printf("Main: creating thread %ld\n", t);
     rc = pthread_create(&thread[t], &attr, BusyWork, (void *)t); 
     if (rc) {
        printf("ERROR; return code from pthread_create() 
               is %d\n", rc);
        exit(-1);
        }
     }

  /* Free attribute and wait for the other threads */
  pthread_attr_destroy(&attr);
  for(t=0; t<NUM_THREADS; t++) {
     rc = pthread_join(thread[t], &status);
     if (rc) {
        printf("ERROR; return code from pthread_join() 
               is %d\n", rc);
        exit(-1);
        }
     printf("Main: completed join with thread %ld having a status   
           of %ld\n",t,(long)status);
     }

printf("Main: program completed. Exiting.\n");
pthread_exit(NULL);
}

This example demonstrates how to "wait" for thread completions by using the Pthread join routine. Since some implementations of Pthreads may not create threads in a joinable state, the threads in this example are explicitly created in a joinable state so that they can be joined later.

DOPIPE

There is examples of using Posix Threads to implement DOPIPE parallelism, but unnecessarily complex. Due to the long length, we won't provide it here. If the reader is interested, it could be found in

  • Pthreads DOPIPE example
  • Comparison among the three

    A unified example

    We use a simple parallel example from Sourceforge.net to show how it will be implemented in the three packages, namely, POSIX Threads, Intel TBB, OpenMP, to show some common and differences among them.

    Following is the original code:

    Grid1 *g = new Grid1(0, n+1);
    Grid1IteratorSub it(1, n, g);
    DistArray x(g), y(g);
    ...
    float e = 0;
    ForEach(int i, it,
       x(i) += ( y(i+1) + y(i-1) )*.5;
       e += sqr( y(i) ); )
    ...
    

    Then we are going to show the implementations in different packages, and also make a brief summary of the three packages.

    In POSIX Thread

    POSIX Thread: Symmetric multi processing, e.g. SMP multi-processor computers, multi-core processors, virtual shared memory computer.

    Data layout: A single global memory. Each thread reads global shared data and writes to a private fraction of global data.

    A simplified translation of the example parallel-for loop is given below.

    Global declaration:

    #include <pthread.h>
    float *x, *y;
    float vec[8];
    int nn, pp;
    

    thread code:

    void *sub1(void *arg) {
       int p = (int)arg;
       float e_local = 0;
       for (int i=1+(nn*p)/pp; i<1+(nn*(p+1))/pp; ++i) {
         x[i] += ( y[i+1] + y[i-1] )*.5;
         e_local += y[i] * y[i];
       }
       vec[p] = e_local;
       return (void*) 0;
    }
    

    main code:

    x = new float[n+1];
    y = new float[n+1];
    ...
    float e = 0;
    int p_threads = 8;
    nn = n-1;
    pp = p_threads;
    pthread_t threads[8];
    pthread_attr_t attr;
    pthread_attr_init(&attr);
    for (int p=0; p<p_threads; ++p)
       pthread_create(&threads[p], &attr,
         sub1, (void *)p);
    for (int p=0; p<p_threads; ++p) {
       pthread_join(threads[p], NULL);
       e += vec[p];
    }
    ...
    delete[] x, y;
    
    In Intel Threading Building Blocks

    Intel TBB: A C++ library for thread programming, e.g. SMP multi-processor computers, multi-core processors, virtual shared memory computer.

    Data layout: A single global memory. Each thread reads global shared data and writes to a private fraction of global data.

    Translation of the example parallel-for loop is given below.

    Global:

    #include "tbb/task_scheduler_init.h"
    #include "tbb/blocked_range.h"
    #include "tbb/parallel_reduce.h"
    #include "tbb/cache_aligned_allocator.h"
    using namespace tbb;
    

    thread code:

    struct sub1 {
       float ee;
       float *x, *y;
       sub1(float *xx, float *yy) : ee(0), x(xx), y(yy) {}
       sub1(sub1& s, split) { ee = 0; x = s.x; y = s.y; }
       void operator() (const blocked_range<int> & r){
         float e = ee;
         for (int i = r.begin(); i!= r.end(); ++i) {
           x[i] += ( y[i+1] + y[i-1] )*.5;
           e += y[i] * y[i];
         }
         ee = e;
       }
       void join(sub1& s) { ee += s.ee; }
    };
    

    main code:

    task_scheduler_init init;
    ...
    float e;
    float *x = cache_aligned_allocator<float>().allocate(n+1);
    float *y = cache_aligned_allocator<float>().allocate(n+1);
    ...
    sub1 s(x, y);
    parallel_reduce(blocked_range<int>(1, n, 1000), s);
    e = s.ee;
    ...
    cache_aligned_allocator<float>().deallocate(x, n+1);
    cache_aligned_allocator<float>().deallocate(y, n+1);
    
    In OpenMP shared memory parallel code annotations

    OpenMP: Usually automatic paralleization with a run-time system based on a thread library.

    A simplified translation of the example parallel-for loop is given below.

    Global:

    float e;
    

    main code:

    float *x = new float[n+1];
    float *y = new float[n+1];
    ...
    e = 0;
    #pragma omp for reduction(+:e)
    for (int i=1; i<n; ++i) {
       x[i] += ( y[i+1] + y[i-1] )*.5;
       e += y[i] * y[i];
    }
    ...
    delete[] x, y;
    

    Summary: Difference among them

    • Pthreads works for all the parallelism and could express functional parallelism easily, but it needs to build specialized synchronization primitives and explicitly privatize variables, makes it more effort needed to switch a serial program in to parallel mode.
    • OpenMP can provide many performance enhancing features, such as atomic, barrier and flush synchronization primitives. It is very simple to use OpenMP to exploit DOALL parallelism, but the syntax for expressing functional parallelism is awkward.
    • Intel TBB relies on generic programming, it performs better with custom iteration spaces or complex reduction operations. Also, it provides generic parallel patterns for parallel while-loops, data-flow pipeline models, parallel sorts and prefixes, so it's better in cases go beyond loop-based parallelism.

    References

    1. OpenMP.org
    2. An Optimal Abtraction Model for Hardware Multithreading in Modern Processor Architectures
    3. Intel Threading Building Blocks 2.2 for Open Source Reference Manual
    4. NCSU CSC 506 Parallel Computing Systems
    5. Sourceforge.net
    6. OpenMP
    7. Barrier Optimization for OpenMP Program
    8. Performance Programming: Theory, Practice and Case Studies
    9. Intel® Threading Building Blocks, OpenMP, or native threads?
    10. POSIX Threads Programming by Blaise Barney, Lawrence Livermore National Laboratory
    11. Programing with POSIX Threads source code