This page will give you a short introduction concerning the concepts enforced throughout libXDispatch and also help you getting started by first integrating libXDispatch into your project and writing the first lines of code afterwards. libXDispatch aims to provide an environment within which you can write parallelized code using the concepts of Grand Central Dispatch while keeping cross-platfrom compatibility and thus leaves it up to you, which operating systems you intend to target.
Although libXDispatch is provided as ready-to-use library as well, the recommended approach is to include the sources into your own development files. This will enable you to use this source tree without needing to have libXDispatch pre-installed on every development platform.
In the following we assume, that we have a project we are currently working on. This project is managed using the subversion software and CMake and has the following structure:
All commands shown will assume your are located within the directory containing the src and libconfig folders as shown above.
So to get started, you could either download the latest sources and copy them within your source tree or simply include the stable branch using the svn:external property:
svn propset svn:external "http://opensource.mlba-team.de/svn/xdispatch/branches/stable libxdispatch .
Please make sure not to copy the sources in a directory called xdispatch as this will result in build errors. Now edit your CMakeLists.txt to contain the following lines before adding your executable from the src folder:
add_subdirectory(libxdispatch) include_directories(libxdispatch/include)
This will ensure your project knows about and builds libXDispatch. The second line will tell your compiler where to search for the headers of libXDispatch. To finish your preparations we finally need to link our executable with libXDispatch:
add_executable(dispatch_demo ${SRC}) target_link_libraries(dispatch_demo xdispatch)
In case you do not intend to use CMake you can of course still download the binaries and configure your environment to link agains libxdispatch.so on Linux, xdispatch.dylib on Mac OS and xdispatch.lib on Windows. Also make sure to add the libXDispatch include folder to your compiler's search path. This will have the same effect as the afore mentioned CMake scripts. Windows will force you to copy the xdispatch.dll and dispatch.dll files to the same directory as your executable as well.
Using libXDispatch within your source code is pretty straight forward as all you need to do is to include the headers within your source files - that's it.
#include <xdispatch/dispatch>
All functions are located in the xdispatch namespace. In the following I will demonstrate some use cases occuring when trying to parallelize the code. I will assume that your are either using gcc-4.5+, Visual Studio 2010 or clang as your compiler as enables us to utilize lambdas. For those not being able to use a "modern" compiler, please have a look at Parallel code using xdispatch::operations.
The most obvious use case is that you want to move some heavy calculation work off the main thread and into a background worker. Now without using libXDispatch, you'd probably be writing something similar to this:
#include <pthread.h> #include <iostream> // declared somewhere else class SomeData { bool finished; pthread_mutex_t lock; ... }; /* The worker function doing all the stuff */ void* do_work(void* dt){ SomeData* data = (SomeData*)dt; // execute the heavy code do_calculations(data); // notify the main thread we are finished pthread_mutex_lock(&data->lock); data->finished = true; pthread_mutex_unlock(%data->lock); } /* This function is getting called from your main thread also powering the user interface */ void some_function(){ SomeData* sd = new SomeData(); fill_data(sd); pthread_t worker; if(pthread_create(&worker, NULL, do_work, NULL, (void*)sd)){ std::cerr << "Failed to create worker thread" << std::endl; return; } pthread_mutex_lock(&sd->lock); while(!sd->finished){ pthread_mutex_unlock(&sd->lock); // process all events on the main thread process_events(); pthread_mutex_lock(&sd->lock); } // ok, now the worker has finished, show the results within the gui show_calc_results(sd); delete sd; }
So this is an example using pthreads. When writing for windows as well, we'd probably need to write another version using WindowsThreads or need to use a library such as OpenThreads or boost::threads. When using libXDispatch, we can express this code much more effectively - and still maintain cross platform compatibility:
#include <xdispatch/dispatch> // declared somewhere else class SomeData { ... }; /* This function is getting called from your main thread also powering the user interface */ void some_function(){ SomeData* sd = new SomeData(); fill_data(sd); xdispatch::global_queue().async(${ // execute the heavy code do_calulations(sd); // notify the gui that we are finished xdispatch::main_queue().async(${ show_calc_results(sd); delete sd; }); }); }
There's no need for manual thread creation and so on. Also note, that we can use all variables declared within some_function() within our lambda code ${ .. }. It's just as easy when you want to parallelize a loop. Let's assume the following piece of code (Please note this is still a very simple calculation):
#include <vector> #include <cmath> // declared somewhere else class SomeData { ... std::vector<double> a; std::vector<double> b; std::vector<double> c; std::vector<double> results; }; void do_calculations(SomeData* sd){ // our output will go in here sd->results = std::vector<double>(sd->a.size()); // the calculation - running on one thread only for(unsigned int i = 0; i < a.size(); i++){ sd->results[i] = 0; for(unsigned int j = 0; j < b.size(); j++){ for(unsigned int z = 0; z < c.size(); z++){ sd->results[i] += std::pow(sd->b[j], sd->a[i]) * std::sin(sd->c[z]); } } } }
Now to parallelize this piece of code using libXDispatch you can simply write:
#include <vector> #include <cmath> #include <xdispatch/dispatch> // declared somewhere else class SomeData { ... std::vector<double> a; std::vector<double> b; std::vector<double> c; std::vector<double> results; }; void do_calculations(SomeData* sd){ // our output will go in here sd->results = std::vector<double>(sd->a.size()); // the calculation - running on multiple threads xdispatch::global_queue().apply($(size_t i){ sd->results[i] = 0; for(unsigned int j = 0; j < b.size(); j++){ for(unsigned int z = 0; z < c.size(); z++){ sd->results[i] += std::pow(sd->b[j], sd->a[i]) * std::sin(sd->c[z]); } } }, a.size()); }
libXDispatch is also providing mechanisms for making some piece of code perfectly threadsafe. So again assume the following piece of code:
#include <pthread.h> static pthread_mutex_t lock = PTHREAD_MUTEX_INITIALIZER; /* So this function is called from several threads */ void worker(){ // some work ... pthread_mutex_lock(&lock); // do some critical work if(already_done){ // we might have finished here pthread_mutex_unlock(&lock); return; } // do some other critical work pthread_mutex_lock(&lock); // some other work ... }
We will have to make sure the mutex is cleanly unlocked whenever leaving the critical section. And what happens if an exception is thrown from within that we do not catch? This might result in a deadlock. All this can be easily resolved by using the following expression:
#include <xdispatch/dispatch> /* So this function is called from several threads */ void worker(){ // some work ... synchronized { // do some critical work if(already_done) // we might have finished here return; // do some other critical work } // some other work ... }
No need to handle the locking by yourself, all is done magically - and it is ensured that the lock will be automatically cleared whenever you leave the section marked by the brackets. For further details about this, please see the documentation on the xdispatch::synclock. Please note that his functionality is available on compilers without lambda support as well.
All the examples shown above can also be written without using lambdas. So for example the parallel loop can also be expressed using an xdispatch::iteration_operation:
#include <vector> #include <cmath> #include <xdispatch/dispatch> // declared somewhere else class SomeData { ... std::vector<double> a; std::vector<double> b; std::vector<double> c; std::vector<double> results; }; class InnerCalculation : public xdispatch::iteration_operation { SomeData* sd; public: InnerCalculation(SomeData* d) : sd(d) {} void operator()(size_t i){ sd->results[i] = 0; for(unsigned int j = 0; j < b.size(); j++){ for(unsigned int z = 0; z < c.size(); z++){ sd->results[i] += std::pow(sd->b[j], sd->a[i]) * std::sin(sd->c[z]); } } } } void do_calculations(SomeData* sd){ // our output will go in here sd->results = std::vector<double>(sd->a.size()); // the calculation - running on multiple threads xdispatch::global_queue().apply(new InnerCalculation(sd), a.size()); }
There is no need to worry about memory leaks - xdispatch will automatically delete the iteration_operation once it has finished execution.
The examples above showed only some of the functionality and power of libXDispatch. Of course there also is a plean C interface and Qt integration provided within QtDispatch. For further exploration, we recommend browsing the API documentation and having a look at the various unittests.
There is also a lot more concepts to explore. For example you could create your own queues and not only use the automatically provided global queues. For understanding the idea of serial and concurrent queues and the usage of setting a target for a queue, we recommend to read the document "Apple Technical Brief on Grand Central Dispatch" and have a look at Apple's Concurrency Programming Guide