Parallel Programming with NVIDIA CUDA
If parallelization of your algorithm is possible, using CUDA will speed up your computations dramatically, allowing you to make the most out of your hardware.
The main challenge consists in deciding how to partition your problem into chunks suitable for parallel execution. As with so many other aspects in parallel programming, this is where experience and—why not—imagination come into play.
Additional techniques offer room for even more improvement. In particular, the on-chip shared memory of each compute node allows further speedup of the computation process.
Alejandro Segovia is a parallel programming advisor for CoroWare. He is also a contributing partner at RealityFrontier. He works in 3-D graphic development and GPU acceleration. Alejandro was recently a visiting scientist at the University of Delaware where he investigated CUDA from an academic standpoint. His findings were published at the IEEE IPCCC Conference in 2009.
- Readers' Choice Awards 2013
- A Plexible Pi
- Linux Kernel News - November 2013
- Advanced Hard Drive Caching Techniques
- Sublime Text: One Editor to Rule Them All?
- Mars Needs Women
- Raspberry Pi: the Perfect Home Server
- Tech Tip: Really Simple HTTP Server with Python
- December 2013 Issue of Linux Journal: Readers' Choice
- RSS Feeds
- rilakkuma onesie
38 min 11 sec ago
- flying squirrel onesie
40 min 20 sec ago
- animal onesies for adults
42 min 19 sec ago
- animal onesies
43 min 28 sec ago
- stitch onesie
44 min 11 sec ago
- totoro onesie
45 min 50 sec ago
- dinosaur onesie
49 min 42 sec ago
- pikachu onesie
53 min 36 sec ago
- While copy.com is nice, it's
5 hours 55 min ago
- Evangelist/Advocate - 5th place - Dedoimedo
7 hours 47 min ago