Parallel Computing in the Computer Science Curriculum > Modules

Modules

Want to know more about modules?

Find out more about modules and their contents.

Have a module of your own?

Contribute to the site by submitting your own module. Your submission will be reviewed by CS In Parallel to determine what categories it should be listed under. After that process, it will become available to all viewers of this site.

The Module Collection



Help

Show all pages

Current Search Limits

Concepts

showing only Message Passing Show all Concepts

Results 1 - 3 of 3 matches

Parallel Computing Concepts
Richard Brown
This concept module will introduce a core of parallel computing notions that CS majors and minors should know in preparation for the era of manycore computing, including parallelism categories, concurrency issues and solutions, and programming strategies.

Distributed Computing Fundamentals
Elizabeth Shoop
Message Passing Interface (MPI) is a programming model widely used for parallel programming in a cluster. Using MPI, programmers can design methods to divide large data and perform the same computing task on segments of it and then and distribute those tasks to multiple processing units within the cluster. In this module, we will learn important and common MPI functions as well as techniques used in 'distributed memory' programming on clusters of networked computers.

Heterogeneous Computing
Elizabeth Shoop;
Message Passing Interface (MPI) is a programming model widely used for parallel programming in a cluster. NVIDIA®'s CUDA, a parallel computing platform and programming model, uses GPU for parallel computation problems. This module will explore ways to combine these two parallel computing platforms to make parallel computation more efficient.



      Next Page »