The document provides an overview of parallel programming using MPI and OpenMP. It discusses key concepts of MPI including message passing, blocking and non-blocking communication, and collective communication operations. It also covers OpenMP parallel programming model including shared memory model, fork/join parallelism, parallel for loops, and shared/private variables. The document is intended as lecture material for an introduction to high performance computing using MPI and OpenMP.