Introduction to Parallel Computing: mudanças entre as edições

De Pontão Nós Digitais
Ir para navegaçãoIr para pesquisar
(→‎Groups to Project 1: postei trabalho do julio)
Linha 142: Linha 142:
* '''Hybrid technologies (Distributed + Cuda)''': Ricardo Dias  ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/dias/apresentacao.pdf slides]) ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/dias/codigo/ code])
* '''Hybrid technologies (Distributed + Cuda)''': Ricardo Dias  ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/dias/apresentacao.pdf slides]) ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/dias/codigo/ code])
* '''Cuda/OpenCL''': Joel Sanchez ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/sanchez/gpu_cuda.pdf slides]) ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/sanchez/codigo/ code])
* '''Cuda/OpenCL''': Joel Sanchez ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/sanchez/gpu_cuda.pdf slides]) ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/sanchez/codigo/ code])
* '''OPENGL/GLSL''' : Luiz Rosalba (Apresentacao: 08/11/2012)
* '''OPENGL/GLSL''' : Luiz Rosalba - 08/11/2012
* '''mapreduce/hadoop''' : Julio Stutz ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/stutz/mapreduce-hadoop.pdf slides])
* '''mapreduce/hadoop''' : Julio Stutz ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/stutz/mapreduce-hadoop.pdf slides])
* '''Google Go''' : Mateus Guida ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/guida/google_go.pdf slides])
* '''Google Go''' : Mateus Guida ([http://www.lems.brown.edu/~rfabbri/stuff/parallel-computing/technologies/guida/google_go.pdf slides])

Edição das 13h11min de 13 de novembro de 2012

This is the main page of a graduate-level course in parallel computing being taught in 2012/2 at the Polytechnic Institute IPRJ/UERJ. It is generally useful for programmers at the advanced level in the fields of scientific and multimedia programming.

Matrix6.png

General Info

  • Instructor: prof. Ricardo Fabbri, Ph.D. Brown University
  • Meeting times: Tues 12:30pm-2pm, Thursdays 2:20pm - 4pm
  • Evaluation criteria: 1 quizz at the end of the term (60%), plus practical projects (40%).
  • Forum for file exchange and discussion: uerj.tk

Pre-requisites

Approximate Content

The course focuses on software techniques for parallel computing. We are aiming at a comprehensive treatment on different types of practical parallel programming techniques

  • process-oriented parallel programming
  • thread programming/thread safety
  • single-core vector instructions
  • multi-processor and multi-core programming
  • mapreduce/hadoop
  • MPI
  • Cuda/OpenCL

Each of the above techniques would merit a course of their own, as done in many of the best universities. Therefore we aim at attaining a practical familiarity with each, in the first half of the course, and we will specialize in the later part of the course as to help the students' graduate research.

Main Resources

Lectures

NOTE: There will be no lectures from Oct 4 - Oct 19 2012

Partial listing & Tentative Outline

  1. Overview of parallel computing: https://computing.llnl.gov/tutorials/parallel_comp/
  2. Review of Linux:
    1. See the book Running Linux from the Recommended Reading
  3. Review of C/C++ (finished 28/Aug)
  4. Fundamental programming techniques: processes and threads
    1. Read The Unix Programming Environment for some classic multi-process programming[5]
    2. What is a process: http://linuxgazette.net/133/saha.html
    3. 30/Aug: We'll be following the McKenney Book, ch 3 at the lab.
    4. Up to (and including) 18/Sept: We followed McKenney's code samples for counting algorithms and his Threading API in detail
      1. What is variable alignment: [6] [7]
      2. How to determine the cache line size for your machine: [8]. This will give you better results for threading algorithms.
      3. Install some documentation useful for POSIX threading: sudo apt-get install glibc-doc manpages-posix-dev
    5. 20/Sept: Finished statistic counters and showed how to read assembly language generated by GCC, and how to disassemble binary objects
    6. 25/Sept: Finished theoretical thread content for the course: up until ch. 5 of McKenney's Book + Semaphores + Thread Safety.
      1. Homework 2 (see below)
  5. Cuda/OpenCL (02/Oct)
    1. Project 1 handout (see below) due Oct. 23
  6. Mapreduce/Hadoop
  7. MPI

Homework

Homework 1

Our insitute is planning on buying a supercomputer. Therefore, our first homework will consist on studying these systems in detail.

  • The class shall be divided into groups of 2 people.
  • Each group must pick one top 500 (http://www.top500.org) supercomputer and write on this wiki:
    • Team number:
    • Supercomputer:
  • Each class will start with a 20min presentation from a random group (or a volunteer group).
  • A handout must also be provided on Tue August 28
  • Your project must not be a plain copy from wikipedia!


Contents of the presentation and report

  • Overview of the system
  • Glossary of terms
  • How to program for the system
  • Software infrastructure
  • How does this differ from a conventional cluster.


Groups

  • Team 1: Julio Stutz & Joel Sanchez
  • Team 2: Claudir Oliver & Luiz Rosalba
  • Team 3: Mateus Guida
    • Supercomputer: Sequoia (Laborátório Nacional Lawrence Livermore - Califórnia - EUA) (report) (slides)
  • Team 4: Lyvia Aloquio & Stella Oggioni
    • Supercomputer: Grifo04 (PETROBRAS/Brazil)
  • Team 5: Ricardo Dias
    • Supercomputer: Mole-8.5 (Institute of Process Engineering, Chinese Academy of Sciences/China) (report) (slides)
  • Team 6: Frederico Santos
    • Supercomputer: Helios (International Fusion Energy Research Centre (IFERC), EU(F4E) - Japan Broader Approach collaboration) (report)
  • Team 7: Thiago Luiz
    • Supercomputer: Pleiades - NASA/Ames Research Center/ NAS - United States.

Homework 2: limit counters

Describe the limit counter implementations of sec. 4.3 in McKenney's Book, answering:

  1. list and explain the advantages of each counter
  2. explain the main concepts of each method in your words

Due date: Oct 1 (was: Sept 27)

Project 1

  • The class will be divided into interest groups
  • The interest groups will each pick one of the following technologies or else propose another one of their liking.
    • mapreduce/hadoop
    • MPI
    • Cuda/OpenCL
    • Hybrid technologies (Distributed + Cuda)
    • Languages designed for concurrency: Google Go, Erlang, etc.
    • Programming for the Playstation 3 parallel architecture[9]
    • OPENGL/GLSL
  • The project will consist in a series of presentations by the group members individually.
  • Grade will be asigned to each presentation individually


Groups to Project 1

  • MPI: Claudir Oliveira - 25/10/2012 (slides) (code)
  • Hybrid technologies (Distributed + Cuda): Ricardo Dias (slides) (code)
  • Cuda/OpenCL: Joel Sanchez (slides) (code)
  • OPENGL/GLSL : Luiz Rosalba - 08/11/2012
  • mapreduce/hadoop : Julio Stutz (slides)
  • Google Go : Mateus Guida (slides)
  • Programming for the Playstation 3 parallel architecture: Thiago Luiz

Project 2

  • This project will consist on a practical programming problem from the student's research. The student is required to describe a problem from his research and present a parallel implementation of some aspect of it.
  • This will be evaluated through actual code and acompanying monograph.

Class exercies

  1. In-class assignment (extra .1 in the final test): Compare 3 different implementation of statistic counters: array of per-thread variables, eventuallly consistent counter and GCC __thread variables
    1. Only Julio Stutz scored the extra +0.1 on the final exam.


Keywords

Portuguese: Programação Paralela, Introdução à Computação Paralela