Numerical Algorithms Quiz questions ITCS4145/5145, Parallel Programming March 14, 2013.

Slides:



Advertisements
Similar presentations
Communication costs of LU decomposition algorithms for banded matrices Razvan Carbunescu 12/02/20111.
Advertisements

Communication-Avoiding Algorithms Jim Demmel EECS & Math Departments UC Berkeley.
Slides for Parallel Programming Techniques & Applications Using Networked Workstations & Parallel Computers 2nd ed., by B. Wilkinson & M
Parallel Jacobi Algorithm Steven Dong Applied Mathematics.
CS 201 Compiler Construction
Block LU Factorization Lecture 24 MA471 Fall 2003.
Optimal PRAM algorithms: Efficiency of concurrent writing “Computer science is no more about computers than astronomy is about telescopes.” Edsger Dijkstra.
Parallel Matrix Operations using MPI CPS 5401 Fall 2014 Shirley Moore, Instructor November 3,
Partitioning and Divide-and-Conquer Strategies ITCS 4/5145 Parallel Computing, UNC-Charlotte, B. Wilkinson, Jan 23, 2013.
Numerical Algorithms ITCS 4/5145 Parallel Computing UNC-Charlotte, B. Wilkinson, 2009.
Numerical Algorithms Matrix multiplication
Parallel Programming: Techniques and Applications Using Networked Workstations and Parallel Computers Chapter 11: Numerical Algorithms Sec 11.2: Implementing.
Advanced Topics in Algorithms and Data Structures Lecture pg 1 Recursion.
Numerical Algorithms • Matrix multiplication
Parallelizing C Programs Using Cilk Mahdi Javadi.
CS Section 600 CS Section 002 Dr. Angela Guercio Spring 2010.
CS603 Process Synchronization February 11, Synchronization: Basics Problem: Shared Resources –Generally data –But could be others Approaches: –Model.
Design of parallel algorithms
Complexity 19-1 Parallel Computation Complexity Andrei Bulatov.
Data Locality CS 524 – High-Performance Computing.
Design of parallel algorithms Matrix operations J. Porras.
Dense Matrix Algorithms CS 524 – High-Performance Computing.
1 Matrix Addition, C = A + B Add corresponding elements of each matrix to form elements of result matrix. Given elements of A as a i,j and elements of.
Basic PRAM algorithms Problem 1. Min of n numbers Problem 2. Computing a position of the first one in the sequence of 0’s and 1’s.
Table of Contents Matrices - Multiplication Assume that matrix A is of order m  n and matrix B is of order p  q. To determine whether or not A can be.
DATA LOCALITY & ITS OPTIMIZATION TECHNIQUES Presented by Preethi Rajaram CSS 548 Introduction to Compilers Professor Carol Zander Fall 2012.
Simple Load Balancing CS550 Operating Systems. Announcements Project will be posted – TBA This project will use the client-server model and will require.
Tuning Libraries to Effectively Exploit Memory Prof. Misha Kilmer Emily Reid Stacey Ecott.
Very Long Instruction Word (VLIW) Architecture. VLIW Machine It consists of many functional units connected to a large central register file Each functional.
“elbowing out” Processors used Speedup Efficiency timeexecution Parallel Processors timeexecution Sequential Efficiency   
1 L ECTURE 2 Matrix Multiplication Tableau Construction Recurrences (Review) Conclusion Merge Sort.
Cache Basics Define temporal and spatial locality.
Slides for Parallel Programming Techniques & Applications Using Networked Workstations & Parallel Computers 2nd ed., by B. Wilkinson & M
Imperialism and World Wars Review Question #1 3 questions.
Lab 2 Parallel processing using NIOS II processors
Time Parallel Simulations I Problem-Specific Approach to Create Massively Parallel Simulations.
Graph Theory. undirected graph node: a, b, c, d, e, f edge: (a, b), (a, c), (b, c), (b, e), (c, d), (c, f), (d, e), (d, f), (e, f) subgraph.
1 Complete this to a Pfaffian orientation (all internal faces have an odd number of clockwise arcs).
Lecture 9 Architecture Independent (MPI) Algorithm Design
Worksheet Answers Matrix worksheet And Matrices Review.
PARALLEL COMPUTATION FOR MATRIX MULTIPLICATION Presented By:Dima Ayash Kelwin Payares Tala Najem.
CS 221 – May 22 Timing (sections 2.6 and 3.6) Speedup Amdahl’s law – What happens if you can’t parallelize everything Complexity Commands to put in your.
Exercise 1: Maximum element for the following code a- what is the basic operation? b- what is C(n)? d-what is the running time?
Classification of parallel computers Limitations of parallel processing.
Sieve of Eratosthenes Quiz questions ITCS4145/5145, Parallel Programming Oct 24, 2013.
Numerical Algorithms Chapter 11.
Quiz for Week #5.
Section 7: Memory and Caches
Sorting Quiz questions
Course Description Algorithms are: Recipes for solving problems.
Assignment 4 – (a) Consider a symmetric MP with two processors and a cache invalidate write-back cache. Each block corresponds to two words in memory.
Set-Associative Cache
Interconnect with Cache Coherency Manager
Parallel Matrix Operations
Lecture 5 PRAM Algorithms (cont.)
Numerical Algorithms • Parallelizing matrix multiplication
Parallel Programming in C with MPI and OpenMP
Numerical Algorithms Quiz questions
Quiz Questions Parallel Programming Parallel Computing Potential
Quiz Questions Seeds pattern programming framework
Quiz Questions Seeds pattern programming framework
Matrix Addition and Multiplication
Quiz Questions Parallel Programming Parallel Computing Potential
Quiz Questions Parallel Programming Parallel Computing Potential
Quiz Questions Parallel Programming Parallel Computing Potential
Matrix Addition, C = A + B Add corresponding elements of each matrix to form elements of result matrix. Given elements of A as ai,j and elements of B as.
Algorithm Analysis T(n) O() Growth Rates 5/21/2019 CS 303 – Big ‘Oh’
Course Description Algorithms are: Recipes for solving problems.
Use of Submatrices If the original matrix is n x n, the four submatrices are n/2 x n/2 each; for simplicity we assume powers of two To find the result.
Presentation transcript:

Numerical Algorithms Quiz questions ITCS4145/5145, Parallel Programming March 14, 2013.

What is meant by the term cost-optimal? a)When the cost of computers is the lowest b)When the parallel time complexity x number of processors used = sequential time complexity c)When the number of lines of code is at the lower bound (minimum) d)When the sequential time complexity x number of processors used = parallel time complexity

What is the sequential time complexity of matrix multiplication (Matrices each n x n) a)O(log n) b)O(n) c)O(n 2 ) d)O(n 3 )

What is the parallel time complexity of matrix multiplication using n 2 processors (Matrices each n x n) a)O(log n) b)O(n) c)O(n 2 ) d)O(n 3 )

How many processors would be needed to perform matrix multiplication using block multiplication with s x s blocks (sub-matrices) and n x n matrices? (Assume n is a multiple of s) a)n/s b)(n/s) 2 c)s 2 d)None of the other answers

What is the parallel time complexity of matrix multiplication using block multiplication with s x s blocks (sub-matrices) using the number of processors from the last question? (Assume n is a multiple of s) a). b). c). d)None of the other answers

Discussion How can block matrix multiplication take advantage of cache memory?