CAPS project-team Compilation et Architectures pour Processeurs Superscalaires et Spécialisés.

Slides:



Advertisements
Similar presentations
DSPs Vs General Purpose Microprocessors
Advertisements

Slides Prepared from the CI-Tutor Courses at NCSA By S. Masoud Sadjadi School of Computing and Information Sciences Florida.
Lecture 8 Dynamic Branch Prediction, Superscalar and VLIW Advanced Computer Architecture COE 501.
1 Advanced Computer Architecture Limits to ILP Lecture 3.
André Seznec Caps Team IRISA/INRIA Design tradeoffs for the Alpha EV8 Conditional Branch Predictor André Seznec, IRISA/INRIA Stephen Felix, Intel Venkata.
Microprocessor Microarchitecture Multithreading Lynn Choi School of Electrical Engineering.
Computers Organization & Assembly Language Chapter 1 THE 80x86 MICROPROCESSOR.
Single-Chip Multiprocessor Nirmal Andrews. Case for single chip multiprocessors Advances in the field of integrated chip processing. - Gate density (More.
Instruction Level Parallelism (ILP) Colin Stevens.
EECS 470 Superscalar Architectures and the Pentium 4 Lecture 12.
Multiscalar processors
State Machines Timing Computer Bus Computer Performance Instruction Set Architectures RISC / CISC Machines.
RISC. Rational Behind RISC Few of the complex instructions were used –data movement – 45% –ALU ops – 25% –branching – 30% Cheaper memory VLSI technology.
The Vector-Thread Architecture Ronny Krashinsky, Chris Batten, Krste Asanović Computer Architecture Group MIT Laboratory for Computer Science
Simultaneous Multithreading: Maximizing On-Chip Parallelism Presented By: Daron Shrode Shey Liggett.
1 Thread level parallelism: It’s time now ! André Seznec IRISA/INRIA CAPS team.
1 Thread level parallelism: It’s time now ! André Seznec IRISA/INRIA CAPS team.
1 Advance Computer Architecture CSE 8383 Ranya Alawadhi.
CAPS project-team Compilation et Architectures pour Processeurs Superscalaires et Spécialisés.
Is Out-Of-Order Out Of Date ? IA-64’s parallel architecture will improve processor performance William S. Worley Jr., HP Labs Jerry Huck, IA-64 Architecture.
Lecture 1 1 Computer Systems Architecture Lecture 1: What is Computer Architecture?
1 Towards Optimal Custom Instruction Processors Wayne Luk Kubilay Atasu, Rob Dimond and Oskar Mencer Department of Computing Imperial College London HOT.
10/27: Lecture Topics Survey results Current Architectural Trends Operating Systems Intro –What is an OS? –Issues in operating systems.
Spring 2003CSE P5481 VLIW Processors VLIW (“very long instruction word”) processors instructions are scheduled by the compiler a fixed number of operations.
CASH: REVISITING HARDWARE SHARING IN SINGLE-CHIP PARALLEL PROCESSOR
1 Vulnerabilities on high-end processors André Seznec IRISA/INRIA CAPS project-team.
Chapter 8 CPU and Memory: Design, Implementation, and Enhancement The Architecture of Computer Hardware and Systems Software: An Information Technology.
Advanced Processor Technology Architectural families of modern computers are CISC RISC Superscalar VLIW Super pipelined Vector processors Symbolic processors.
1 CAPS Compilers Activities IRISA Campus Universitaire de Beaulieu Rennes.
A few issues on the design of future multicores André Seznec IRISA/INRIA.
VLIW Digital Signal Processor Michael Chang. Alison Chen. Candace Hobson. Bill Hodges.
André Seznec Caps Team IRISA/INRIA 1 High Performance Microprocessors André Seznec IRISA/INRIA
Chapter 1 Introduction to the Systems Approach
CAPS team Compiler and Architecture for superscalar and embedded processors.
Next Generation ISA Itanium / IA-64. Operating Environments IA-32 Protected Mode/Real Mode/Virtual Mode - if supported by the OS IA-64 Instruction Set.
© Wen-mei Hwu and S. J. Patel, 2005 ECE 511, University of Illinois Lecture 4: Microarchitecture: Overview and General Trends.
DSP Architectures Additional Slides Professor S. Srinivasan Electrical Engineering Department I.I.T.-Madras, Chennai –
CAPS project-team Compilation et Architectures pour Processeurs Superscalaires et Spécialisés.
Dept. of Computer Science - CS6461 Computer Architecture CS6461 – Computer Architecture Fall 2015 Lecture 1 – Introduction Adopted from Professor Stephen.
Power Awareness through Selective Dynamically Optimized Traces Roni Rosner, Yoav Almog, Micha Moffie, Naftali Schwartz and Avi Mendelson – Intel Labs,
Hybrid Multi-Core Architecture for Boosting Single-Threaded Performance Presented by: Peyman Nov 2007.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
Effective ahead pipelining of instruction block address generation André Seznec and Antony Fraboulet IRISA/ INRIA.
Chao Han ELEC6200 Computer Architecture Fall 081ELEC : Han: PowerPC.
Application Domains for Fixed-Length Block Structured Architectures ACSAC-2001 Gold Coast, January 30, 2001 ACSAC-2001 Gold Coast, January 30, 2001.
Architectural Effects on DSP Algorithms and Optimizations Sajal Dogra Ritesh Rathore.
On-chip Parallelism Alvin R. Lebeck CPS 220/ECE 252.
Computer Architecture: Multi-Core Processors: Why? Prof. Onur Mutlu Carnegie Mellon University.
“Processors” issues for LQCD January 2009 André Seznec IRISA/INRIA.
Fall 2012 Parallel Computer Architecture Lecture 4: Multi-Core Processors Prof. Onur Mutlu Carnegie Mellon University 9/14/2012.
Topics to be covered Instruction Execution Characteristics
CS Lecture 20 The Case for a Single-Chip Multiprocessor
Computer Architecture Principles Dr. Mike Frank
Visit for more Learning Resources
Hyperthreading Technology
CS775: Computer Architecture
Levels of Parallelism within a Single Processor
Computer Architecture Lecture 4 17th May, 2006
IA-64 Microarchitecture --- Itanium Processor
Mattan Erez The University of Texas at Austin
Computer Architecture: A Science of Tradeoffs
EE 4xx: Computer Architecture and Performance Programming
Overview Prof. Eric Rotenberg
Mattan Erez The University of Texas at Austin
Levels of Parallelism within a Single Processor
CAPS project-team Compilation et Architectures pour Processeurs Superscalaires et Spécialisés.
ALF Amdhal’s Law is Forever
Research: Past, Present and Future
Spring’19 Prof. Eric Rotenberg
Presentation transcript:

CAPS project-team Compilation et Architectures pour Processeurs Superscalaires et Spécialisés

André Seznec CAPS project-team Irisa-Inria 2 History of CAPS project-team  Project-team created in 1994:  “Compiler Parallel Architectures and Systems”  Common focus: high performance through optimizing the memory hierarchy  Comes from supercomputer architecture group:  Involved in Marie mini-supercomputer design  late participation in ACRI

André Seznec CAPS project-team Irisa-Inria 3 CAPS: Compiler and Architecture for Superscalar and Special purpose processors  Two interacting activities  microprocessor architecture (A. Seznec, P. Michaud) High performance Migrating high performance concepts to embedded systems  Performance oriented compilation (F. Bodin) High performance Embedded processors + Recently: Worst case execution time analysis (I. Puaut)

André Seznec CAPS project-team Irisa-Inria 4 CAPS « missions »  Defining the tradeoffs between:  what should be done through hardware  what can be done by the compiler  for maximum performance  or for minimum cost  or for minimum size, power..

André Seznec CAPS project-team Irisa-Inria 5 Issues on high performance processor architecture  Memory hierarchy management:  1 cycle L1 – 10 cycles L2 – 30 cycles L3 – 200 cycles memory  Branch prediction :  30 cycles penalty x N instructions per cycle  Single cycle next instruction block address generation ?  Complexity quadratic with issue width:  Register file, bypass network, issue logic  Single chip hardware thread parallelism is available:  How do we exploit it ?  Power/temperature

André Seznec CAPS project-team Irisa-Inria 6 Issues on code generation/software environments for embedded processors  ILP, caches are entering embedded processor world  Code generation must manage them  Binary compatibility is not critical, time-to-market is critical  Retargetable platforms are wanted: ISAs, architecture  Performance is not the only ultimate goal:  Code size/ performance  Power/ performance  System cost/ performance

André Seznec CAPS project-team Irisa-Inria 7 Recent scientific contributions (1) Processor architecture  Global history branch predictors and instruction fetch front-end  2bcgskew used in Compaq EV8  Pipelining the I-fetch front end  Limiting hardware complexity on superscalar processors  Dataflow prescheduling: instruction window  WSRS architecture: register file, bypass network and issue logic  Thread parallelism and single chip parallelism :  CASH: CMP and SMT hybrid  Execution migration: single thread on a multicore, to use all the cache space

André Seznec CAPS project-team Irisa-Inria 8 Recent scientific contributions (2) architecture/compiler interaction  ISA simulation:  ABSCISS : ISA and architecture retargetable high speed simulator for VLIW processor  IATO : simulation of out-of-order execution IA64  Low power and architecture configurability:  Cache reconfiguration at software level on phase basis  Hardware/software speculative management of data path and register file width  SWARP: retargetable C-to-C preprocessor to enhance multimedia instruction use

André Seznec CAPS project-team Irisa-Inria 9 Recent scientific contributions (3) compiler and software environments  Artificial intelligence in performance tuning  CAHT: case based reasoning for assisting performance tuning  Automatic derivation of compiler heuristics: using machine learning to derive compiler heuristics  Performance code size tradeoffs:  Iterative compilation  Mixing interpretation on compressed code and native execution

André Seznec CAPS project-team Irisa-Inria 10 “New-CAPS” objectives (1)  High-end microprocessor architecture: From “ultimate performance” to “ maintaining performance to cheaper”  Migrating “high-end” concepts to embedded processors:  (limited) O-O-O execution  Compiler/architecture power management

André Seznec CAPS project-team Irisa-Inria 11 “New-CAPS” research objectives (2)  Embedded systems are more and more complex:  performance often comes with unpredictability and unstabibility Dimensioning a system ? Real time constraints ?  Research on performance predictability and stability:  Predictable/stable performance oriented code generation  Predictable/stable performance oriented architecture

André Seznec CAPS project-team Irisa-Inria 12 “New-CAPS” research objectives (3)  On-chip thread parallelism is a new opportunity:  Homogeneous: SMT/CMP Tradeoffs, sharing, synchronization  Heterogeneous: single ISA –Power, performance, multiple ISAs (e.g. SoC) –Thread extraction

What can we bring in SCIPARC at architecture level ?

André Seznec CAPS project-team Irisa-Inria 14 CAPS pipeline background  « ancient » background in hardware management of ILP: both research and implementation  decoupled pipeline architectures: Involved in the design of Marie mini-supercomputer  OPAC, an hardware matrix floating-point coprocessor 1991: 300 ICs, a VLSI sequencer,..

André Seznec CAPS project-team Irisa-Inria CAPS background in microarchitecture  Solid knowledge in microprocessor architecture  technological watch on microprocessors  + research on processor architecture  + A. Seznec worked at Alpha Development Group in : Defined the EV8 branch predictor  + P. Michaud worked at Intel ( )

André Seznec CAPS project-team Irisa-Inria 16 Background in memory hierarchy  Interleaved memories for vector supercomputers (research):  + A. Seznec participated at Tarantula project: vector extension to Compaq EV8  International CAPS visibility in cache architecture :  skewed associative cache  + decoupled sectored cache

André Seznec CAPS project-team Irisa-Inria 17 Our expertise may help to define next machine in SCIPARC  Bring pipeline definition expertise  Bring memory hierarchy definition expertise  Help to remain simple  Help to enlarge possible application domains