EGRE 426 Fall 09 Chapter Three

Slides:



Advertisements
Similar presentations
Chapter Three.
Advertisements

1 Chapter Three Last revision: 4/17/ Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture.
©UCB CPSC 161 Lecture 6 Prof. L.N. Bhuyan
1 CONSTRUCTING AN ARITHMETIC LOGIC UNIT CHAPTER 4: PART II.
Chapter 3 Arithmetic for Computers. Multiplication More complicated than addition accomplished via shifting and addition More time and more area Let's.
Arithmetic CPSC 321 Computer Architecture Andreas Klappenecker.
Arithmetic CPSC 321 Computer Architecture Andreas Klappenecker.
Arithmetic II CPSC 321 E. J. Kim. Today’s Menu Arithmetic-Logic Units Logic Design Revisited Faster Addition Multiplication (if time permits)
1 Representing Numbers Using Bases Numbers in base 10 are called decimal numbers, they are composed of 10 numerals ( ספרות ) = 9* * *10.
1  2004 Morgan Kaufmann Publishers Chapter Three.
1 Chapter 4: Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture Assembly Language and.
Chapter Four Arithmetic and Logic Unit
Arithmetic I CPSC 321 Andreas Klappenecker. Administrative Issues Office hours of TA Praveen Bhojwani: M 1:00pm-3:00pm.
Arithmetic-Logic Units CPSC 321 Computer Architecture Andreas Klappenecker.
1  2004 Morgan Kaufmann Publishers Lectures for 3rd Edition Note: these lectures are often supplemented with other materials and also problems from the.
1 ECE369 Chapter 3. 2 ECE369 Multiplication More complicated than addition –Accomplished via shifting and addition More time and more area.
1  1998 Morgan Kaufmann Publishers Chapter Four Arithmetic for Computers.
Chapter 3 Arithmetic for Computers. Arithmetic Where we've been: Abstractions: Instruction Set Architecture Assembly Language and Machine Language What's.
Arithmetic for Computers
(original notes from Prof. J. Kelly Flanagan)
1 Bits are just bits (no inherent meaning) — conventions define relationship between bits and numbers Binary numbers (base 2)
1 CHAPTER 4: PART I ARITHMETIC FOR COMPUTERS. 2 The MIPS ALU We’ll be working with the MIPS instruction set architecture –similar to other architectures.
1 CS/COE0447 Computer Organization & Assembly Language Chapter 3.
1 Modified from  Modified from 1998 Morgan Kaufmann Publishers Chapter Three: Arithmetic for Computers citation and following credit line is included:
Computer Arithmetic.
1  2004 Morgan Kaufmann Publishers Instructions: bne $t4,$t5,Label Next instruction is at Label if $t4≠$t5 beq $t4,$t5,Label Next instruction is at Label.
Computing Systems Basic arithmetic for computers.
1 ECE369 Sections 3.5, 3.6 and ECE369 Number Systems Fixed Point: Binary point of a real number in a certain position –Can treat real numbers as.
Chapter 6-1 ALU, Adder and Subtractor
07/19/2005 Arithmetic / Logic Unit – ALU Design Presentation F CSE : Introduction to Computer Architecture Slides by Gojko Babić.
1 EGRE 426 Fall 08 Chapter Three. 2 Arithmetic What's up ahead: –Implementing the Architecture 32 operation result a b ALU.
1  1998 Morgan Kaufmann Publishers Arithmetic Where we've been: –Performance (seconds, cycles, instructions) –Abstractions: Instruction Set Architecture.
Csci 136 Computer Architecture II – Constructing An Arithmetic Logic Unit Xiuzhen Cheng
Computing Systems Designing a basic ALU.
1 Lecture 6 BOOLEAN ALGEBRA and GATES Building a 32 bit processor PH 3: B.1-B.5.
CDA 3101 Fall 2013 Introduction to Computer Organization The Arithmetic Logic Unit (ALU) and MIPS ALU Support 20 September 2013.
1  2004 Morgan Kaufmann Publishers Performance is specific to a particular program/s –Total execution time is a consistent summary of performance For.
1 ELEN 033 Lecture 4 Chapter 4 of Text (COD2E) Chapters 3 and 4 of Goodman and Miller book.
1  2004 Morgan Kaufmann Publishers Lets Build a Processor Almost ready to move into chapter 5 and start building a processor First, let’s review Boolean.
Computer Architecture Lecture Notes Spring 2005 Dr. Michael P. Frank Competency Area 4: Computer Arithmetic.
Addition, Subtraction, Logic Operations and ALU Design
CML CML CS 230: Computer Organization and Assembly Language Aviral Shrivastava Department of Computer Science and Engineering School of Computing and Informatics.
순천향대학교 정보기술공학부 이 상 정 1 3. Arithmetic for Computers.
Prof. Hsien-Hsin Sean Lee
Lec 11Systems Architecture1 Systems Architecture Lecture 11: Arithmetic for Computers Jeremy R. Johnson Anatole D. Ruslanov William M. Mongan Some or all.
10/7/2004Comp 120 Fall October 7 Read 5.1 through 5.3 Register! Questions? Chapter 4 – Floating Point.
1 Arithmetic Where we've been: –Abstractions: Instruction Set Architecture Assembly Language and Machine Language What's up ahead: –Implementing the Architecture.
Computer Arthmetic Chapter Four P&H. Data Representation Why do we not encode numbers as strings of ASCII digits inside computers? What is overflow when.
9/23/2004Comp 120 Fall September Chapter 4 – Arithmetic and its implementation Assignments 5,6 and 7 posted to the class web page.
EE204 L03-ALUHina Anwar Khan EE204 Computer Architecture Lecture 03- ALU.
May 2, 2001System Architecture I1 Systems Architecture I (CS ) Lecture 11: Arithmetic for Computers * Jeremy R. Johnson May 2, 2001 *This lecture.
By Wannarat Computer System Design Lecture 3 Wannarat Suntiamorntut.
1 CPTR 220 Computer Organization Computer Architecture Assembly Programming.
Based on slides from D. Patterson and www-inst.eecs.berkeley.edu/~cs152/ COM 249 – Computer Organization and Assembly Language Chapter 3 Arithmetic For.
1 Chapter 3 Arithmetic for Computers Lecture Slides are from Prof. Jose Delgado-Frias, Mr. Paul Wettin, and Prof. Valeriu Beiu (Washington State University.
1 (Based on text: David A. Patterson & John L. Hennessy, Computer Organization and Design: The Hardware/Software Interface, 3 rd Ed., Morgan Kaufmann,
Computer System Design Lecture 3
Computer Arthmetic Chapter Four P&H.
Arithmetic Where we've been:
Arithmetic Logical Unit
Computer Arithmetic Multiplication, Floating Point
CS/COE0447 Computer Organization & Assembly Language
October 17 Chapter 4 – Floating Point Read 5.1 through 5.3 1/16/2019
CS/COE0447 Computer Organization & Assembly Language
COMS 361 Computer Organization
A 1-Bit Arithmetic Logic Unit
October 8 Rules for Programming Assignments ASK QUESTIONS!
Chapter 3 Arithmetic for Computers
COMS 361 Computer Organization
Presentation transcript:

EGRE 426 Fall 09 Chapter Three

Arithmetic What's up ahead: Implementing the Architecture ALU 32 operation result a b ALU

Numbers Bits are just bits (no inherent meaning) — conventions define relationship between bits and numbers Binary numbers (base 2) 0000 0001 0010 0011 0100 0101 0110 0111 1000 1001... decimal: 0...2n-1 Of course it gets more complicated: numbers are finite (overflow) fractions and real numbers negative numbers e.g., no MIPS subi instruction; addi can add a negative number) How do we represent negative numbers? i.e., which bit patterns will represent which numbers?

Possible Representations Sign Magnitude: One's Complement Two's Complement 000 = +0 000 = +0 000 = +0 001 = +1 001 = +1 001 = +1 010 = +2 010 = +2 010 = +2 011 = +3 011 = +3 011 = +3 100 = -0 100 = -3 100 = -4 101 = -1 101 = -2 101 = -3 110 = -2 110 = -1 110 = -2 111 = -3 111 = -0 111 = -1 Issues: balance, number of zeros, ease of operations Which one is best? Why?

MIPS 32 bit signed numbers: 0000 0000 0000 0000 0000 0000 0000 0000two = 0ten 0000 0000 0000 0000 0000 0000 0000 0001two = + 1ten 0000 0000 0000 0000 0000 0000 0000 0010two = + 2ten ... 0111 1111 1111 1111 1111 1111 1111 1110two = + 2,147,483,646ten 0111 1111 1111 1111 1111 1111 1111 1111two = + 2,147,483,647ten 1000 0000 0000 0000 0000 0000 0000 0000two = – 2,147,483,648ten 1000 0000 0000 0000 0000 0000 0000 0001two = – 2,147,483,647ten 1000 0000 0000 0000 0000 0000 0000 0010two = – 2,147,483,646ten ... 1111 1111 1111 1111 1111 1111 1111 1101two = – 3ten 1111 1111 1111 1111 1111 1111 1111 1110two = – 2ten 1111 1111 1111 1111 1111 1111 1111 1111two = – 1ten maxint minint

Two's Complement Operations Negating a two's complement number: invert all bits and add 1 Easier rule: Start at least significant bit. Copy through fist “1”. Then invert each bit. Example 0010101100 1101010100 remember: “negate” and “invert” are quite different! Converting n bit numbers into numbers with more than n bits: MIPS 16 bit immediate gets converted to 32 bits for arithmetic copy the most significant bit (the sign bit) into the other bits 0010 -> 0000 0010 1010 -> 1111 1010 "sign extension" (lbu vs. lb) Load byte lbu rt,ofst(rs) # rt  024||MB(rs + ofst) or rt  MB(rs + ofst) lb rt,ofst(rs) # rt  MB(rs + ofst;7)24||MB(rs + ofst) or rt  MB(rs + ofst)

Addition & Subtraction Just like in grade school (carry/borrow 1s) 0111 ( 7) 0111 ( 7) 0110 ( 6) + 0110 ( 6) - 0110 (-6) - 0101 (-5) Two's complement operations easy subtraction using addition of negative numbers 0111 ( 7) + 1010 (-6) Overflow (result too large for finite computer word): e.g., adding two n-bit numbers does not yield an n-bit number 0111 ( 7) + 0001 ( 1) note that overflow term is somewhat misleading, 1000 (-8) it does not mean a carry “overflowed”

Detecting Overflow No overflow when adding a positive and a negative number No overflow when signs are the same for subtraction Overflow occurs when the value affects the sign: overflow when adding two positives yields a negative or, adding two negatives gives a positive or, subtract a negative from a positive and get a negative or, subtract a positive from a negative and get a positive Consider the operations A + B, and A – B Can overflow occur if B is 0 ? Can overflow occur if A is 0 ?

Two’s Complement Arithmetic Representation of 4 bit words in two’s complement.. Decimal value Binary   Two’s complement -8 -1000 1000 -7 -0111 1001 -6 -0110 1010 -5 -0101 1011 -4 -0100 1100 -3 -0011 1101 -2 -0010 1110 -1 -0001 1111 0000 1 0001 2 0010 3 0011 4 0100 5 0101 6 0110 7 0111

When does overflow occur When does overflow occur? Let S be the sign bit (bit 3) of the result, let Co be the carry out of the sign bit, Let SA and SB be the sign bits of the two numbers added. Then form the previous table we can construct a Karnaough map for overflow. Thus, overflow is given by . Using Co and Cin, by inspection Ov = Cin  Co

Effects of Overflow An exception (interrupt) occurs Control jumps to predefined address for exception Interrupted address is saved for possible resumption Details based on software system / language example: flight control vs. homework assignment Don't always want to detect overflow — new MIPS instructions: addu, addiu, subu note: addiu ignores overflow. still sign-extends! note: sltu, sltiu for unsigned comparisons (no sign-extension)

Review: Boolean Algebra & Gates See Appendix B Problem: Consider a logic function with three inputs: A, B, and C. Output D is true if at least one input is true Output E is true if exactly two inputs are true Output F is true only if all three inputs are true Show the truth table for these three functions. Show the Boolean equations for these three functions. Show an implementation consisting of inverters, AND, and OR gates.

Review: The Multiplexor Selects one of the inputs to be the output, based on a control input Lets build our ALU using a MUX: S C A B 1 note: we call this a 2-input mux even though it has 3 inputs!

An ALU (arithmetic logic unit) Let's build an ALU to support the and and or instructions we'll just build a 1 bit ALU, and use 32 of them Possible Implementation (sum-of-products): operation result a b

Different Implementations Not easy to decide the “best” way to build something Don't want too many inputs to a single gate Don't want to have to go through too many gates for our purposes, ease of comprehension is important Let's look at a 1-bit ALU for addition: How could we build a 1-bit ALU for add, and, and or? How could we build a 32-bit ALU? cout = a b + a cin + b cin sum = a xor b xor cin

Building a 32 bit ALU

What about subtraction (a – b) ? Two's complement approach: just negate b and add. How do we negate? A very clever solution:

Tailoring the ALU to the MIPS Need to support the set-on-less-than instruction (slt) remember: slt is an arithmetic instruction produces a 1 if rs < rt and 0 otherwise use subtraction: (a-b) < 0 implies a < b Need to support test for equality (beq $t5, $t6, $t7) use subtraction: (a-b) = 0 implies a = b

Supporting slt Can we figure out the idea? If a < b then a – b < 0 and set i.e. (a-b) Bit 31 = 1 else set = 0.

Test for equality Notice control lines: 000 = and 001 = or 010 = add 110 = subtract 111 = slt Note: zero is a 1 when the result is zero!

Conclusion We can build an ALU to support the MIPS instruction set key idea: use multiplexor to select the output we want we can efficiently perform subtraction using two’s complement we can replicate a 1-bit ALU to produce a 32-bit ALU Important points about hardware all of the gates are always working the speed of a gate is affected by the number of inputs to the gate the speed of a circuit is affected by the number of gates in series (on the “critical path” or the “deepest level of logic”) Our primary focus: comprehension, however, Clever changes to organization can improve performance (similar to using better algorithms in software) we’ll look at two examples for addition and multiplication

Lab 3 – Build modified one bit ALU in VHDL p g You should fully document your code, and in the lab report briefly explain the decisions you made. Include in your lab report simulation waveforms to verify the ALU1 operation, and before 3:00 p.m. on 9/21/09 email to jhtucker@vcu.edu your source code for the lab. Use “EGRE426 Lab 3-your name” as the subject line. The correct simulation waveform is shown below. Turn in your lab report at the beginning of class on 9/22/09.

ARCHITECTURE TEST OF TB IS SIGNAL a, b, cin, cout, g, p: std_logic := '0'; SIGNAL ainv, binv, less, r_and, r_or, r_add, r_less: std_logic := '0'; BEGIN less <= not less after 10 ns; cin <= not cin after 20 ns; a <= not a after 40 ns; b <= not b after 80 ns; binv <= not binv after 160 ns; ainv <= not ainv when (binv'event and binv = '0'); alu1_and: entity alu1 PORT MAP (a,b,ainv,binv,less,"00",cin,R_and); alu1_or: entity alu1 PORT MAP (a,b,ainv,binv,less,"01",cin,R_or); alu1_add: entity alu1 PORT MAP (a,b,ainv,binv,less,"10",cin,R_add, cout,p,g); alu1_less: entity alu1 PORT MAP (a,b,ainv,binv,less,"11",cin,R_less, cout,p,g); process begin wait on less, cin, a, b, binv; -- Wait for input to change wait for 9 ns; -- Let settle assert r_and = (a and b) report "AND error."; assert r_or = (a or b) report "OR error."; assert r_add = ((a xor b) xor cin) report "ADD error."; assert (r_less = less) report "LESS error"; end process; END ARCHITECTURE TEST;

Problem: ripple carry adder is slow Is a 32-bit ALU as fast as a 1-bit ALU? Is there more than one way to do addition? two extremes: ripple carry and sum-of-products Can you see the ripple? How could you get rid of it? c1 = b0c0 + a0c0 + a0b0 c2 = b1c1 + a1c1 + a1b1 c2 = c3 = b2c2 + a2c2 + a2b2 c3 = c4 = b3c3 + a3c3 + a3b3 c4 = Not feasible! Why?

Carry-lookahead adder See B-39 An approach in-between our two extremes Motivation: If we didn't know the value of carry-in, what could we do? When would we always generate a carry? gi = ai bi When would we propagate the carry? pi = ai + bi Did we get rid of the ripple? c1 = g0 + p0c0 c2 = g1 + p1c1 c3 = g2 + p2c2 c4 = g3 + p3c3 Feasible! Why?

Use principle to build bigger adders Can’t build a 16 bit adder this way... (too big) Could use ripple carry of 4-bit CLA adders Better: use the CLA principle again! See p B44-B45

Multiplication More complicated than addition accomplished via shifting and addition More time and more area Let's look at 3 versions based on gradeschool algorithm 0010 (multiplicand) __x_1011 (multiplier) 0010 0010 0000 0010 . 0010110 Negative numbers: convert and multiply there are better techniques, we won’t look at them

Multiplication: Implementation 0010 (multiplicand x_1011 (multiplier) 0010 1011 0010 0101 0000 0010 0010 0001 0010110 Two 64 bit registers, one 32 bit reg, ALU is 64 bits wide.

Second Version One 64 bit reg, two 32 bit reg, 32 bit ALU.

Final Version One 32 bit reg, one 64 bit reg, 32 bit ALU

Faster Multiplication: See Booths algorithm for a faster serial approach. For a fully parallel approach consider: Worst case delay for S5 = 7 full adders.

Carry Save Adders are a faster approach. Worst case delay for S5 = 6 FA’s, faster using one CLA. Other techniques exist.

Floating Point (a brief look) We need a way to represent numbers with fractions, e.g., 3.1416 very small numbers, e.g., .000000001 very large numbers, e.g., 3.15576 ´ 109 Representation: sign, exponent, significand: (–1)sign ´ significand ´ 2exponent more bits for significand gives more accuracy more bits for exponent increases range IEEE 754 floating point standard: single precision: 8 bit exponent, 23 bit significand double precision: 11 bit exponent, 52 bit significand

IEEE 754 floating-point standard Leading “1” bit of significand is implicit Exponent is “biased” to make sorting easier all 0s is smallest exponent all 1s is largest bias of 127 for single precision and 1023 for double precision summary: (–1)sign ´ (1+significand) ´ 2exponent – bias Example: decimal: -.75 = -3/4 = -3/22 binary: -.11 = -1.1 x 2-1 floating point: exponent = 126 = 01111110 IEEE single precision: 1,01111110,10000000000000000000000 1,10000001,0100000000000000000000 = ? Negative, exponent of 129 – 127 = +2, Significan = 1.012 -1.012x22 = -101.02 = -5.0

Floating point addition

Floating Point Complexities Operations are somewhat more complicated (see text) In addition to overflow we can have “underflow” Accuracy can be a big problem IEEE 754 keeps two extra bits, guard and round four rounding modes positive divided by zero yields “infinity” zero divide by zero yields “not a number” other complexities Implementing the standard can be tricky Not using the standard can be even worse see text for description of 80x86 and Pentium bug!

Chapter Three Summary Computer arithmetic is constrained by limited precision Bit patterns have no inherent meaning but standards do exist two’s complement IEEE 754 floating point Computer instructions determine “meaning” of the bit patterns Performance and accuracy are important so there are many complexities in real machines (i.e., algorithms and implementation). We are ready to move on (and implement the processor)