ece/cs 552: introduction to superscalar processors instructor: mikko h lipasti fall 2010 university...

60
ECE/CS 552: ECE/CS 552: Introduction to Introduction to Superscalar Processors Superscalar Processors Instructor: Mikko H Lipasti Fall 2010 University of Wisconsin- Madison Lecture notes partially based on notes by John P. Shen

Upload: marilyn-grant

Post on 21-Dec-2015

217 views

Category:

Documents


1 download

TRANSCRIPT

ECE/CS 552: Introduction to ECE/CS 552: Introduction to Superscalar ProcessorsSuperscalar Processors

Instructor: Mikko H Lipasti

Fall 2010University of Wisconsin-Madison

Lecture notes partially based on notes by John P. Shen

© Shen, Lipasti2

Limitations of Scalar PipelinesLimitations of Scalar Pipelines

Scalar upper bound on throughput– IPC <= 1 or CPI >= 1

Inefficient unified pipeline– Long latency for each instruction

Rigid pipeline stall policy– One stalled instruction stalls all newer

instructions

© Shen, Lipasti3

Parallel PipelinesParallel Pipelines

(a) No Parallelism (b) Temporal Parallelism

(c) Spatial Parallelism

(d) Parallel Pipeline

© Shen, Lipasti4

Intel Pentium Parallel PipelineIntel Pentium Parallel Pipeline

IF

D1

D2

EX

WB

IF IF

D1 D1

D2 D2

EX EX

WB WB

U - Pipe V - Pipe

© Shen, Lipasti5

Diversified PipelinesDiversified Pipelines

• • •

• • •

• • •

• • •IF

ID

RD

WB

ALU MEM1 FP1 BR

MEM2 FP2

FP3

EX

© Shen, Lipasti6

Power4 Diversified PipelinesPower4 Diversified PipelinesPCI-Cache

BR Scan

BR Predict

Fetch Q

Decode

Reorder BufferBR/CRIssue Q

CRUnit

BRUnit

FX/LD 1Issue Q

FX1Unit LD1

Unit

FX/LD 2Issue Q

LD2Unit

FX2Unit

FPIssue Q

FP1Unit

FP2Unit

StQ

D-Cache

© Shen, Lipasti7

Rigid Pipeline Stall PolicyRigid Pipeline Stall Policy

Bypassing of StalledInstruction

Stalled Instruction

Backward Propagationof Stalling

Not Allowed

© Shen, Lipasti8

Dynamic PipelinesDynamic Pipelines

• • •

• • •

• • •

• • •IF

ID

RD

WB

ALU MEM1 FP1 BR

MEM2 FP2

FP3

EX

DispatchBuffer

ReorderBuffer

( in order )

( out of order )

( out of order )

( in order )

© Shen, Lipasti9

Interstage BuffersInterstage Buffers

• • •

• • •

• • •

Stage i

Buffer (n)

Stage i +1

Stage i

Buffer (> n)

Stage i + 1

n

n

Stage i

Buffer (1)

Stage i + 1

(a) (b)

(c)

• • •

( in order )

( out of order )_

( in order )1

1

• • •

( in order )

© Shen, Lipasti10

Superscalar Pipeline StagesSuperscalar Pipeline Stages

Instruction Buffer

Fetch

Dispatch Buffer

Decode

Issuing Buffer

Dispatch

Completion Buffer

Execute

Store Buffer

Complete

Retire

In Program

Order

In Program

Order

Outof

Order

© Shen, Lipasti11

Limitations of Scalar PipelinesLimitations of Scalar Pipelines

Scalar upper bound on throughput– IPC <= 1 or CPI >= 1– Solution: wide (superscalar) pipeline

Inefficient unified pipeline– Long latency for each instruction– Solution: diversified, specialized pipelines

Rigid pipeline stall policy– One stalled instruction stalls all newer instructions– Solution: Out-of-order execution, distributed

execution pipelines

© Shen, Lipasti12

Impediments to High IPCImpediments to High IPC

I-cache

FETCH

DECODE

COMMIT

D-cache

BranchPredictor Instruction

Buffer

StoreQueue

ReorderBuffer

Integer Floating-point Media Memory

Instruction

RegisterData

MemoryData

Flow

EXECUTE

(ROB)

Flow

Flow

© Shen, Lipasti13

Superscalar Pipeline DesignSuperscalar Pipeline Design

Instruction Fetching IssuesInstruction Decoding IssuesInstruction Dispatching IssuesInstruction Execution IssuesInstruction Completion & Retiring Issues

© Shen, Lipasti14

Instruction FetchInstruction Fetch

Challenges:– Branches: control dependences

– Branch target misalignment

– Instruction cache misses Solutions

– Alignment hardware

– Prediction/speculation

Instruction Memory

PC

3 instructions fetched

Objective: Fetch multiple instructions per cycle

© Shen, Lipasti15

Fetch AlignmentFetch Alignment

© Shen, Lipasti16

Branches – MIPSBranches – MIPS

6 Types of BranchesJump (uncond, no save PC, imm)Jump and link (uncond, save PC, imm)Jump register (uncond, no save PC, register)Jump and link register (uncond, save PC, register)Branch (conditional, no save PC, PC+imm)Branch and link (conditional, save PC, PC+imm)

© Shen, Lipasti17

Disruption of Sequential Control FlowDisruption of Sequential Control Flow

Instruction/Decode Buffer

Fetch

Dispatch Buffer

Decode

Reservation

Dispatch

Reorder/

Store Buffer

Complete

Retire

StationsIssue

Execute

FinishCompletion Buffer

Branch

© Shen, Lipasti18

Branch PredictionBranch Prediction

Target address generation Target Speculation– Access register:

PC, General purpose register, Link register

– Perform calculation: +/- offset, autoincrement

Condition resolution Condition speculation– Access register:

Condition code register, General purpose register

– Perform calculation: Comparison of data register(s)

© Shen, Lipasti19

Target Address GenerationTarget Address Generation

Decode Buffer

Fetch

Dispatch Buffer

Decode

Reservation

Dispatch

Store Buffer

Complete

Retire

StationsIssue

Execute

Finish Completion Buffer

Branch

PC-rel.

Reg.ind.

Reg.ind.withoffset

© Shen, Lipasti20

Condition ResolutionCondition Resolution

Decode Buffer

Fetch

Dispatch Buffer

Decode

Reservation

Dispatch

Store Buffer

Complete

Retire

StationsIssue

Execute

Finish Completion Buffer

Branch

CCreg.

GPreg.valuecomp.

© Shen, Lipasti21

Branch Instruction SpeculationBranch Instruction Speculation

Decode Buffer

Fetch

Dispatch Buffer

Decode

Reservation

Dispatch

StationsIssue

Execute

Finish Completion Buffer

Branch

to I-cache

PC(seq.) = FA (fetch address)PC(seq.)Branch

Predictor(using a BTB)

Spec. target

BTBupdate

Prediction

(target addr.and history)

Spec. cond.

FA-mux

© Shen, Lipasti22

Static Branch PredictionStatic Branch Prediction

Single-direction– Always not-taken: Intel i486

Backwards Taken/Forward Not Taken– Loop-closing branches have negative offset– Used as backup in Pentium Pro, II, III, 4

© Shen, Lipasti23

Static Branch PredictionStatic Branch Prediction

Profile-based1. Instrument program binary

2. Run with representative (?) input set

3. Recompile programa. Annotate branches with hint bits, or

b. Restructure code to match predict not-taken

Performance: 75-80% accuracy– Much higher for “easy” cases

© Shen, Lipasti24

Dynamic Branch PredictionDynamic Branch Prediction

Main advantages:– Learn branch behavior autonomously

No compiler analysis, heuristics, or profiling

– Adapt to changing branch behaviorProgram phase changes branch behavior

First proposed in 1980– US Patent #4,370,711, Branch predictor using

random access memory, James. E. SmithContinually refined since then

© Shen, Lipasti25

Smith Predictor HardwareSmith Predictor Hardware

Jim E. Smith. A Study of Branch Prediction Strategies. International Symposium on Computer Architecture, pages 135-148, May 1981

Widely employed: Intel Pentium, PowerPC 604, PowerPC 620, etc.

Branch Address

Branch Prediction

m

2m k-bit counters

most significant bit

Saturating CounterIncrement/Decrement

Branch Outcome

Updated Counter Value

© Shen, Lipasti26

Two-level Branch PredictionTwo-level Branch Prediction

BHR adds global branch history– Provides more context– Can differentiate multiple instances of the same static branch– Can correlate behavior across multiple static branches

BHR0110

PC = 01011010010101

010110

000000000001000010000011

010100010101010110010111

111110111111

PHT

1 0

1 Branch Prediction

© Shen, Lipasti27

Combining or Hybrid PredictorsCombining or Hybrid Predictors

Select “best” history Reduce interference w/partial updates Scott McFarling. Combining Branch Predictors. TN-36,

Digital Equipment Corporation Western Research Laboratory, June 1993.

Branch Address

P0 P1

Branch Prediction

Meta-Prediction

Branch Address

M

PApgshare

M

Branch Prediction

© Shen, Lipasti28

Branch Target PredictionBranch Target PredictionBranch Address

Branch ...target tag target tag target tag

= = =

OR

Branch Target Buffer

+

Size ofInstruction

Branch Target

BTB Hit?

DirectionPredictor

not-takentarget

taken-target0 1

Partial tags sufficient in BTB

© Shen, Lipasti29

Return Address StackReturn Address Stack

For each call/return pair:– Call: push return address onto hardware stack– Return: pop return address from hardware stack

Bra nch Address

Size ofInstruction

BTB

Target Prediction

ReturnAddress

BTB

Target Predictionis this a return?

Bra nch Address

(a) (b)

© Shen, Lipasti30

Branch SpeculationBranch Speculation

Leading Speculation– Typically done during the Fetch stage– Based on potential branch instruction(s) in the current fetch

group Trailing Confirmation

– Typically done during the Branch Execute stage– Based on the next Branch instruction to finish execution

NT T NT T NT TNT T

NT T NT T

NT T (TAG 1)

(TAG 2)

(TAG 3)

© Shen, Lipasti31

Branch SpeculationBranch Speculation Leading Speculation

1. Tag speculative instructions2. Advance branch and following instructions3. Buffer addresses of speculated branch

instructions Trailing Confirmation

1. When branch resolves, remove/deallocate speculation tag

2. Permit completion of branch and following instructions

© Shen, Lipasti32

Branch SpeculationBranch Speculation

Start new correct path– Must remember the alternate (non-predicted) path

Eliminate incorrect path– Must ensure that the mis-speculated instructions

produce no side effects

NT T NT T NT TNT T

NT T NT T

NT T

(TAG 2)

(TAG 3) (TAG 1)

© Shen, Lipasti33

Mis-speculation RecoveryMis-speculation Recovery Start new correct path

1. Update PC with computed branch target (if predicted NT)

2. Update PC with sequential instruction address (if predicted T)

3. Can begin speculation again at next branch Eliminate incorrect path

1. Use tag(s) to deallocate resources occupied by speculative instructions

2. Invalidate all instructions in the decode and dispatch buffers, as well as those in reservation stations

© Shen, Lipasti34

Summary: Instruction FetchSummary: Instruction Fetch Fetch group alignment Target address generation

– Branch target buffer– Return address stack

Target condition generation– Static prediction– Dynamic prediction

Speculative execution– Tagging/tracking instructions– Recovering from mispredicted branches

© Shen, Lipasti35

Issues in DecodingIssues in Decoding

Primary Tasks– Identify individual instructions (!)– Determine instruction types– Determine dependences between instructions

Two important factors– Instruction set architecture– Pipeline width

© Shen, Lipasti36

Pentium Pro Fetch/DecodePentium Pro Fetch/Decode

© Shen, Lipasti37

Predecoding in the AMD K5Predecoding in the AMD K5

© Shen, Lipasti38

Dependence CheckingDependence Checking

Trailing instructions in fetch group– Check for dependence on leading instructions

Dest Src0 Src1 Dest Src0 Src1 Dest Src0 Src1 Dest Src0 Src1

?= ?= ?= ?= ?= ?=

?= ?= ?= ?=

?= ?=

© Shen, Lipasti39

Instruction Dispatch and IssueInstruction Dispatch and Issue

Parallel pipeline– Centralized instruction fetch– Centralized instruction decode

Diversified pipeline– Distributed instruction execution

© Shen, Lipasti40

Necessity of Instruction DispatchNecessity of Instruction Dispatch

© Shen, Lipasti41

Centralized Reservation StationCentralized Reservation Station

© Shen, Lipasti42

Distributed Reservation StationDistributed Reservation Station

© Shen, Lipasti43

Issues in Instruction ExecutionIssues in Instruction Execution

Parallel execution units– Bypassing is a real challenge

Resolving register data dependences– Want out-of-order instruction execution

Resolving memory data dependences– Want loads to issue as soon as possible

Maintaining precise exceptions– Required by the ISA

© Shen, Lipasti44

Bypass NetworksBypass Networks

O(n2) interconnect from/to FU inputs and outputs Associative tag-match to find operands Solutions (hurt IPC, help cycle time)

– Use RF only (IBM Power4) with no bypass network

– Decompose into clusters (Alpha 21264)

PCI-Cache

BR Scan

BR Predict

Fetch Q

Decode

Reorder BufferBR/CRIssue Q

CRUnit

BRUnit

FX/LD 1Issue Q

FX1Unit LD1

Unit

FX/LD 2Issue Q

LD2Unit

FX2Unit

FPIssue Q

FP1Unit

FP2Unit

StQ

D-Cache

© Shen, Lipasti45

The Big PictureThe Big PictureINSTRUCTION PROCESSING CONSTRAINTS

Resource Contention Code Dependences

Control Dependences Data Dependences

True Dependences

Anti-Dependences Output Dependences

Storage Conflicts

(Structural Dependences)

(RAW)

(WAR) (WAW)

© Shen, Lipasti46

Register Data DependencesRegister Data Dependences

Program data dependences cause hazards– True dependences (RAW)– Antidependences (WAR)– Output dependences (WAW)

When are registers read and written?– Out of program order!– Hence, any/all of these can occur

Solution to all three: register renaming

© Shen, Lipasti47

Register Renaming: WAR/WAWRegister Renaming: WAR/WAW

Widely employed (Core i7, Athlon/Phenom, …) Resolving WAR/WAW:

– Each register write gets unique “rename register”

– Writes are committed in program order at Writeback

– WAR and WAW are not an issue All updates to “architected state” delayed till writeback Writeback stage always later than read stage

– Reorder Buffer (ROB) enforces in-order writeback

Add R3 <= … P32 <= …Sub R4 <= … P33 <= …And R3 <= … P35 <= …

© Shen, Lipasti48

Register Renaming: RAWRegister Renaming: RAW

In order, at dispatch:– Source registers checked to see if “in flight”

Register map table keeps track of this If not in flight, can be read from the register file If in flight, look up “rename register” tag (IOU)

– Then, allocate new register for register write

Add R3 <= R2 + R1 P32 <= P2 + P1Sub R4 <= R3 + R1 P33 <= P32 + P1And R3 <= R4 & R2 P35 <= P33 + P2

© Shen, Lipasti49

Register Renaming: RAWRegister Renaming: RAW

Advance instruction to reservation station– Wait for rename register tag to trigger issue

Reservation station enables out-of-order issue– Newer instructions can bypass stalled

instructions

© Shen, Lipasti50

““Dataflow Engine” for Dynamic ExecutionDataflow Engine” for Dynamic ExecutionDispatch Buffer

Reservation

Dispatch

Complete

Stations

Reorder Buffer

Branch

Reg. File Ren. Reg.

Forwardingresults toRes. Sta. &

AllocateReorderBufferentries

Reg. Write Back

rename

Managed as a queue;Maintains sequential orderof all Instructions in flight

Integer Integer Float.- Load/Point Store

registers

© Shen, Lipasti51

Instruction Processing StepsInstruction Processing Steps•DISPATCH:

•Read operands from Register File (RF) and/or Rename Buffers (RRB)

•Rename destination register and allocate RRF entry

•Allocate Reorder Buffer (ROB) entry

•Advance instruction to appropriate Reservation Station (RS)

•EXECUTE:

•RS entry monitors bus for register Tag(s) to latch in pending operand(s)

•When all operands ready, issue instruction into Functional Unit (FU) and deallocate RS entry (no further stalling in execution pipe)

•When execution finishes, broadcast result to waiting RS entries, RRB entry, and ROB entry

•COMPLETE:

•Update architected register from RRB entry, deallocate RRB entry, and if it is a store instruction, advance it to Store Buffer

•Deallocate ROB entry and instruction is considered architecturally completed

© Shen, Lipasti52

Physical Register FilePhysical Register File

Used in MIPS R10000, Pentium 4, AMD Bulldozer All registers in one place

– Always accessed right before EX stage– No copying to real register file at commit

Fet

ch

Dec

ode

Ren

ame

Issu

eR

F R

ead

Exe

cute

Age

n-D

$

ALU

RF

Writ

eD

$

Loa

d R

F W

rite

Physical Register File

Map TableR0 => P7

R1 => P3

R31 => P39

© Shen, Lipasti53

Managing Physical RegistersManaging Physical Registers

What to do when all physical registers are in use?– Must release them somehow to avoid stalling

– Maintain free list of “unused” physical registers Release when no more uses are possible

– Sufficient: next write commits

Map TableR0 => P7

R1 => P3

R31 => P39

Add R3 <= R2 + R1 P32 <= P2 + P1Sub R4 <= R3 + R1 P33 <= P32 + P1……And R3 <= R4 & R2 P35 <= P33 + P2

Release P32 (previous R3)

when this instruction completes execution

© Shen, Lipasti54

Memory Data Memory Data DependencesDependences WAR/WAW: stores commit in order

– Hazards not possible. Why? RAW: loads must check pending stores

– Store queue keeps track of pending store addresses

– Loads check against these addresses– Similar to register bypass logic– Comparators are 32 or 64 bits wide (address

size) Major source of complexity in modern

designs– Store queue lookup is position-based– What if store address is not yet known?

StoreQueue

Load/Store RS

Agen

Reorder Buffer

Mem

© Shen, Lipasti55

Increasing Memory BandwidthIncreasing Memory BandwidthDispatch Buffer

Dispatch

RS’s

Branch

Reg. File Ren. Reg.

Reg. Write Back

Reorder Buff.

Integer Integer Float.-

Point

Load/

Store

Data Cache

Complete

Retire

Store Buff.

Load/

Store

Missed loads

© Shen, Lipasti56

Issues in Completion/RetirementIssues in Completion/Retirement

Out-of-order execution– ALU instructions

– Load/store instructions In-order completion/retirement

– Precise exceptions Solutions

– Reorder buffer retires instructions in order

– Store queue retires stores in order

– Exceptions can be handled at any instruction boundary by reconstructing state out of ROB/SQ

© Shen, Lipasti57

A Dynamic Superscalar ProcessorA Dynamic Superscalar Processor

© Shen, Lipasti58

Superscalar SummarySuperscalar Summary

I-cache

FETCH

DECODE

COMMIT

D-cache

BranchPredictor Instruction

Buffer

StoreQueue

ReorderBuffer

Integer Floating-point Media Memory

Instruction

RegisterData

MemoryData

Flow

EXECUTE

(ROB)

Flow

Flow

© Shen, Lipasti59

Landscape of Microprocessor Families

0

0.5

1

0 500 1000 1500 2000 2500 3000 3500

Frequency (MHz)

SP

EC

int2

000/

MH

z

Intel-x86

AMD-x86

Power

Itanium

700500

300

100

PIII

P4

Athlon

** Data source www.spec.org

Power4

NWD

900

11001900 SpecINT 2000 1300

1500

Opteron

800 MHz

Extreme

Power 3

Power5

PSC

DTN

1700

Itanium

CPIPathLength

FrequencyePerformanc CPU

[John DeVale & Bryan Black, 2005]

© Shen, Lipasti60

Superscalar SummarySuperscalar Summary

Instruction flow– Branches, jumps, calls: predict target, direction– Fetch alignment– Instruction cache misses

Register data flow– Register renaming: RAW/WAR/WAW

Memory data flow– In-order stores: WAR/WAW– Store queue: RAW– Data cache misses: missed load buffers