Date post: | 21-Dec-2015 |
Category: |
Documents |
View: | 225 times |
Download: | 0 times |
CS252/KubiatowiczLec 23.1
11/17/99
CS252Graduate Computer Architecture
Lecture 23
I/O Continued
November 17, 1999
Prof. John Kubiatowicz
CS252/KubiatowiczLec 23.2
11/17/99
Review: Disk Device Terminology
Disk Latency = Queuing Time + Seek Time + Rotation Time + Xfer Time + Ctrl Time
Order of magnitude times for 4K byte transfers:Seek: 12 ms or less
Rotate: 4.2 ms @ 7200 rpm = 0.5 rev/(7200 rpm/60m/s) (8.3 ms @ 3600 rpm )Xfer: 1 ms @ 7200 rpm (2 ms @ 3600 rpm)
Ctrl: 2 ms (big variation)
Disk Latency = Queuing Time + (12 + 4.2 + 1 + 2)ms = QT + 19.2msAverage Service Time = 19.2 ms
CS252/KubiatowiczLec 23.3
11/17/99
But: What about queue time?
Or: why nonlinear response
Response time = Queue + Device Service time
100%
ResponseTime (ms)
Throughput (% total BW)
0
100
200
300
0%
Proc
Queue
IOC Device
Metrics: Response Time Throughput
CS252/KubiatowiczLec 23.4
11/17/99
Departure to discuss queueing theory
(On board)
CS252/KubiatowiczLec 23.5
11/17/99
Introduction to Queueing Theory
• More interested in long term, steady state than in startup => Arrivals = Departures
• Little’s Law: Mean number tasks in system = arrival rate x mean reponse time– Observed by many, Little was first to prove
• Applies to any system in equilibrium, as long as nothing in black box is creating or destroying tasks
Arrivals Departures
CS252/KubiatowiczLec 23.6
11/17/99
A Little Queuing Theory: Notation
• Queuing models assume state of equilibrium: input rate = output rate
• Notation: r average number of arriving customers/second
Tser average time to service a customer (tradtionally µ = 1/ Tser )u server utilization (0..1): u = r x Tser (or u = r / Tser )Tq average time/customer in queue Tsys average time/customer in system: Tsys = Tq + Tser
Lq average length of queue: Lq = r x Tq
Lsys average length of system: Lsys = r x Tsys
• Little’s Law: Lengthsystem = rate x Timesystem (Mean number customers = arrival rate x mean service time)
Proc IOC Device
Queue server
System
CS252/KubiatowiczLec 23.7
11/17/99
A Little Queuing Theory
• Service time completions vs. waiting time for a busy server: randomly arriving event joins a queue of arbitrary length when server is busy, otherwise serviced immediately– Unlimited length queues key simplification
• A single server queue: combination of a servicing facility that accomodates 1 customer at a time (server) + waiting area (queue): together called a system
• Server spends a variable amount of time with customers; how do you characterize variability?– Distribution of a random variable: histogram? curve?
Proc IOC Device
Queue server
System
CS252/KubiatowiczLec 23.8
11/17/99
A Little Queuing Theory
• Server spends a variable amount of time with customers– Weighted mean m1 = (f1 x T1 + f2 x T2 +...+ fn x Tn)/F (F=f1 + f2...)– variance = (f1 x T12 + f2 x T22 +...+ fn x Tn2)/F – m12
» Must keep track of unit of measure (100 ms2 vs. 0.1 s2 )– Squared coefficient of variance: C = variance/m12
» Unitless measure (100 ms2 vs. 0.1 s2)
• Exponential distribution C = 1 : most short relative to average, few others long; 90% < 2.3 x average, 63% < average
• Hypoexponential distribution C < 1 : most close to average, C=0.5 => 90% < 2.0 x average, only 57% < average
• Hyperexponential distribution C > 1 : further from average C=2.0 => 90% < 2.8 x average, 69% < average
Proc IOC Device
Queue server
System
Avg.
CS252/KubiatowiczLec 23.9
11/17/99
A Little Queuing Theory: Variable Service Time
• Server spends a variable amount of time with customers– Weighted mean m1 = (f1xT1 + f2xT2 +...+ fnXTn)/F (F=f1+f2+...)– Squared coefficient of variance C
• Disk response times C 1.5 (majority seeks < average)• Yet usually pick C = 1.0 for simplicity• Another useful value is average time
must wait for server to complete task: m1(z)– Not just 1/2 x m1 because doesn’t capture variance– Can derive m1(z) = 1/2 x m1 x (1 + C)– No variance => C= 0 => m1(z) = 1/2 x m1
Proc IOC Device
Queue server
System
CS252/KubiatowiczLec 23.10
11/17/99
A Little Queuing Theory:Average Wait Time
• Calculating average wait time in queue Tq
– If something at server, it takes to complete on average m1(z)– Chance server is busy = u; average delay is u x m1(z)– All customers in line must complete; each avg Tser
Tq = u x m1(z) + Lq x Ts er= 1/2 x u x Tser x (1 + C) + Lq x Ts er
Tq = 1/2 x u x Ts er x (1 + C) + r x Tq x Ts er
Tq = 1/2 x u x Ts er x (1 + C) + u x Tq
Tq x (1 – u) = Ts er x u x (1 + C) /2Tq = Ts er x u x (1 + C) / (2 x (1 – u))
• Notation: r average number of arriving customers/second
Tser average time to service a customeru server utilization (0..1): u = r x Tser
Tq average time/customer in queueLq average length of queue:Lq= r x Tq
CS252/KubiatowiczLec 23.11
11/17/99
A Little Queuing Theory: M/G/1 and M/M/1
• Assumptions so far:– System in equilibrium– Time between two successive arrivals in line are random– Server can start on next customer immediately after prior
finishes– No limit to the queue: works First-In-First-Out
– Afterward, all customers in line must complete; each avg Tser
• Described “memoryless” or Markovian request arrival (M for C=1 exponentially random), General service distribution (no restrictions), 1 server: M/G/1 queue
• When Service times have C = 1, M/M/1 queueTq = Tser x u x (1 + C) /(2 x (1 – u)) = Tser x u / (1 – u)
Tser average time to service a customeru server utilization (0..1): u = r x Tser
Tq average time/customer in queue
CS252/KubiatowiczLec 23.12
11/17/99
A Little Queuing Theory: An Example
• processor sends 10 x 8KB disk I/Os per second, requests & service exponentially distrib., avg. disk service = 20 ms
• On average, how utilized is the disk?– What is the number of requests in the queue?– What is the average time spent in the queue?– What is the average response time for a disk request?
• Notation: r average number of arriving customers/second = 10
Tser average time to service a customer = 20 ms (0.02s)u server utilization (0..1): u = r x Tser= 10/s x .02s = 0.2Tq average time/customer in queue = Tser x u / (1 – u)
= 20 x 0.2/(1-0.2) = 20 x 0.25 = 5 ms (0 .005s)Tsys average time/customer in system: Tsys =Tq +Tser= 25 msLq average length of queue:Lq= r x Tq
= 10/s x .005s = 0.05 requests in queueLsys average # tasks in system: Lsys = r x Tsys = 10/s x .025s = 0.25
CS252/KubiatowiczLec 23.13
11/17/99
A Little Queuing Theory: Another Example
• processor sends 20 x 8KB disk I/Os per sec, requests & service exponentially distrib., avg. disk service = 12 ms
• On average, how utilized is the disk?– What is the number of requests in the queue?– What is the average time a spent in the queue?– What is the average response time for a disk request?
• Notation: r average number of arriving customers/second= 20
Tser average time to service a customer= 12 msu server utilization (0..1): u = r x Tser= 20/s x .012s = 0.24Tq average time/customer in queue = Ts er x u / (1 – u)
= 12 x 0.24/(1-0.24) = 12 x 0.32 = 3.8 msTsys average time/customer in system: Tsys =Tq +Tser= 15.8 msLq average length of queue:Lq= r x Tq
= 20/s x .0038s = 0.076 requests in queue Lsys average # tasks in system : Lsys = r x Tsys = 20/s x .016s = 0.32
CS252/KubiatowiczLec 23.14
11/17/99
A Little Queuing Theory:Yet Another Example
• Suppose processor sends 10 x 8KB disk I/Os per second, squared coef. var.(C) = 1.5, avg. disk service time = 20 ms
• On average, how utilized is the disk?– What is the number of requests in the queue?– What is the average time a spent in the queue?– What is the average response time for a disk request?
• Notation: r average number of arriving customers/second= 10
Tser average time to service a customer= 20 msu server utilization (0..1): u = r x Tser= 10/s x .02s = 0.2Tq average time/customer in queue = Tser x u x (1 + C) /(2 x (1 – u))
= 20 x 0.2(2.5)/2(1 – 0.2) = 20 x 0.32 = 6.25 ms Tsys average time/customer in system: Tsys = Tq +Tser= 26 msLq average length of queue:Lq= r x Tq
= 10/s x .006s = 0.06 requests in queueLsys average # tasks in system :Lsys = r x Tsys = 10/s x .026s = 0.26
CS252/KubiatowiczLec 23.15
11/17/99
Manufacturing Advantages
of Disk Arrays
14”10”5.25”3.5”
3.5”
Disk Array: 1 disk design
Conventional: 4 disk designs
Low End High End
Disk Product Families
CS252/KubiatowiczLec 23.16
11/17/99
Replace Small # of Large Disks with Large # of Small
Disks! (1988 Disks)
Data Capacity
Volume
Power
Data Rate
I/O Rate
MTTF
Cost
IBM 3390 (K)
20 GBytes
97 cu. ft.
3 KW
15 MB/s
600 I/Os/s
250 KHrs
$250K
IBM 3.5" 0061
320 MBytes
0.1 cu. ft.
11 W
1.5 MB/s
55 I/Os/s
50 KHrs
$2K
x70
23 GBytes
11 cu. ft.
1 KW
120 MB/s
3900 IOs/s
??? Hrs
$150K
Disk Arrays have potential for
large data and I/O rates
high MB per cu. ft., high MB per KW
reliability?
CS252/KubiatowiczLec 23.17
11/17/99
Array Reliability
• Reliability of N disks = Reliability of 1 Disk ÷ N
50,000 Hours ÷ 70 disks = 700 hours
Disk system MTTF: Drops from 6 years to 1 month!
• Arrays (without redundancy) too unreliable to be useful!
Hot spares support reconstruction in parallel with access: very high media availability can be achievedHot spares support reconstruction in parallel with access: very high media availability can be achieved
CS252/KubiatowiczLec 23.18
11/17/99
Redundant Arrays of Disks
• Files are "striped" across multiple spindles• Redundancy yields high data availability
Disks will fail
Contents reconstructed from data redundantly stored in the array
Capacity penalty to store it
Bandwidth penalty to update
Mirroring/Shadowing (high capacity cost)
Horizontal Hamming Codes (overkill)
Parity & Reed-Solomon Codes
Failure Prediction (no capacity overhead!)VaxSimPlus — Technique is controversial
Techniques:
CS252/KubiatowiczLec 23.19
11/17/99
Redundant Arrays of DisksRAID 1: Disk Mirroring/Shadowing
• Each disk is fully duplicated onto its "shadow" Very high availability can be achieved
• Bandwidth sacrifice on write: Logical write = two physical writes
• Reads may be optimized
• Most expensive solution: 100% capacity overhead
Targeted for high I/O rate , high availability environments
recoverygroup
CS252/KubiatowiczLec 23.20
11/17/99
Redundant Arrays of Disks RAID 3: Parity Disk
P100100111100110110010011
. . .
logical record 10010011
11001101
10010011
00110000
Striped physicalrecords
• Parity computed across recovery group to protect against hard disk failures 33% capacity cost for parity in this configuration wider arrays reduce capacity costs, decrease expected availability, increase reconstruction time• Arms logically synchronized, spindles rotationally synchronized logically a single high capacity, high transfer rate disk
Targeted for high bandwidth applications: Scientific, Image Processing
CS252/KubiatowiczLec 23.21
11/17/99
Redundant Arrays of Disks RAID 5+: High I/O Rate Parity
A logical writebecomes fourphysical I/Os
Independent writespossible because ofinterleaved parity
Reed-SolomonCodes ("Q") forprotection duringreconstruction
A logical writebecomes fourphysical I/Os
Independent writespossible because ofinterleaved parity
Reed-SolomonCodes ("Q") forprotection duringreconstruction
D0 D1 D2 D3 P
D4 D5 D6 P D7
D8 D9 P D10 D11
D12 P D13 D14 D15
P D16 D17 D18 D19
D20 D21 D22 D23 P
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.Disk Columns
IncreasingLogical
Disk Addresses
Stripe
StripeUnit
Targeted for mixedapplications
CS252/KubiatowiczLec 23.22
11/17/99
Problems of Disk Arrays: Small Writes
D0 D1 D2 D3 PD0'
+
+
D0' D1 D2 D3 P'
newdata
olddata
old parity
XOR
XOR
(1. Read) (2. Read)
(3. Write) (4. Write)
RAID-5: Small Write Algorithm
1 Logical Write = 2 Physical Reads + 2 Physical Writes
CS252/KubiatowiczLec 23.23
11/17/99
Subsystem Organization
hostarray
controller
single boarddisk
controller
single boarddisk
controller
single boarddisk
controller
single boarddisk
controller
hostadapter
manages interfaceto host, DMA
control, buffering,parity logic
physical devicecontrol
often piggy-backedin small format devices
striping software off-loaded from host to array controller
no applications modifications
no reduction of host performance
CS252/KubiatowiczLec 23.24
11/17/99
System Availability: Orthogonal RAIDs
ArrayController
StringController
StringController
StringController
StringController
StringController
StringController
. . .
. . .
. . .
. . .
. . .
. . .
Data Recovery Group: unit of data redundancy
Redundant Support Components: fans, power supplies, controller, cables
End to End Data Integrity: internal parity protected data paths
CS252/KubiatowiczLec 23.25
11/17/99
System-Level Availability
Fully dual redundantI/O Controller I/O Controller
Array Controller Array Controller
. . .
. . .
. . .
. . . . . .
.
.
.RecoveryGroup
Goal: No SinglePoints ofFailure
Goal: No SinglePoints ofFailure
host host
with duplicated paths, higher performance can beobtained when there are no failures
CS252/KubiatowiczLec 23.26
11/17/99
Review: Storage System Issues
• Historical Context of Storage I/O• Secondary and Tertiary Storage Devices• Storage I/O Performance Measures• Processor Interface Issues• A Little Queuing Theory• Redundant Arrarys of Inexpensive Disks (RAID)• I/O Buses• ABCs of UNIX File Systems• I/O Benchmarks• Comparing UNIX File System Performance
CS252/KubiatowiczLec 23.27
11/17/99
CS 252 Administrivia• Upcoming schedule of project events in CS 252
– Friday Nov 12: finish I/O? Start multiprocessing/networking– Remaining 3 lectures before Thanksgiving: multiprocessing– Wednesday Dec 1: Midterm I – Friday Dec 3: Esoteric computation.
» Quantum/DNA/Nano computing
– Next week: Midproject meetings. Tuesday? (Sharad?)
– Tue/Wed Dec 7/8 for oral reports?– Friday Dec 10: project reports due.
Get moving!!!
CS252/KubiatowiczLec 23.28
11/17/99
A Bus Is:• shared communication link• single set of wires used to connect
multiple subsystems
• A Bus is also a fundamental tool for composing large, complex systems– systematic means of abstraction
Control
Datapath
Memory
ProcessorInput
Output
What is a bus?
CS252/KubiatowiczLec 23.29
11/17/99
Buses
CS252/KubiatowiczLec 23.30
11/17/99
• Versatility:– New devices can be added easily– Peripherals can be moved between computer
systems that use the same bus standard
• Low Cost:– A single set of wires is shared in multiple ways
MemoryProcesser
I/O Device
I/O Device
I/O Device
Advantages of Buses
CS252/KubiatowiczLec 23.31
11/17/99
• It creates a communication bottleneck– The bandwidth of that bus can limit the maximum I/O
throughput
• The maximum bus speed is largely limited by:– The length of the bus– The number of devices on the bus– The need to support a range of devices with:
» Widely varying latencies » Widely varying data transfer rates
MemoryProcesser
I/O Device
I/O Device
I/O Device
Disadvantage of Buses
CS252/KubiatowiczLec 23.32
11/17/99
• Control lines:– Signal requests and acknowledgments– Indicate what type of information is on the data lines
• Data lines carry information between the source and the destination:– Data and Addresses– Complex commands
Data Lines
Control Lines
General Organization of a Bus
CS252/KubiatowiczLec 23.33
11/17/99
• A bus transaction includes two parts:– Issuing the command (and address) – request– Transferring the data – action
• Master is the one who starts the bus transaction by:– issuing the command (and address)
• Slave is the one who responds to the address by:– Sending data to the master if the master ask for data– Receiving data from the master if the master wants to
send data
BusMaster
BusSlave
Master issues command
Data can go either way
Master versus Slave
CS252/KubiatowiczLec 23.34
11/17/99
DMA (Direct Memory Access)?• Typical I/O devices must transfer large
amounts of data to memory of processor:– Disk must transfer complete block (4K? 16K?)– Large packets from network– Regions of frame buffer
• DMA gives external device ability to write memory directly: much lower overhead than having processor request one word at a time.– Processor (or at least memory system) acts like slave
• Issue: Cache coherence:– What if I/O devices write data that is currently in
processor Cache? » The processor may never see new data!
– Solutions: » Flush cache on every I/O operation (expensive)» Have hardware invalidate cache lines (remember
“Coherence” cache misses?)
CS252/KubiatowiczLec 23.35
11/17/99
Types of Buses• Processor-Memory Bus (design specific)
– Short and high speed– Only need to match the memory system
» Maximize memory-to-processor bandwidth– Connects directly to the processor– Optimized for cache block transfers
• I/O Bus (industry standard)– Usually is lengthy and slower– Need to match a wide range of I/O devices– Connects to the processor-memory bus or backplane bus
• Backplane Bus (standard or proprietary)– Backplane: an interconnection structure within the chassis– Allow processors, memory, and I/O devices to coexist– Cost advantage: one bus for all components
CS252/KubiatowiczLec 23.36
11/17/99
Processor/MemoryBus
PCI Bus
I/O Busses
Example: Pentium System Organization
CS252/KubiatowiczLec 23.37
11/17/99
A Computer System with One Bus:
Backplane Bus
• A single bus (the backplane bus) is used for:– Processor to memory communication– Communication between I/O devices and memory
• Advantages: Simple and low cost• Disadvantages: slow and the bus can
become a major bottleneck• Example: IBM PC - AT
Processor Memory
I/O Devices
Backplane Bus
CS252/KubiatowiczLec 23.38
11/17/99
A Two-Bus System
• I/O buses tap into the processor-memory bus via bus adaptors:– Processor-memory bus: mainly for processor-memory traffic– I/O buses: provide expansion slots for I/O devices
• Apple Macintosh-II– NuBus: Processor, memory, and a few selected I/O devices– SCCI Bus: the rest of the I/O devices
Processor Memory
I/OBus
Processor Memory Bus
BusAdaptor
BusAdaptor
BusAdaptor
I/OBus
I/OBus
CS252/KubiatowiczLec 23.39
11/17/99
A Three-Bus System
• A small number of backplane buses tap into the processor-memory bus– Processor-memory bus is only used for processor-memory
traffic– I/O buses are connected to the backplane bus
• Advantage: loading on the processor bus is greatly reduced
Processor Memory
Processor Memory Bus
BusAdaptor
BusAdaptor
BusAdaptor
I/O BusBackplane Bus
I/O Bus
CS252/KubiatowiczLec 23.40
11/17/99
North/South Bridge architectures: separate buses
• Separate sets of pins for different functions– Memory bus – Caches– Graphics bus (for fast frame buffer)– I/O buses are connected to the backplane bus
• Advantage: – Buses can run at different speeds– Much less overall loading!
Processor MemoryProcessor Memory Bus
BusAdaptor
BusAdaptor
I/O BusBackplane Bus
I/O Bus
“backsidecache”
CS252/KubiatowiczLec 23.41
11/17/99
Bunch of Wires
Physical / Mechanical Characterisics – the connectors
Electrical Specification
Timing and Signaling Specification
Transaction Protocol
What defines a bus?
CS252/KubiatowiczLec 23.42
11/17/99
•Synchronous Bus:– Includes a clock in the control lines– A fixed protocol for communication that is relative to the
clock– Advantage: involves very little logic and can run very fast– Disadvantages:
» Every device on the bus must run at the same clock rate
» To avoid clock skew, they cannot be long if they are fast
•Asynchronous Bus:– It is not clocked– It can accommodate a wide range of devices– It can be lengthened without worrying about clock skew– It requires a handshaking protocol
Synchronous and Asynchronous Bus
CS252/KubiatowiczLec 23.43
11/17/99
° ° °Master Slave
Control LinesAddress LinesData Lines
Bus Master: has ability to control the bus, initiates transaction
Bus Slave: module activated by the transaction
Bus Communication Protocol: specification of sequence of events and timing requirements in transferring information.
Asynchronous Bus Transfers: control lines (req, ack) serve to orchestrate sequencing.
Synchronous Bus Transfers: sequence relative to common clock.
Busses so far
CS252/KubiatowiczLec 23.44
11/17/99
Bus Transaction
• Arbitration: Who gets the bus• Request: What do we want to do• Action: What happens in response
CS252/KubiatowiczLec 23.45
11/17/99
• One of the most important issues in bus design:– How is the bus reserved by a device that wishes to use it?
• Chaos is avoided by a master-slave arrangement:– Only the bus master can control access to the bus:
It initiates and controls all bus requests– A slave responds to read and write requests
• The simplest system:– Processor is the only bus master– All bus requests must be controlled by the processor– Major drawback: the processor is involved in every
transaction
BusMaster
BusSlave
Control: Master initiates requests
Data can go either way
Arbitration: Obtaining Access to the Bus
CS252/KubiatowiczLec 23.46
11/17/99
Multiple Potential Bus Masters:
the Need for Arbitration• Bus arbitration scheme:– A bus master wanting to use the bus asserts the bus request– A bus master cannot use the bus until its request is granted– A bus master must signal to the arbiter after finish using the bus
• Bus arbitration schemes usually try to balance two factors:– Bus priority: the highest priority device should be serviced first– Fairness: Even the lowest priority device should never
be completely locked out from the bus
• Bus arbitration schemes can be divided into four broad classes:– Daisy chain arbitration– Centralized, parallel arbitration– Distributed arbitration by self-selection: each device wanting the bus
places a code indicating its identity on the bus.– Distributed arbitration by collision detection:
Each device just “goes for it”. Problems found after the fact.
CS252/KubiatowiczLec 23.47
11/17/99
The Daisy Chain Bus
Arbitrations Scheme
• Advantage: simple• Disadvantages:
– Cannot assure fairness: A low-priority device may be locked out indefinitely
– The use of the daisy chain grant signal also limits the bus speed
BusArbiter
Device 1HighestPriority
Device NLowestPriority
Device 2
Grant Grant Grant
Release
Request
wired-OR
CS252/KubiatowiczLec 23.48
11/17/99
• Used in essentially all processor-memory busses and in high-speed I/O busses
BusArbiter
Device 1 Device NDevice 2
Grant Req
Centralized Parallel Arbitration
CS252/KubiatowiczLec 23.49
11/17/99
• All agents operate synchronously• All can source / sink data at same rate• => simple protocol
– just manage the source and target
Simplest bus paradigm
CS252/KubiatowiczLec 23.50
11/17/99
• Even memory busses are more complex than this– memory (slave) may take time to respond– it may need to control data rate
BReq
BG
Cmd+AddrR/WAddress
Data1 Data2Data
Simple Synchronous Protocol
CS252/KubiatowiczLec 23.51
11/17/99
• Slave indicates when it is prepared for data xfer
• Actual transfer goes at bus rate
BReq
BG
Cmd+AddrR/WAddress
Data1 Data2Data Data1
Wait
Typical Synchronous Protocol
CS252/KubiatowiczLec 23.52
11/17/99
• Separate versus multiplexed address and data lines:– Address and data can be transmitted in one bus cycle
if separate address and data lines are available– Cost: (a) more bus lines, (b) increased complexity
• Data bus width:– By increasing the width of the data bus, transfers of multiple words
require fewer bus cycles– Example: SPARCstation 20’s memory bus is 128 bit wide– Cost: more bus lines
• Block transfers:– Allow the bus to transfer multiple words in back-to-back bus cycles– Only one address needs to be sent at the beginning– The bus is not released until the last word is transferred– Cost: (a) increased complexity
(b) decreased response time for request
Increasing the Bus Bandwidth
CS252/KubiatowiczLec 23.53
11/17/99
• Overlapped arbitration– perform arbitration for next transaction during current
transaction
• Bus parking– master can holds onto bus and performs multiple
transactions as long as no other master makes request
• Overlapped address / data phases (prev. slide)– requires one of the above techniques
• Split-phase (or packet switched) bus– completely separate address and data phases– arbitrate separately for each– address phase yield a tag which is matched with data phase
• ”All of the above” in most modern memory buses
Increasing Transaction Rate on Multimaster Bus
CS252/KubiatowiczLec 23.54
11/17/99
Bus MBus Summit Challenge XDBus
Originator Sun HP SGI Sun
Clock Rate (MHz) 40 60 48 66
Address lines 36 48 40 muxed
Data lines 64 128 256 144 (parity)
Data Sizes (bits) 256 512 1024 512
Clocks/transfer 4 5 4?
Peak (MB/s) 320(80) 960 1200 1056
Master Multi Multi Multi Multi
Arbitration Central Central Central Central
Slots 16 9 10
Busses/system 1 1 1 2
Length 13 inches 12? inches 17 inches
1993 MP Server Memory Bus Survey: GTL revolution
CS252/KubiatowiczLec 23.55
11/17/99
Address
Data
Read
Req
Ack
Master Asserts Address
Master Asserts Data
Next Address
Write Transaction
t0 t1 t2 t3 t4 t5
• t0 : Master has obtained control and asserts address, direction, data
• Waits a specified amount of time for slaves to decode target
• t1: Master asserts request line• t2: Slave asserts ack, indicating data received• t3: Master releases req• t4: Slave releases ack
Asynchronous Handshake
CS252/KubiatowiczLec 23.56
11/17/99
Address
Data
Read
Req
Ack
Master Asserts Address Next Address
t0 t1 t2 t3 t4 t5
• t0 : Master has obtained control and asserts address, direction, data
• Waits a specified amount of time for slaves to decode target\
• t1: Master asserts request line• t2: Slave asserts ack, indicating ready to transmit data• t3: Master releases req, data received• t4: Slave releases ack
Read Transaction
Slave Data
CS252/KubiatowiczLec 23.57
11/17/99
Bus SBus TurboChannelMicroChannel PCI
Originator Sun DEC IBM Intel
Clock Rate (MHz) 16-25 12.5-25 async 33
Addressing Virtual Physical Physical Physical
Data Sizes (bits) 8,16,32 8,16,24,32 8,16,24,32,648,16,24,32,64
Master Multi Single Multi Multi
Arbitration Central Central Central Central
32 bit read (MB/s) 33 25 2033
Peak (MB/s) 89 84 75 111 (222)
Max Power (W) 16 26 13 25
1993 Backplane/IO Bus Survey
CS252/KubiatowiczLec 23.58
11/17/99
• Examples– graphics– fast networks
• Limited number of devices• Data transfer bursts at full rate• DMA transfers important
– small controller spools stream of bytes to or from memory
• Either side may need to squelch transfer– buffers fill up
High Speed I/O Bus
CS252/KubiatowiczLec 23.59
11/17/99
• All signals sampled on rising edge• Centralized Parallel Arbitration
– overlapped with previous transaction
• All transfers are (unlimited) bursts• Address phase starts by asserting
FRAME#• Next cycle “initiator” asserts cmd and
address• Data transfers happen on when
– IRDY# asserted by master when ready to transfer data
– TRDY# asserted by target when ready to transfer data
– transfer when both asserted on rising edge
• FRAME# deasserted when master intends to complete only one more data transfer
PCI Read/Write Transactions
CS252/KubiatowiczLec 23.60
11/17/99
– Turn-around cycle on any signal driven by more than one agent
PCI Read Transaction
CS252/KubiatowiczLec 23.61
11/17/99
PCI Write Transaction
CS252/KubiatowiczLec 23.62
11/17/99
• Push bus efficiency toward 100% under common simple usage– like RISC
• Bus Parking– retain bus grant for previous master until another makes
request– granted master can start next transfer without arbitration
• Arbitrary Burst length– initiator and target can exert flow control with xRDY– target can disconnect request with STOP (abort or retry)– master can disconnect by deasserting FRAME– arbiter can disconnect by deasserting GNT
• Delayed (pended, split-phase) transactions– free the bus after request to slow device
PCI Optimizations
CS252/KubiatowiczLec 23.63
11/17/99
Summary• Buses are an important technique for
building large-scale systems– Their speed is critically dependent on factors such as
length, number of devices, etc.– Critically limited by capacitance– Tricks: esoteric drive technology such as GTL
• Important terminology:– Master: The device that can initiate new transactions– Slaves: Devices that respond to the master
• Two types of bus timing:– Synchronous: bus includes clock– Asynchronous: no clock, just REQ/ACK strobing
• Direct Memory Access (dma) allows fast, burst transfer into processor’s memory:– Processor’s memory acts like a slave– Probably requires some form of cache-coherence so that
DMA’ed memory can be invalidated from cache.
CS252/KubiatowiczLec 23.64
11/17/99
Summary of Bus Options
Option High performance Low cost
Bus width Separate address Multiplex address& data lines & data lines
Data width Wider is faster Narrower is cheaper (e.g., 32 bits)
(e.g., 8 bits)
Transfer size Multiple words hasSingle-word transfer less bus overhead is simpler
Bus masters Multiple Single master(requires arbitration)
(no arbitration)
Clocking Synchronous Asynchronous
CS252/KubiatowiczLec 23.65
11/17/99
Processor Interface Issues
• Processor interface– Interrupts– Memory mapped I/O
• I/O Control Structures– Polling– Interrupts– DMA– I/O Controllers– I/O Processors
• Capacity, Access Time, Bandwidth• Interconnections
– Busses
CS252/KubiatowiczLec 23.66
11/17/99
I/O Interface
Independent I/O Bus
CPU
Interface Interface
Peripheral Peripheral
Memory
memorybus
Seperate I/O instructions (in,out)
CPU
Interface Interface
Peripheral Peripheral
Memory
Lines distinguish between I/O and memory transferscommon memory
& I/O bus
VME busMultibus-IINubus
40 Mbytes/secoptimistically
10 MIP processorcompletelysaturates the bus!
CS252/KubiatowiczLec 23.67
11/17/99
Memory Mapped I/O
Single Memory & I/O Bus No Separate I/O Instructions
CPU
Interface Interface
Peripheral Peripheral
Memory
ROM
RAM
I/O$
CPU
L2 $
Memory Bus
Memory Bus Adaptor
I/O bus
CS252/KubiatowiczLec 23.68
11/17/99
Programmed I/O (Polling)
CPU
IOC
device
Memory
Is thedata
ready?
readdata
storedata
yesno
done? no
yes
busy wait loopnot an efficient
way to use the CPUunless the device
is very fast!
but checks for I/O completion can bedispersed amongcomputationallyintensive code
CS252/KubiatowiczLec 23.69
11/17/99
Interrupt Driven Data TransferCPU
IOC
device
Memory
addsubandornop
readstore...rti
memory
userprogram(1) I/O
interrupt
(2) save PC
(3) interruptservice addr
interruptserviceroutine(4)
Device xfer rate = 10 MBytes/sec => 0 .1 x 10 sec/byte => 0.1 µsec/byte => 1000 bytes = 100 µsec 1000 transfers x 100 µsecs = 100 ms = 0.1 CPU seconds
-6
User program progress only halted during actual transfer
1000 transfers at 1 ms each: 1000 interrupts @ 2 µsec per interrupt 1000 interrupt service @ 98 µsec each = 0.1 CPU seconds
Still far from device transfer rate! 1/2 in interrupt overhead
CS252/KubiatowiczLec 23.70
11/17/99
Direct Memory Access
CPU
IOC
device
Memory DMAC
Time to do 1000 xfers at 1 msec each:1 DMA set-up sequence @ 50 µsec1 interrupt @ 2 µsec1 interrupt service sequence @ 48 µsec
.0001 second of CPU time
CPU sends a starting address, direction, and length count to DMAC. Then issues "start".
DMAC provides handshake signals for PeripheralController, and Memory Addresses and handshakesignals for Memory.
0ROM
RAM
Peripherals
DMACn
Memory Mapped I/O
CS252/KubiatowiczLec 23.71
11/17/99
Input/Output Processors
CPU IOP
Mem
D1
D2
Dn
. . .main memory
bus
I/Obus
CPU
IOP
issues instruction to IOP
interrupts when done(1)
memory
(2)
(3)
(4)
Device to/from memorytransfers are controlledby the IOP directly.
IOP steals memory cycles.
OP Device Address
target devicewhere cmnds are
looks in memory for commands
OP Addr Cnt Other
whatto do
whereto putdata
howmuch
specialrequests
CS252/KubiatowiczLec 23.72
11/17/99
Relationship to Processor Architecture
• I/O instructions have largely disappeared• Interrupt vectors have been replaced by
jump tablesPC <- M [ IVA + interrupt number ]PC <- IVA + interrupt number
• Interrupts:– Stack replaced by shadow registers– Handler saves registers and re-enables higher priority
int's– Interrupt types reduced in number; handler must query
interrupt controller
CS252/KubiatowiczLec 23.73
11/17/99
Relationship to Processor Architecture
• Caches required for processor performance cause problems for I/O– Flushing is expensive, I/O polutes cache– Solution is borrowed from shared memory
multiprocessors "snooping"
• Virtual memory frustrates DMA• Load/store architecture at odds with atomic
operations– load locked, store conditional
• Stateful processors hard to context switch
CS252/KubiatowiczLec 23.74
11/17/99
Summary• Disk industry growing rapidly, improves:
– bandwidth 40%/yr , – areal density 60%/year, $/MB faster?
• queue + controller + seek + rotate + transfer• Advertised average seek time benchmark much
greater than average seek time in practice• Response time vs. Bandwidth tradeoffs• Queueing theory: or
• Value of faster response time:– 0.7sec off response saves 4.9 sec and 2.0 sec (70%) total time
per transaction => greater productivity– everyone gets more done with faster response,
but novice with fast response = expert with slow
1
121
xCW
1
xW
CS252/KubiatowiczLec 23.75
11/17/99
Summary: Relationship to Processor Architecture
• I/O instructions have disappeared• Interrupt vectors have been replaced by jump
tables• Interrupt stack replaced by shadow registers• Interrupt types reduced in number• Caches required for processor performance
cause problems for I/O• Virtual memory frustrates DMA• Load/store architecture at odds with atomic
operations• Stateful processors hard to context switch