+ All Categories
Home > Documents > ©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA 1 Introduction to HiperDispatch...

©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA 1 Introduction to HiperDispatch...

Date post: 02-Jan-2016
Category:
Upload: chad-brown
View: 213 times
Download: 0 times
Share this document with a friend
20
1 ©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpex Introduction to HiperDispatch Management Mode with z10 NCACMG meeting June 11, 2008
Transcript

1©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Introduction toHiperDispatch Management Mode

with z10

NCACMG meetingJune 11, 2008

2©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Highlights of HiperDispatch

•Why HiperDispatch Management Mode?• Hardware cache reload reducing performance

• Large number of active logical processors causing multiprocessor effect

•Redesign of z/OS Dispatcher• Multiple affinity dispatch queues

• Dynamic LPAR weight distribution to logical processors

•Redesign of PR/SM algorithms• Establish affinity nodes of up to 4 processors

• Establish High Polarity logical processors

3©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

z10 HARDWARE WITHIN A BOOK

4©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

EXAMPLE MULTIPLE BOOKSt

5©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

LPARLPAR WEIGHTWEIGHT SHARESHARE #CP#CP #LCP#LCP %CP/LCP%CP/LCP

LPAR1LPAR1 400400 40%40% 4.84.8 55 96%96%

LPAR2LPAR2 400400 40%40% 4.84.8 1212 40%40%

LPAR3LPAR3 150150 15%15% 1.81.8 33 60%60%

LPAR4LPAR4 5050 5%5% 0.60.6 22 30%30%

TOTALTOTAL 10001000 100%100% 12.012.0 2222

CPC has 12 physical processors

Common PR/SM Definition Problem

6©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Common PR/SM Definition Problem

Consider simple PR/SM definition for production LPAR:

• 12 logical processors in one LPAR

• PR/SM gives equal share to each logical processor

Resulting logical processor busy:

• Perhaps 40% per processor if using share

• Causes multiprocessor effect

• Each logical processor has low access to physical processor, which can cause “Short Engine” effect

Solution is counter-intuitive to management!

7©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Brief HiperDispatch overview – z/OS(Normal operation - not MASTER or SYSSTC)

• z/OS knows the total weight allocated to the LPAR

• z/OS periodically provides PR/SM with weight (and corresponding share) of each logical processor

• High polarity logical processors = 100% share of equivalent physical processor (up to 4 logical processors per affinity pool)

• Medium polarity logical processors = less than 100% share of equivalent physical processor, share is divided equally among logical processors in affinity pool (up to 2 logical processors)

• Low polarity logical processors = 0% share of equivalent physical processor

• Dynamic management and adjustment of number of logical processors in each category.

8©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

z/OS AFFINITY DISPATCH QUEUES

9©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Brief HiperDispatch overview – PR/SM

• PR/SM creates affinity nodes of up to four physical processors, and dispatches the logical processors to physical processors in the affinity node

• High polarity logical processors(100% share of physical processor)• Dispatch to physical processor affinity node• Pseudo-dedicated processors PR/SM maintains affinity between

logical and physical

• Medium polarity• Total share less than 1.5 equivalent physical processors• Dispatch to affinity group of physical processors• Logical processors have equal share

• Low polarity • 0% share• PR/SM “parks” these logical processors

10©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

PR/SM AFFINITY NODE – HIGH POLARITY PROCESSORS

11©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

HiperDispatch Benefits

• Optimizes reuse of hardware cache

• L1 and L1.5 cache reused by dispatch to same processor

• L2 cache reused by dispatch to same book

• Minimize number of logical processors

• Reduce multiprocessor overhead

• Minimize “short engine” effect

• IBM estimates up to 10% improvement in ITR – depends on:

• The number of physical processors

• The number of logical processors assigned to LPAR

• The logical/physical processor ratio

• The size of locality of memory reference (working set)

12©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Performance Considerations

• Serialization of resource use by applications

• Application might wait longer for access to processor

• Performance goals might need to be adjusted

• Repeatability of application execution time

• Impact of HiperDispatch on LPARs not in HiperDispatch Management Mode

• Could have available capacity reduced because of high polarity processors dedicated by HiperDispatch

• Could have skew of capacity to logical processors because of HiperDispatch minimizing logical processors

13©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

Implementation Considerations

• Not for LPARs with low share of CPC capacity

• Minimum 0.5 equivalent physical processor required

• 1.5 equivalent physical processors for High polarity

• Not for LPARs that have 2 or 3 logical processors

• Not much benefit for small z10 environments

• Review IRD specifications

• Review zIIP and zAAP specifications

• Review performance goals and goal importance

14©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

RMF variables for HiperDispatch

SMF70PFL BIT MEANING WHEN SETBIT MEANING WHEN SET

0 Content of SMF70UPI valid0 Content of SMF70UPI valid

1 Group flag1 Group flag

2 Polarization flag. This partition is vertically polarized.2 Polarization flag. This partition is vertically polarized.

That is, HiperDispatch mode is active. The SMF70POW fieldsThat is, HiperDispatch mode is active. The SMF70POW fields

in the logical processor data section are valid for logicalin the logical processor data section are valid for logical

processors of this partition.processors of this partition.

SMF70HHFSMF70HHF BIT MEANING WHEN SETBIT MEANING WHEN SET

0 HiperDispatch supported0 HiperDispatch supported

1 HiperDispatch is active1 HiperDispatch is active

2 HiperDispatch status changed during interval2 HiperDispatch status changed during interval

SMF70POWSMF70POW Weight for the logical processor when HiperDispatch mode is Weight for the logical processor when HiperDispatch mode is active. The value may be the same or different for all shared active. The value may be the same or different for all shared logical processors of the type described by this PR/SM record.logical processors of the type described by this PR/SM record.

SMF70PATSMF70PAT Logical processor parked time.Logical processor parked time.

These variables require APAR OA24074 and APAR OA12774

15©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

CPExpert initial analysis

• CPU access might be denied because of HiperDispatch

• Service class period missed goal

• HiperDispatch minimizes logical processors – fewer “servers”

• Serialization of resource queues

• HiperDispatch could not be activated for LPAR with low weight

• LPAR weight less than 0.5 physical processor

• HiperDispatch could not establish a high polarity processor

• LPAR weight less than 1.5 physical processor

16©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

CPExpert initial analysis

• HiperDispatch might not be appropriate for LPAR

• Small number of logical processors assigned

• Most logical processors were parked

• Potentially inadequate logical processors assigned to LPAR

• 90% of LPAR capacity share used for RMF interval

• No logical processors parked

• Excess capacity on CPC

• # logical processors less than shared processors on CPC

• Potentially inadequate zAAP processors assigned to LPAR

• Potentially inadequate zIIP processors assigned to LPAR

17©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

CPExpert initial analysis

• HiperDispatch was turned off because IRD decreased weight

• New weight yields less than 0.5 physical processor share

• Increase Minimum Weight for IRD

• High polarity processor not established, IRD decreased weight

• New weight yields less than 1.5 physical processor share

• Increase Minimum Weight for IRD

• Information findings

• HiperDispatch was turned back on, IRD increased weight

• High polarity processor established, IRD increased weight

• HiperDispatch was turned on/off by operator action

** = Applies to LPARs not in HiperDispatch Management Mode

18©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

CPExpert initial analysis(LPARs not in HiperDispatch Management Mode)

• LPAR could not use the defined logical processors

• HiperDispatch used pseudo-dedicated physical processors

• Number of “free” physical processors less than logical processors assigned to LPAR

• LPAR’s available capacity is diminished significantly

• LPAR could not use the defined logical zAAP processors

• LPAR could not use the defined logical zIIP processors

• Logical processors in LPAR had skewed share of capacity

• HiperDispatch uses large percent of physical processors

• LPAR cannot use “equal share” access to physical processors

19©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

References• System z9 109 PR/SM Planning Guide (SB10-7041-01) - Chapter 3: Determining the

Characteristics of Logical Partitions

• System z10 PR/SM Planning Guide (SB10-7153-00) - Chapter 3: Determining the Characteristics of Logical Partitions

• z/OS Intelligent Resource Director (SG24-5922) - Chapter 3: How WLM LPAR CPU Management works (Section 3.5: LPAR Weights)

• IBM System z10 Enterprise Class Technical Introduction (SG24-7515-00)

• IBM System z10 Enterprise Class Technical Guide (SG24-7516-00)

• “WLM–Update for z/OS Release 9 and IBM System z10, The Latest and Greatest” (Horst Sinram,WLM Development, IBM Corporation, Boeblingen, Germany), Session 2540, SHARE Orlando 2008

• “What’s New with z/OS Release 10" (John Eells, IBM Poughkeepsie), Session 2838 SHARE Orlando 2008

• WP101229 "z/OS: Planning Considerations for HiperDispatch Mode" (Steve Grabarits, IBM System z Performance)

20©Copyright 2008, Computer Management Sciences, Inc., Hartfield, VA www.cpexpert.com

For more information, please contact

Don DeeseComputer Management Sciences, Inc.634 Lakeview DriveHartfield, VA 23071-3113

Phone: (804) 776-7109Fax: (804) 776-7139email: [email protected]

Visit www.cpexpert.com/papers.htm for draft of “Introduction to HiperDispatch Management Mode with z10"


Recommended