SlideShare a Scribd company logo
1 of 31
Download to read offline
Technical University/Symposia materials may not be reproduced in whole or in part without the prior written permission of IBM.
9.0
© Copyright IBM Corporation 2015
Parallel Sysplex Performance Topics
Session z011601
Martin Packer
IBM
Abstract	
Over recent years RMF's Parallel Sysplex instrumentation has improved
tremendously.
For example, Coupling Facility CPU reporting was enhanced, to give more
granularity. And Coupling Facility links were reported on much better. Also RMF
support of XCF was enhanced.
This presentation outlines my experience with this important new
instrumentation, from a number of perspectives.
Topics	
§ Structure-Level CPU	
§ CPU / LPAR Match Up Between 70-1 and 74-4	
§ Structure Duplexing	
§ XCF Traffic
§ Coupling Facility Link Information
§ Thin Interrupts Instrumentation
§ Conclusions and Musings
Structure-Level CPU
Structure-Level CPU Consumption	
§  SMF 74-4 Field: R744SETM
– “Structure Execution Time”
§ Always 100% Capture Ratio
– Adds up to R744PBSY
§ Multiple uses:
– Capacity planning for changing request rates
– Examine which structures are large consumers
– Compute CPU cost of a request
• And compare to service time
• Interesting number is “non-CPU” element of service time - as we shall see
– Understand whether CPU per request has degraded
– Estimating Structure Duplexing cost
NOTE: Need to collect 74-4 data from all z/OS systems sharing to get total request rate
- Otherwise “CPU per request” calculation will overestimate
CPU By Structure – For Capacity Planning
Structure CPU Experiment	
§ All requests were Sync	
§ One minute RMF intervals	
– Sorted by request rate increasing	
§ Run was 1-way DB2 Datasharing	
– Only really active structures ISGLOCK and LOCK1
§ Red lines are CPU time per request	
§ Blue lines are Service time per request
§ ISGLOCK case: “low volume”	
– Shows amortization of some fixed cost effect	
– CF used IC links
§ LOCK1 case: “high volume”	
– More reliable for capacity planning	
– CF used a mixture of ISC and ICB links
ISGLOCK Requests
0
2
4
6
8
10
12
14
16
0 10 20 30 40 50 60 70
Requests / Second
Microseconds
CPU Time Service Time
3us?
LOCK1 Requests
0
2
4
6
8
10
12
750 800 850 900
Requests / Second
Microseconds
CPU Time Service Time
3.5us?
And From My Travels...	
§ Next chart isn't from the experiment just described
– A real customer system
– An old case so numbers old
§ A Group Buffer Pool
§ Not duplexed
§ ISC-Connected
– Necessary for the customer's estate
§ Clearly something goes wrong at about 1100 requests / second
– Especially in response time terms but also CPU
• (Coupling Facility not CPU constrained)
§ Options include
– Managing the request rate to below 1100 / sec
– Working on the request mix
– Infrastructure reconfiguration
25us?
CPU / LPAR Match Up Between
70-1 and 74-4
λ Managed out of Pool 5 in modern processor families
λ Pool numbers given in SMF 70 as index into table of labels
λ Recommendation: Manage in reporting as a separate pool
λ Follow special CF sizing guidelines
λ Especially for takeover situations
λ Always runs at full speed
λ So good technology match for coupled z/OS images on same footprint
λ Another good reason to use ICFs is IC links
λ Shared ICFs strongly discouraged for Production
λ Especially if the CF image has Dynamic Dispatch turned on
λ Unconvinced Thin Interrupts negates this advice
λ Should not run ANY coupling facility above 50% busy
λ Especially if we need to be able to recover structures onto it
Internal Coupling Facility - Basics
ICF CPU Instrumentation	
§ SMF 74-4 view different from SMF 70-1 LPAR view of processor busy
• R744PBSY is CPU time processing requests
• R744PWAI is CPU time while CFCC is not processing requests but it is still using CF
cycles
• For Dynamic Dispatch PWAI is time when not processing CF requests but Logical
CP not yet taken back by PR/SM
• CF Thin Interrupts will reduce this
• For dedicated or non-Dynamic Dispatch cases sum is constant
• For Dynamic Dispatch sum can vary.
§ Number of defined processors is number of CF Processor Data sections in 74-4
• Fields for dedicated (R744FPDN) and shared (R744FPSN) processors
• Also whether individual engine is dedicated (R744PTYP) and its weight (R744PWGT)
§ PBSY and PWAI Can be examined down to Coupling Facility engine level
CF LPAR Identification In SMF 70-1 Is Was Complex	
§ Need to match LPARs in SMF 70-1 with coupling facilities
in SMF 74-4 to get proper CPU picture
§ 74-4 has machine serial number
– Allows correlation in most cases
• But LPAR names and CF Names often don't match
• Often multiple CF's in same footprint with similar configuration
• Sometimes there are multiple CF's with the same name
• My code – in extremis – uses the presence of IC links to
determine “colocality”
• [I'm slowly learning :-) not all CF LPARs are in Pool 5]
Additional Instrumentation - OA21140	
§ Everyone has this support
§ Ancient APAR integrated into recent z/OS releases
§ Introduced to support zHPF
– Has other SMF and reporting improvements
• HiperDispatch Vertical Polarisation indicators at ENGINE level
– Type 70
• Normalisation factor for zIIP – Type 70
§ Adds CF LPAR Partition Number
– Allows matching with SMF 70-1
§ RMF Level (SMFxxSRL) changed to X'55'
Structure Duplexing
Structure Duplexing Performance	
§ Additional Traffic	
– For lock structures duplexing generates double the traffic	
– Otherwise only the writes are duplicated	
– Additional CPU cost	
§ Additional Physical Resources	
– A second coupling facility	
• Documented in 74-4	
– Additional memory – but “white space” rules say “not really”	
– Additional links – to second coupling facility and between it and the
primary	
• Documented in SMF 74-4
SMD = System-Managed Duplexing
UMD = User-Managed Duplexing
Structure Duplexing Performance - Response Times	
§ For SMD structures both requests must complete	
– Response time is that of the slowest	
• So all requests are essentially with “remote” response times	
• High likelihood of requests becoming asynchronous	
• For low contention rates applications might experience longer lock acquisition
times	
§ For UMD structures both requests must complete	
– But only for writes	
– So writes performed with “remote” response times	
– With high a read-to-write ratio request response times might not be
significantly extended	
– Only example: DB2 Group Buffer Pools
	
§ Response time elongation measured by RMF PR WT and PR CMP times	
– Former suggests better link infrastructure	
– Latter suggests a more capable peer coupling facility
XCF Traffic
XCF – Groups Worth Looking At	
§ 74-2 has “job name” as well as member name
§ You can now answer more detailed questions about traffic:	
– For most XCF groups member name is useless	
§ Traffic valuable for three reasons:	
– Explains links, structure, buffering etc demand	
– You can look better at eg DB2 IRLM Global Lock tuning
– You can see topology without special instrumentation
– For example, DB2 IRLM address spaces
– For example, CICS
– I've never seen a customer using other than DFHIR000 group
Coupling Facility Link Informa<on
Coupling Facility Path Information	
§ Dramatically improved in CFLEVEL 18 (zEC12)	
– RMF APAR OA37826	
• SMF 74-4	
• Coupling Facility Activity Report	
– Configuration:	
• Detailed adapter and link type, PCHID, CHPID	
– OA37826 gives CHPID even without CFLEVEL 18	
• Infiniband, ISC, and ICA-SR only	
– Performance:	
• “Degraded” flag
• If this flag set then call your Customer Engineer	
• Channel Path Latency Time (R744HLAT)	
– Divide by 10 us to give distance estimate in Postprocessor Report	
– Would be interesting if it degraded (as it shouldn’t)
R744HOPM - Channel path operation mode	
Value Meaning	
X'01' CFP path supporting a 1.0625 Gbit/s data rate	
X'02' CFP path supporting a 2.125 Gbit/s data rate	
X'10' CIB path operating at 1x bandwidth using the IFB protocol, adapter type HCA2-O LR	
X'11' CIB path operating at 12x bandwidth using the IFB protocol, adapter type HCA2-O	
X'20' CIB path operating at 1x bandwidth using the IFB protocol, adapter type HCA3-O LR	
X'21' CIB path operating at 12x bandwidth using the IFB protocol,	
adapter type HCA3-O	
X'30' CIB path operating at 12x bandwidth using the IFB3 protocol,	
adapter type HCA3-O
X’40’ CS5 path operating at 8x bandwidth using the PCIe third generation protocol,
adapter type PCIe-O ßICA-SR
I’ve Blogged On This Subject Numerous Times
•  System	zEC12	CFLEVEL	18	RMF	Instrumenta8on	
Improvements	
•  Coupling	Facility	Topology	Informa8on	-	A	Con8nuing	
Journey	
•  The	Missing	Link?	
•  The	Effect	Of	CF	Structure	Distance	
•  What's	The	Latency,	Kenneth?	
•  What's	The	Latency	Really?	
•  And	Latency	Once	More
Coupling Facility Thin Interrupts Instrumenta<on
•  Logical	engines	acquired	and	released	in	a	more	8mely	fashion	
•  See	Barbara	Weiler	paper:
Coupling	Thin	Interrupts	and	Coupling	Facility	Performance	in	Shared	
Processor	Environments	
•  If	you	have	SMF	74-4	for	a	Shared	CF	Engine	Coupling	Facility	
•  With	OA42682	flag	byte	R744FFLG	Bit	4	is	set	if	DYNDISP=THIN	
•  If	CFLEVEL	(R744FLVL)	>	18	
•  R744PWAI	will	be	reduced,	compared	to	DYNDISP=NO	or	YES	
•  But	many	“shared	engine”	CFs	go	“undocumented”	in	SMF	74-4	
•  For	these	expect	SMF70PDT	-	SMF70EDT	to	be	greater	than	before	
•  PR/SM	playing	a	more	concerted	role
Conclusions and Musings	
§ I think we've come a long way with Coupling Facility CPU	
– Capacity Planning is now down to the structure level	
• But not to the structure-by-system level	
– We can now tie up the Coupling Facility and LPAR views of CPU	
§ I'd encourage you to revisit your Parallel Sysplex reporting	
– Including for all the other aspects we didn't have time for	
§ Structure Duplexing needs particular care	
– A very useful resilience feature that has performance
considerations	
§ XCF Traffic a subject worthy of study
§ Coupling Facility Link Information a useful emergent topic

More Related Content

What's hot

I Know What You Did THIS Summer
I Know What You Did THIS SummerI Know What You Did THIS Summer
I Know What You Did THIS SummerMartin Packer
 
FlashCopy and DB2 for z/OS
FlashCopy and DB2 for z/OSFlashCopy and DB2 for z/OS
FlashCopy and DB2 for z/OSFlorence Dubois
 
DB2 Data Sharing Performance
DB2 Data Sharing PerformanceDB2 Data Sharing Performance
DB2 Data Sharing PerformanceMartin Packer
 
Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)
Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)
Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)Florence Dubois
 
DB2 for z/OS and DASD-based Disaster Recovery - Blowing away the myths
DB2 for z/OS and DASD-based Disaster Recovery - Blowing away the mythsDB2 for z/OS and DASD-based Disaster Recovery - Blowing away the myths
DB2 for z/OS and DASD-based Disaster Recovery - Blowing away the mythsFlorence Dubois
 
DB2 for z/OS - Starter's guide to memory monitoring and control
DB2 for z/OS - Starter's guide to memory monitoring and controlDB2 for z/OS - Starter's guide to memory monitoring and control
DB2 for z/OS - Starter's guide to memory monitoring and controlFlorence Dubois
 
Coupling Facility CPU
Coupling Facility CPUCoupling Facility CPU
Coupling Facility CPUMartin Packer
 
DB2 Data Sharing Performance for Beginners
DB2 Data Sharing Performance for BeginnersDB2 Data Sharing Performance for Beginners
DB2 Data Sharing Performance for BeginnersMartin Packer
 
Even More Fun With DDF
Even More Fun With DDFEven More Fun With DDF
Even More Fun With DDFMartin Packer
 
Educational seminar lessons learned from customer db2 for z os health check...
Educational seminar   lessons learned from customer db2 for z os health check...Educational seminar   lessons learned from customer db2 for z os health check...
Educational seminar lessons learned from customer db2 for z os health check...John Campbell
 
Best practices for DB2 for z/OS log based recovery
Best practices for DB2 for z/OS log based recoveryBest practices for DB2 for z/OS log based recovery
Best practices for DB2 for z/OS log based recoveryFlorence Dubois
 
A First Look at the DB2 10 DSNZPARM Changes
A First Look at the DB2 10 DSNZPARM ChangesA First Look at the DB2 10 DSNZPARM Changes
A First Look at the DB2 10 DSNZPARM ChangesWillie Favero
 
Top 5 performance and capacity challenges for z/OS
Top 5 performance and capacity challenges for z/OS Top 5 performance and capacity challenges for z/OS
Top 5 performance and capacity challenges for z/OS Metron
 
IMS09 ims v14 higlights
IMS09   ims v14 higlightsIMS09   ims v14 higlights
IMS09 ims v14 higlightsRobert Hain
 
Introduction to FlashCopy
Introduction to FlashCopy Introduction to FlashCopy
Introduction to FlashCopy HelpSystems
 
IBM DB2 Analytics Accelerator Trends & Directions by Namik Hrle
IBM DB2 Analytics Accelerator  Trends & Directions by Namik Hrle IBM DB2 Analytics Accelerator  Trends & Directions by Namik Hrle
IBM DB2 Analytics Accelerator Trends & Directions by Namik Hrle Surekha Parekh
 

What's hot (19)

Much Ado about CPU
Much Ado about CPUMuch Ado about CPU
Much Ado about CPU
 
I Know What You Did THIS Summer
I Know What You Did THIS SummerI Know What You Did THIS Summer
I Know What You Did THIS Summer
 
FlashCopy and DB2 for z/OS
FlashCopy and DB2 for z/OSFlashCopy and DB2 for z/OS
FlashCopy and DB2 for z/OS
 
DB2 Data Sharing Performance
DB2 Data Sharing PerformanceDB2 Data Sharing Performance
DB2 Data Sharing Performance
 
Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)
Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)
Db2 for z/OS and FlashCopy - Practical use cases (June 2019 Edition)
 
DB2 for z/OS and DASD-based Disaster Recovery - Blowing away the myths
DB2 for z/OS and DASD-based Disaster Recovery - Blowing away the mythsDB2 for z/OS and DASD-based Disaster Recovery - Blowing away the myths
DB2 for z/OS and DASD-based Disaster Recovery - Blowing away the myths
 
DB2 for z/OS - Starter's guide to memory monitoring and control
DB2 for z/OS - Starter's guide to memory monitoring and controlDB2 for z/OS - Starter's guide to memory monitoring and control
DB2 for z/OS - Starter's guide to memory monitoring and control
 
Coupling Facility CPU
Coupling Facility CPUCoupling Facility CPU
Coupling Facility CPU
 
Much Ado About CPU
Much Ado About CPUMuch Ado About CPU
Much Ado About CPU
 
DB2 Data Sharing Performance for Beginners
DB2 Data Sharing Performance for BeginnersDB2 Data Sharing Performance for Beginners
DB2 Data Sharing Performance for Beginners
 
Even More Fun With DDF
Even More Fun With DDFEven More Fun With DDF
Even More Fun With DDF
 
Educational seminar lessons learned from customer db2 for z os health check...
Educational seminar   lessons learned from customer db2 for z os health check...Educational seminar   lessons learned from customer db2 for z os health check...
Educational seminar lessons learned from customer db2 for z os health check...
 
Best practices for DB2 for z/OS log based recovery
Best practices for DB2 for z/OS log based recoveryBest practices for DB2 for z/OS log based recovery
Best practices for DB2 for z/OS log based recovery
 
A First Look at the DB2 10 DSNZPARM Changes
A First Look at the DB2 10 DSNZPARM ChangesA First Look at the DB2 10 DSNZPARM Changes
A First Look at the DB2 10 DSNZPARM Changes
 
Top 5 performance and capacity challenges for z/OS
Top 5 performance and capacity challenges for z/OS Top 5 performance and capacity challenges for z/OS
Top 5 performance and capacity challenges for z/OS
 
DB2 Through My Eyes
DB2 Through My EyesDB2 Through My Eyes
DB2 Through My Eyes
 
IMS09 ims v14 higlights
IMS09   ims v14 higlightsIMS09   ims v14 higlights
IMS09 ims v14 higlights
 
Introduction to FlashCopy
Introduction to FlashCopy Introduction to FlashCopy
Introduction to FlashCopy
 
IBM DB2 Analytics Accelerator Trends & Directions by Namik Hrle
IBM DB2 Analytics Accelerator  Trends & Directions by Namik Hrle IBM DB2 Analytics Accelerator  Trends & Directions by Namik Hrle
IBM DB2 Analytics Accelerator Trends & Directions by Namik Hrle
 

Similar to Munich 2016 - Z011601 Martin Packer - Parallel Sysplex Performance Topics topics

2009-01-28 DOI NBC Red Hat on System z Performance Considerations
2009-01-28 DOI NBC Red Hat on System z Performance Considerations2009-01-28 DOI NBC Red Hat on System z Performance Considerations
2009-01-28 DOI NBC Red Hat on System z Performance ConsiderationsShawn Wells
 
CPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCoburn Watson
 
3.3 gpp NR USER Plane introduction
3.3 gpp NR USER Plane introduction3.3 gpp NR USER Plane introduction
3.3 gpp NR USER Plane introductionSaurabh Verma
 
SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)
SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)
SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)Yuuki Takano
 
Инновации Cisco для операторов связи
Инновации Cisco для операторов связиИнновации Cisco для операторов связи
Инновации Cisco для операторов связиCisco Russia
 
Benchmarking distributed tracers
Benchmarking distributed tracersBenchmarking distributed tracers
Benchmarking distributed tracersQiao Han
 
VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series VMworld
 
Revisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS SchedulerRevisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS SchedulerYongseok Oh
 
Technical sales education enterprise- svc and ibm flash best practices update
Technical sales education   enterprise- svc and ibm flash best practices updateTechnical sales education   enterprise- svc and ibm flash best practices update
Technical sales education enterprise- svc and ibm flash best practices updatesolarisyougood
 
6 open capi_meetup_in_japan_final
6 open capi_meetup_in_japan_final6 open capi_meetup_in_japan_final
6 open capi_meetup_in_japan_finalYutaka Kawai
 
High-performance 32G Fibre Channel Module on MDS 9700 Directors:
High-performance 32G Fibre Channel Module on MDS 9700 Directors:High-performance 32G Fibre Channel Module on MDS 9700 Directors:
High-performance 32G Fibre Channel Module on MDS 9700 Directors:Tony Antony
 
3 g huawei ran resource monitoring and management recommended
3 g huawei ran resource monitoring and management recommended3 g huawei ran resource monitoring and management recommended
3 g huawei ran resource monitoring and management recommendedMery Koto
 
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Amazon Web Services
 

Similar to Munich 2016 - Z011601 Martin Packer - Parallel Sysplex Performance Topics topics (20)

OpenPOWER Webinar
OpenPOWER Webinar OpenPOWER Webinar
OpenPOWER Webinar
 
2009-01-28 DOI NBC Red Hat on System z Performance Considerations
2009-01-28 DOI NBC Red Hat on System z Performance Considerations2009-01-28 DOI NBC Red Hat on System z Performance Considerations
2009-01-28 DOI NBC Red Hat on System z Performance Considerations
 
Fast Convergence in IP Network
Fast Convergence in IP Network Fast Convergence in IP Network
Fast Convergence in IP Network
 
CPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performance
 
Technology (1)
Technology (1)Technology (1)
Technology (1)
 
Security-as-a-Service using SDN
Security-as-a-Service using SDNSecurity-as-a-Service using SDN
Security-as-a-Service using SDN
 
3.3 gpp NR USER Plane introduction
3.3 gpp NR USER Plane introduction3.3 gpp NR USER Plane introduction
3.3 gpp NR USER Plane introduction
 
SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)
SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)
SF-TAP: Scalable and Flexible Traffic Analysis Platform (USENIX LISA 2015)
 
Инновации Cisco для операторов связи
Инновации Cisco для операторов связиИнновации Cisco для операторов связи
Инновации Cisco для операторов связи
 
Benchmarking distributed tracers
Benchmarking distributed tracersBenchmarking distributed tracers
Benchmarking distributed tracers
 
VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series
 
Revisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS SchedulerRevisiting CephFS MDS and mClock QoS Scheduler
Revisiting CephFS MDS and mClock QoS Scheduler
 
Technical sales education enterprise- svc and ibm flash best practices update
Technical sales education   enterprise- svc and ibm flash best practices updateTechnical sales education   enterprise- svc and ibm flash best practices update
Technical sales education enterprise- svc and ibm flash best practices update
 
RISC.ppt
RISC.pptRISC.ppt
RISC.ppt
 
13 risc
13 risc13 risc
13 risc
 
6 open capi_meetup_in_japan_final
6 open capi_meetup_in_japan_final6 open capi_meetup_in_japan_final
6 open capi_meetup_in_japan_final
 
13 risc
13 risc13 risc
13 risc
 
High-performance 32G Fibre Channel Module on MDS 9700 Directors:
High-performance 32G Fibre Channel Module on MDS 9700 Directors:High-performance 32G Fibre Channel Module on MDS 9700 Directors:
High-performance 32G Fibre Channel Module on MDS 9700 Directors:
 
3 g huawei ran resource monitoring and management recommended
3 g huawei ran resource monitoring and management recommended3 g huawei ran resource monitoring and management recommended
3 g huawei ran resource monitoring and management recommended
 
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
 

More from Martin Packer

Munich 2016 - Z011597 Martin Packer - How To Be A Better Performance Specialist
Munich 2016 - Z011597 Martin Packer - How To Be A Better Performance SpecialistMunich 2016 - Z011597 Martin Packer - How To Be A Better Performance Specialist
Munich 2016 - Z011597 Martin Packer - How To Be A Better Performance SpecialistMartin Packer
 
Munich 2016 - Z011599 Martin Packer - More Fun With DDF
Munich 2016 - Z011599 Martin Packer - More Fun With DDFMunich 2016 - Z011599 Martin Packer - More Fun With DDF
Munich 2016 - Z011599 Martin Packer - More Fun With DDFMartin Packer
 
I Know What You Did Last Summer
I Know What You Did Last SummerI Know What You Did Last Summer
I Know What You Did Last SummerMartin Packer
 
Optimizing z/OS Batch
Optimizing z/OS BatchOptimizing z/OS Batch
Optimizing z/OS BatchMartin Packer
 
Memory Matters in 2011
Memory Matters in 2011Memory Matters in 2011
Memory Matters in 2011Martin Packer
 
Curt Cotner DDF Inactive Threads Support DB2 Version 3
Curt Cotner DDF Inactive Threads Support DB2 Version 3Curt Cotner DDF Inactive Threads Support DB2 Version 3
Curt Cotner DDF Inactive Threads Support DB2 Version 3Martin Packer
 

More from Martin Packer (8)

Munich 2016 - Z011597 Martin Packer - How To Be A Better Performance Specialist
Munich 2016 - Z011597 Martin Packer - How To Be A Better Performance SpecialistMunich 2016 - Z011597 Martin Packer - How To Be A Better Performance Specialist
Munich 2016 - Z011597 Martin Packer - How To Be A Better Performance Specialist
 
Munich 2016 - Z011599 Martin Packer - More Fun With DDF
Munich 2016 - Z011599 Martin Packer - More Fun With DDFMunich 2016 - Z011599 Martin Packer - More Fun With DDF
Munich 2016 - Z011599 Martin Packer - More Fun With DDF
 
Time For DIME
Time For DIMETime For DIME
Time For DIME
 
I Know What You Did Last Summer
I Know What You Did Last SummerI Know What You Did Last Summer
I Know What You Did Last Summer
 
Optimizing z/OS Batch
Optimizing z/OS BatchOptimizing z/OS Batch
Optimizing z/OS Batch
 
Much Ado About CPU
Much Ado About CPUMuch Ado About CPU
Much Ado About CPU
 
Memory Matters in 2011
Memory Matters in 2011Memory Matters in 2011
Memory Matters in 2011
 
Curt Cotner DDF Inactive Threads Support DB2 Version 3
Curt Cotner DDF Inactive Threads Support DB2 Version 3Curt Cotner DDF Inactive Threads Support DB2 Version 3
Curt Cotner DDF Inactive Threads Support DB2 Version 3
 

Recently uploaded

The Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfThe Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfSeasiaInfotech2
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsMemoori
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embeddingZilliz
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr LapshynFwdays
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 

Recently uploaded (20)

DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
The Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfThe Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdf
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial Buildings
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embedding
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 

Munich 2016 - Z011601 Martin Packer - Parallel Sysplex Performance Topics topics

  • 1. Technical University/Symposia materials may not be reproduced in whole or in part without the prior written permission of IBM. 9.0 © Copyright IBM Corporation 2015 Parallel Sysplex Performance Topics Session z011601 Martin Packer IBM
  • 2. Abstract Over recent years RMF's Parallel Sysplex instrumentation has improved tremendously. For example, Coupling Facility CPU reporting was enhanced, to give more granularity. And Coupling Facility links were reported on much better. Also RMF support of XCF was enhanced. This presentation outlines my experience with this important new instrumentation, from a number of perspectives.
  • 3. Topics § Structure-Level CPU § CPU / LPAR Match Up Between 70-1 and 74-4 § Structure Duplexing § XCF Traffic § Coupling Facility Link Information § Thin Interrupts Instrumentation § Conclusions and Musings
  • 5. Structure-Level CPU Consumption §  SMF 74-4 Field: R744SETM – “Structure Execution Time” § Always 100% Capture Ratio – Adds up to R744PBSY § Multiple uses: – Capacity planning for changing request rates – Examine which structures are large consumers – Compute CPU cost of a request • And compare to service time • Interesting number is “non-CPU” element of service time - as we shall see – Understand whether CPU per request has degraded – Estimating Structure Duplexing cost NOTE: Need to collect 74-4 data from all z/OS systems sharing to get total request rate - Otherwise “CPU per request” calculation will overestimate
  • 6. CPU By Structure – For Capacity Planning
  • 7. Structure CPU Experiment § All requests were Sync § One minute RMF intervals – Sorted by request rate increasing § Run was 1-way DB2 Datasharing – Only really active structures ISGLOCK and LOCK1 § Red lines are CPU time per request § Blue lines are Service time per request § ISGLOCK case: “low volume” – Shows amortization of some fixed cost effect – CF used IC links § LOCK1 case: “high volume” – More reliable for capacity planning – CF used a mixture of ISC and ICB links
  • 8. ISGLOCK Requests 0 2 4 6 8 10 12 14 16 0 10 20 30 40 50 60 70 Requests / Second Microseconds CPU Time Service Time 3us?
  • 9. LOCK1 Requests 0 2 4 6 8 10 12 750 800 850 900 Requests / Second Microseconds CPU Time Service Time 3.5us?
  • 10. And From My Travels... § Next chart isn't from the experiment just described – A real customer system – An old case so numbers old § A Group Buffer Pool § Not duplexed § ISC-Connected – Necessary for the customer's estate § Clearly something goes wrong at about 1100 requests / second – Especially in response time terms but also CPU • (Coupling Facility not CPU constrained) § Options include – Managing the request rate to below 1100 / sec – Working on the request mix – Infrastructure reconfiguration
  • 11. 25us?
  • 12. CPU / LPAR Match Up Between 70-1 and 74-4
  • 13. λ Managed out of Pool 5 in modern processor families λ Pool numbers given in SMF 70 as index into table of labels λ Recommendation: Manage in reporting as a separate pool λ Follow special CF sizing guidelines λ Especially for takeover situations λ Always runs at full speed λ So good technology match for coupled z/OS images on same footprint λ Another good reason to use ICFs is IC links λ Shared ICFs strongly discouraged for Production λ Especially if the CF image has Dynamic Dispatch turned on λ Unconvinced Thin Interrupts negates this advice λ Should not run ANY coupling facility above 50% busy λ Especially if we need to be able to recover structures onto it Internal Coupling Facility - Basics
  • 14. ICF CPU Instrumentation § SMF 74-4 view different from SMF 70-1 LPAR view of processor busy • R744PBSY is CPU time processing requests • R744PWAI is CPU time while CFCC is not processing requests but it is still using CF cycles • For Dynamic Dispatch PWAI is time when not processing CF requests but Logical CP not yet taken back by PR/SM • CF Thin Interrupts will reduce this • For dedicated or non-Dynamic Dispatch cases sum is constant • For Dynamic Dispatch sum can vary. § Number of defined processors is number of CF Processor Data sections in 74-4 • Fields for dedicated (R744FPDN) and shared (R744FPSN) processors • Also whether individual engine is dedicated (R744PTYP) and its weight (R744PWGT) § PBSY and PWAI Can be examined down to Coupling Facility engine level
  • 15. CF LPAR Identification In SMF 70-1 Is Was Complex § Need to match LPARs in SMF 70-1 with coupling facilities in SMF 74-4 to get proper CPU picture § 74-4 has machine serial number – Allows correlation in most cases • But LPAR names and CF Names often don't match • Often multiple CF's in same footprint with similar configuration • Sometimes there are multiple CF's with the same name • My code – in extremis – uses the presence of IC links to determine “colocality” • [I'm slowly learning :-) not all CF LPARs are in Pool 5]
  • 16. Additional Instrumentation - OA21140 § Everyone has this support § Ancient APAR integrated into recent z/OS releases § Introduced to support zHPF – Has other SMF and reporting improvements • HiperDispatch Vertical Polarisation indicators at ENGINE level – Type 70 • Normalisation factor for zIIP – Type 70 § Adds CF LPAR Partition Number – Allows matching with SMF 70-1 § RMF Level (SMFxxSRL) changed to X'55'
  • 18. Structure Duplexing Performance § Additional Traffic – For lock structures duplexing generates double the traffic – Otherwise only the writes are duplicated – Additional CPU cost § Additional Physical Resources – A second coupling facility • Documented in 74-4 – Additional memory – but “white space” rules say “not really” – Additional links – to second coupling facility and between it and the primary • Documented in SMF 74-4 SMD = System-Managed Duplexing UMD = User-Managed Duplexing
  • 19. Structure Duplexing Performance - Response Times § For SMD structures both requests must complete – Response time is that of the slowest • So all requests are essentially with “remote” response times • High likelihood of requests becoming asynchronous • For low contention rates applications might experience longer lock acquisition times § For UMD structures both requests must complete – But only for writes – So writes performed with “remote” response times – With high a read-to-write ratio request response times might not be significantly extended – Only example: DB2 Group Buffer Pools § Response time elongation measured by RMF PR WT and PR CMP times – Former suggests better link infrastructure – Latter suggests a more capable peer coupling facility
  • 20.
  • 21.
  • 22.
  • 24. XCF – Groups Worth Looking At § 74-2 has “job name” as well as member name § You can now answer more detailed questions about traffic: – For most XCF groups member name is useless § Traffic valuable for three reasons: – Explains links, structure, buffering etc demand – You can look better at eg DB2 IRLM Global Lock tuning – You can see topology without special instrumentation – For example, DB2 IRLM address spaces – For example, CICS – I've never seen a customer using other than DFHIR000 group
  • 26. Coupling Facility Path Information § Dramatically improved in CFLEVEL 18 (zEC12) – RMF APAR OA37826 • SMF 74-4 • Coupling Facility Activity Report – Configuration: • Detailed adapter and link type, PCHID, CHPID – OA37826 gives CHPID even without CFLEVEL 18 • Infiniband, ISC, and ICA-SR only – Performance: • “Degraded” flag • If this flag set then call your Customer Engineer • Channel Path Latency Time (R744HLAT) – Divide by 10 us to give distance estimate in Postprocessor Report – Would be interesting if it degraded (as it shouldn’t)
  • 27.
  • 28. R744HOPM - Channel path operation mode Value Meaning X'01' CFP path supporting a 1.0625 Gbit/s data rate X'02' CFP path supporting a 2.125 Gbit/s data rate X'10' CIB path operating at 1x bandwidth using the IFB protocol, adapter type HCA2-O LR X'11' CIB path operating at 12x bandwidth using the IFB protocol, adapter type HCA2-O X'20' CIB path operating at 1x bandwidth using the IFB protocol, adapter type HCA3-O LR X'21' CIB path operating at 12x bandwidth using the IFB protocol, adapter type HCA3-O X'30' CIB path operating at 12x bandwidth using the IFB3 protocol, adapter type HCA3-O X’40’ CS5 path operating at 8x bandwidth using the PCIe third generation protocol, adapter type PCIe-O ßICA-SR
  • 29. I’ve Blogged On This Subject Numerous Times •  System zEC12 CFLEVEL 18 RMF Instrumenta8on Improvements •  Coupling Facility Topology Informa8on - A Con8nuing Journey •  The Missing Link? •  The Effect Of CF Structure Distance •  What's The Latency, Kenneth? •  What's The Latency Really? •  And Latency Once More
  • 30. Coupling Facility Thin Interrupts Instrumenta<on •  Logical engines acquired and released in a more 8mely fashion •  See Barbara Weiler paper: Coupling Thin Interrupts and Coupling Facility Performance in Shared Processor Environments •  If you have SMF 74-4 for a Shared CF Engine Coupling Facility •  With OA42682 flag byte R744FFLG Bit 4 is set if DYNDISP=THIN •  If CFLEVEL (R744FLVL) > 18 •  R744PWAI will be reduced, compared to DYNDISP=NO or YES •  But many “shared engine” CFs go “undocumented” in SMF 74-4 •  For these expect SMF70PDT - SMF70EDT to be greater than before •  PR/SM playing a more concerted role
  • 31. Conclusions and Musings § I think we've come a long way with Coupling Facility CPU – Capacity Planning is now down to the structure level • But not to the structure-by-system level – We can now tie up the Coupling Facility and LPAR views of CPU § I'd encourage you to revisit your Parallel Sysplex reporting – Including for all the other aspects we didn't have time for § Structure Duplexing needs particular care – A very useful resilience feature that has performance considerations § XCF Traffic a subject worthy of study § Coupling Facility Link Information a useful emergent topic