The Database Sizing Workflow
Upcoming SlideShare
Loading in...5
×
 

Like this? Share it with your network

Share

The Database Sizing Workflow

on

  • 154 views

 

Statistics

Views

Total Views
154
Views on SlideShare
154
Embed Views
0

Actions

Likes
1
Downloads
9
Comments
0

0 Embeds 0

No embeds

Accessibility

Categories

Upload Details

Uploaded via as Microsoft PowerPoint

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment
  • Outline: <br /> Ultimate Exadata IO monitoring – Flash, HardDisk , & Write back cache overhead http://www.kylehailey.com/oaktable-world/agenda/ <br /> I’ll do a session highlighting a very write intensive OLTP Exadata environment and will discuss the different ways to monitor IO from the database and storage layer perspective and correlating it back to the application by mining the dba_hist_sqlstat data. I’ll also touch on utilizing the OEM12c Metric Extensions and BI Publisher integration to ultimately scale the monitoring to a bunch of Exadata environments. It’s going to be a fun hacking session. <br /> &gt; discuss the capacity doodle <br /> &gt; the variables <br /> &gt; monitoring <br /> &gt; the reclaim <br /> &gt; highlight issue on very write intensive OLTP environment <br /> &gt; monitoring problem <br /> on OEM perf page &gt; show IO perf page not accounting the flash IOs <br /> ** partly because some people in the team have access to only limited view of things <br /> ** or they have difficulty interpreting the numbers, they need simple stuff <br /> on OEM12c storage grid perf &gt; although 12c has exadata IOs monitoring but, <br /> I&apos;d like to get the IOPS number separated by flash and disk <br /> &gt; wbfc patent <br /> &gt; write back cache http://goo.gl/2WCmw <br /> &gt; exadata oltp optimizations <br /> &gt; discuss about the basic architecture <br /> &gt; discuss different ways to monitor IO (email to randy) http://goo.gl/i660CZ <br /> Different views of IO performance <br /> SECTION 1: USER IO wait class and cell single block reads latency with curve fitting <br /> SECTION 2: Small IOPS vs Large IOPS <br /> SECTION 3: Flash vs HD IOPS <br /> SECTION 4: Flash vs HD IOPS with read/write breakdown <br /> SECTION 5: IO throughput read/write MB/s <br /> SECTION 6: Drill down on smart scans affecting cell single block latency on 24hour period <br /> &gt; IO workload correlate up to the topevents and sqlstat data <br /> &gt; causal links - produce analysis which relates database load to application processing creating a strong understanding <br /> front to back as an enabler to ‘fix’ <br /> &gt; feedback loop on what is working and what is not <br /> &gt; track IO config changes - IORM (topevents data) <br /> &gt; basic, auto, low latency... and when it is applicable <br /> &gt; scaling it! <br /> &gt; metrics extension <br /> &gt; BIP <br /> &gt; show data model <br /> &gt; email everyday <br />
  • Just a brief introduction of myself.. <br />
  • And this is what the tar files looks like and it’s just a simple CSV output of AWR data <br />
  • And what makes the tableau really interesting is it automatically creates “dimensions” out of those CSV files <br /> My objective on this image is to quickly see the utilization of CPU if I combine particular instances and I can do that by just pulling the Total Oracle CPU seconds metric on the graph and that’s the boxed line chart at the bottom and that&apos;s the sum of Total Oracle CPU seconds of the instances that are selected on the right hand side portion of the graph.  <br /> So let’s say I want to consolidated the 3 instances on a single 24cores compute node.. (24cores x 3600 seconds = 86400 seconds of CPU capacity) I’ll be able to tell from the workload trend that it can fit on that box and I’m expecting the highest CPU Utilization that I’ll have is about 69% (60000/86400) <br /> And you can also right click on this and do a “View Data” <br />
  • So how it works is whatever SNAP_ID on the selected instances that falls on a specific hour dimension will get summed. So this tool automatically takes care of snap interval differences of the databases which is tedious to do manually. <br />

The Database Sizing Workflow Presentation Transcript

  • 1. The Database Sizing Workflow Presented by: Karl Arao 1
  • 2. whoami Karl Arao • Senior Technical Consultant @ Enkitec • Performance and Capacity Planning Enthusiast 7+ years DBA experience Oracle ACE, OCP-DBA, RHCE, OakTable Blog: karlarao.wordpress.com Wiki: karlarao.tiddlyspot.com Twitter: @karlarao www.enkitec.com 2
  • 3. www.enkitec.com 3 200+ 3
  • 4. Agenda • The sizing scenarios/objective • The general sizing workflow – Extract – Visualize – Model – Project • Putting it all together: Real Sizing Scenarios www.enkitec.com 4
  • 5. www.enkitec.com 5
  • 6. The sizing scenarios/objective • Consolidation, HW refresh, platform migration – How many can fit? – Can I combine A + B + ½ of C? – What's the ideal hardware to buy - "right sizing" www.enkitec.com 6
  • 7. The sizing workflow – Extract • Workload data – Visualize • Consolidated peak workload – Model • Provisioning plan – Project • “Headroom” www.enkitec.com 7
  • 8. www.enkitec.com 8
  • 9. Extract www.enkitec.com 9
  • 10. AWR data • Top Events – AAS CPU, latency, wait class • SYSSTAT – PGA, SGA, physical memory, Executes/sec • IO – IOPS breakdown, MB/s • CPU – Load Average, NUM_CPUs, • Storage – total storage size, per tablespace size • Services – distribution of workload/modules • Top SQL – PIOs, LIOs, modules, SQL type, SQL_ID, PX Correlate across months of workload data! http://goo.gl/7uCk7w
  • 11. www.enkitec.com 11
  • 12. www.enkitec.com 12 OS data
  • 13. Visualize www.enkitec.com 13
  • 14. Visualize – Workload Characterization General Workload • top events • load profile (exec/sec) • top modules/services CPU usage • CPU, cpuwait, scheduler SGA/PGA IOPS, MB/s, latency • IO breakdown • read/write ratio Storage Size www.enkitec.com 14
  • 15. • Tableau auto creates a time dimension for the time column “MM/DD/YY HH24:MI:SS” of AWR csv output www.enkitec.com 15
  • 16. www.enkitec.com 16 • Summary and Underlying data 1-2AM 2-3AM
  • 17. www.enkitec.com 17 Consolidated CPU usage
  • 18. Model www.enkitec.com 18
  • 19. What to model? • the provisioning plan – instance mapping – node failure scenarios – resource management • backups, test/dev, DR, ZFS • hardware options • memory upgrade • redundancy (normal or high) www.enkitec.com 19
  • 20. www.enkitec.com 20
  • 21. Projection www.enkitec.com 21
  • 22. www.enkitec.com 22
  • 23. Putting it all together www.enkitec.com 23
  • 24. Summary • The sizing scenarios/objective • The 4 points of the sizing worklflow www.enkitec.com 24
  • 25. References • Where did my CPU go? (webinar) http://www.youtube.com/watch?v=WXktSUbE4AU (paper) http://goo.gl/qP1xqr • Book: Computer Architecture: A Quantitative Approach 5th Ed - Chapter1 Section1.10 Putting it all together Perf, Price, Power http://goo.gl/MXigAQ • Book: The Art of Scalability - Ch11 “Headroom” http://theartofscalability.com • Viz Example: CPU sizing 15 vs 60 mins snap interval http://goo.gl/rOJ9M4 • Viz Example: Different views of IO performance http://goo.gl/i660CZ • Exadata Provisioning Worksheet http://www.slideshare.net/karlarao/pape- rkaraoconsolidation-successstory www.enkitec.com 25 karl.arao@enkitec.com karlarao.wordpress.com karlarao.tiddlyspot.com @karlarao