SlideShare a Scribd company logo
WHY DBA IS NEEDED IN  ORACLE DWH PROJECTS? By  Anurag Vidyarthi  (Accenture,Norway)
Main feature of DWH projects ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Partitioning ,[object Object],[object Object],[object Object],[object Object]
Parallelism ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Compression  ,[object Object],[object Object],[object Object],[object Object]
Materialized Views(MVs) ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Demand for Performance scalability ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Demand for Performance scalability ,[object Object],[object Object],[object Object],[object Object],[object Object]
Typical Question : Query is taking time   ,[object Object],[object Object],[object Object],[object Object]
Miscellaneous DBA Tasks ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
ETL Job Analysis ,[object Object],[object Object],[object Object],[object Object],[object Object]
Pro-active co-ordination ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Capacity Planning ,[object Object],[object Object],[object Object],[object Object],[object Object]
Thanks ,[object Object]

More Related Content

Similar to Why dba needed in dwh projects

When & Why\'s of Denormalization
When & Why\'s of DenormalizationWhen & Why\'s of Denormalization
When & Why\'s of Denormalization
Aliya Saldanha
 
The High Performance DBA Optimizing Databases For High Performance
The High Performance DBA Optimizing Databases For High PerformanceThe High Performance DBA Optimizing Databases For High Performance
The High Performance DBA Optimizing Databases For High Performance
Embarcadero Technologies
 
White Paper - Lepide SQL Storage Manager
White Paper - Lepide SQL Storage ManagerWhite Paper - Lepide SQL Storage Manager
White Paper - Lepide SQL Storage Manager
Sumant Kumar
 

Similar to Why dba needed in dwh projects (20)

Teradata sql-tuning-top-10
Teradata sql-tuning-top-10Teradata sql-tuning-top-10
Teradata sql-tuning-top-10
 
Capacity management for ETL System
Capacity management for ETL SystemCapacity management for ETL System
Capacity management for ETL System
 
When & Why\'s of Denormalization
When & Why\'s of DenormalizationWhen & Why\'s of Denormalization
When & Why\'s of Denormalization
 
Data mining and warehousing (uca15 e04)
Data mining and warehousing (uca15 e04)Data mining and warehousing (uca15 e04)
Data mining and warehousing (uca15 e04)
 
Monitorando performance no Azure SQL Database
Monitorando performance no Azure SQL DatabaseMonitorando performance no Azure SQL Database
Monitorando performance no Azure SQL Database
 
SQL Server 2017 - Adaptive Query Processing and Automatic Query Tuning
SQL Server 2017 - Adaptive Query Processing and Automatic Query TuningSQL Server 2017 - Adaptive Query Processing and Automatic Query Tuning
SQL Server 2017 - Adaptive Query Processing and Automatic Query Tuning
 
ETL•Accelerator
ETL•AcceleratorETL•Accelerator
ETL•Accelerator
 
05_DP_300T00A_Optimize.pptx
05_DP_300T00A_Optimize.pptx05_DP_300T00A_Optimize.pptx
05_DP_300T00A_Optimize.pptx
 
The High Performance DBA Optimizing Databases For High Performance
The High Performance DBA Optimizing Databases For High PerformanceThe High Performance DBA Optimizing Databases For High Performance
The High Performance DBA Optimizing Databases For High Performance
 
Dremel Paper Review
Dremel Paper ReviewDremel Paper Review
Dremel Paper Review
 
Query Tuning Azure SQL Databases
Query Tuning Azure SQL DatabasesQuery Tuning Azure SQL Databases
Query Tuning Azure SQL Databases
 
IRJET- Physical Database Design Techniques to improve Database Performance
IRJET-	 Physical Database Design Techniques to improve Database PerformanceIRJET-	 Physical Database Design Techniques to improve Database Performance
IRJET- Physical Database Design Techniques to improve Database Performance
 
How to Achieve Fast Data Performance in Big Data, Logical Data Warehouse, and...
How to Achieve Fast Data Performance in Big Data, Logical Data Warehouse, and...How to Achieve Fast Data Performance in Big Data, Logical Data Warehouse, and...
How to Achieve Fast Data Performance in Big Data, Logical Data Warehouse, and...
 
Understanding System Performance
Understanding System PerformanceUnderstanding System Performance
Understanding System Performance
 
Design and development of oracle database system
Design and development of oracle database systemDesign and development of oracle database system
Design and development of oracle database system
 
KoprowskiT_SQLSat409_MaintenancePlansForBeginners
KoprowskiT_SQLSat409_MaintenancePlansForBeginnersKoprowskiT_SQLSat409_MaintenancePlansForBeginners
KoprowskiT_SQLSat409_MaintenancePlansForBeginners
 
KoprowskiT_SQLSaturday409_MaintenancePlansForBeginners
KoprowskiT_SQLSaturday409_MaintenancePlansForBeginnersKoprowskiT_SQLSaturday409_MaintenancePlansForBeginners
KoprowskiT_SQLSaturday409_MaintenancePlansForBeginners
 
ADV Slides: Comparing the Enterprise Analytic Solutions
ADV Slides: Comparing the Enterprise Analytic SolutionsADV Slides: Comparing the Enterprise Analytic Solutions
ADV Slides: Comparing the Enterprise Analytic Solutions
 
White Paper - Lepide SQL Storage Manager
White Paper - Lepide SQL Storage ManagerWhite Paper - Lepide SQL Storage Manager
White Paper - Lepide SQL Storage Manager
 
Practical SQL query monitoring and optimization
Practical SQL query monitoring and optimizationPractical SQL query monitoring and optimization
Practical SQL query monitoring and optimization
 

Recently uploaded

Future Visions: Predictions to Guide and Time Tech Innovation, Peter Udo Diehl
Future Visions: Predictions to Guide and Time Tech Innovation, Peter Udo DiehlFuture Visions: Predictions to Guide and Time Tech Innovation, Peter Udo Diehl
Future Visions: Predictions to Guide and Time Tech Innovation, Peter Udo Diehl
Peter Udo Diehl
 
Essentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with ParametersEssentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with Parameters
Safe Software
 

Recently uploaded (20)

Transcript: Selling digital books in 2024: Insights from industry leaders - T...
Transcript: Selling digital books in 2024: Insights from industry leaders - T...Transcript: Selling digital books in 2024: Insights from industry leaders - T...
Transcript: Selling digital books in 2024: Insights from industry leaders - T...
 
Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...
 
ODC, Data Fabric and Architecture User Group
ODC, Data Fabric and Architecture User GroupODC, Data Fabric and Architecture User Group
ODC, Data Fabric and Architecture User Group
 
Future Visions: Predictions to Guide and Time Tech Innovation, Peter Udo Diehl
Future Visions: Predictions to Guide and Time Tech Innovation, Peter Udo DiehlFuture Visions: Predictions to Guide and Time Tech Innovation, Peter Udo Diehl
Future Visions: Predictions to Guide and Time Tech Innovation, Peter Udo Diehl
 
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
 
Essentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with ParametersEssentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with Parameters
 
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
 
Neuro-symbolic is not enough, we need neuro-*semantic*
Neuro-symbolic is not enough, we need neuro-*semantic*Neuro-symbolic is not enough, we need neuro-*semantic*
Neuro-symbolic is not enough, we need neuro-*semantic*
 
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdfFIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
 
JMeter webinar - integration with InfluxDB and Grafana
JMeter webinar - integration with InfluxDB and GrafanaJMeter webinar - integration with InfluxDB and Grafana
JMeter webinar - integration with InfluxDB and Grafana
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
 
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
 
Demystifying gRPC in .Net by John Staveley
Demystifying gRPC in .Net by John StaveleyDemystifying gRPC in .Net by John Staveley
Demystifying gRPC in .Net by John Staveley
 
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
 
IOS-PENTESTING-BEGINNERS-PRACTICAL-GUIDE-.pptx
IOS-PENTESTING-BEGINNERS-PRACTICAL-GUIDE-.pptxIOS-PENTESTING-BEGINNERS-PRACTICAL-GUIDE-.pptx
IOS-PENTESTING-BEGINNERS-PRACTICAL-GUIDE-.pptx
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
 
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdfFIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
 
Exploring UiPath Orchestrator API: updates and limits in 2024 🚀
Exploring UiPath Orchestrator API: updates and limits in 2024 🚀Exploring UiPath Orchestrator API: updates and limits in 2024 🚀
Exploring UiPath Orchestrator API: updates and limits in 2024 🚀
 

Why dba needed in dwh projects

  • 1. WHY DBA IS NEEDED IN ORACLE DWH PROJECTS? By Anurag Vidyarthi (Accenture,Norway)
  • 2.
  • 3.
  • 4.
  • 5.
  • 6.
  • 7.
  • 8.
  • 9.
  • 10.
  • 11.
  • 12.
  • 13.
  • 14.

Editor's Notes

  1. A DBA is required in handling the above mentioned aspects of DWH projects.
  2. Worked Example : A customer wishes to build a Data Warehouse to support their customer information system. The database consists of about 10 core tables that comprise about 60 Gigabytes of data. They are going to perform a great deal of table scan activities as they mine the data for purchasing trends. They know the data is coming from a number of legacy systems and anticipate many problems consolidating the data. They anticipate regularly sweeping or scrubbing entire tables and creating updated copies. For this reason they know the processing requirements are going to be extremely high and for this reason have ordered a 20 CPU Sparc Server 2000E. They have configured 2 Gigabytes of memory into the system and now they need estimates as to the size of the I/O subsystem. They have requested that the database is fully mirrored in case of media failure. Assessment of Database Size and Volume based Estimate (base data+indexes+temp) * Factor for Admin. ( 60 + 15 + 30 ) * 1.7 = 179 Giga Bytes The factors are arbitrary and are based upon experience they can be altered when applicable. The disks available are 4 Gig Volumes so the number of disks prior to mirroring will be 179 / 4 = 45 Disk Drives Assessment of Disk Drives from CPU performance #CPUS * Disks/CPU estimate 20 * 8 = 160 Disk Drives In this case the number of disk drives to keep the CPUs busy exceeds the storage requirement. Using a volume based estimate could mean that there could be an I/O bottle neck on the database. Mirroring the database would double the number of disk drives to 90. Mirroring will also assist query processing as reads can be satisfied by multiple spindles. However the scale up is not perfect so we can assume an I/O scale up of about 50%. This means that the effective number of drives using a volume based estimate would approach about ( 1.5 * 45 ) 68 drives again well short of the 160 drives calculated to keep the CPUs busy. At this point these estimates need to be communicated to the project team and cost benefit assessment needs to be made. To get all CPUs busy will require doubling the number of disk drives. An assessment needs to be made to see if budget limitations allow the project buy almost twice the number of disks. At this point it usually becomes a judgment issue and compromises are required. Other factors that should be considered when making this estimate. Is the number of CPUs going to increase in the future Are the CPUs going to upgraded to faster ones in the future What are the chances of the database further increasing after phase one of the project. All these issues encourage early investment in disks however budgets will usually override the technical arguments. It is stated that this process is a very imprecise science and I/O subsystem capacity planning is one of the hardest to get correct. However if the database administrators cannot negotiate enough I/O resources there will be I/O bottlenecks later. If obtaining hardware resources are not an issue 160 Disks should be obtained as part of creating a balanced system.
  3. Links… https://www.indiana.edu/~dbateam/databases/oracle/dwperf.pdf