This document discusses decision support, data warehousing, and online analytical processing (OLAP). It defines these terms and compares online transaction processing (OLTP) to OLAP. It describes the evolution of decision support from batch reports to integrated data warehouses. The benefits of separating data warehouses from operational databases are outlined. Common architectures and the design/operational process are summarized.
In computing, a data warehouse (DW, DWH), or an enterprise data warehouse (EDW), is a database used for reporting (1) and data analysis (2). Integrating data from one or more disparate sources creates a central repository of data, a data warehouse (DW). Data warehouses store current and historical data and are used for creating trending reports for senior management reporting such as annual and quarterly comparisons.
1) Data warehousing aims to bring together information from multiple sources to provide a consistent database for decision support queries and analytical applications, offloading these tasks from operational transaction systems.
2) OLAP is focused on efficient multidimensional analysis of large data volumes for decision making, while OLTP is aimed at reliable processing of high-volume transactions.
3) A data warehouse is a subject-oriented, integrated collection of historical and summarized data used for analysis and decision making, separate from operational databases.
- Data warehousing aims to help knowledge workers make better decisions by integrating data from multiple sources and providing historical and aggregated data views. It separates analytical processing from operational processing for improved performance.
- A data warehouse contains subject-oriented, integrated, time-variant, and non-volatile data to support analysis. It is maintained separately from operational databases. Common schemas include star schemas and snowflake schemas.
- Online analytical processing (OLAP) supports ad-hoc querying of data warehouses for analysis. It uses multidimensional views of aggregated measures and dimensions. Relational and multidimensional OLAP are common architectures. Measures are metrics like sales, and dimensions provide context like products and time periods.
What is a Data Warehouse and How Do I Test It?RTTS
ETL Testing: A primer for Testers on Data Warehouses, ETL, Business Intelligence and how to test them.
Are you hearing and reading about Big Data, Enterprise Data Warehouses (EDW), the ETL Process and Business Intelligence (BI)? The software markets for EDW and BI are quickly approaching $22 billion, according to Gartner, and Big Data is growing at an exponential pace.
Are you being tasked to test these environments or would you like to learn about them and be prepared for when you are asked to test them?
RTTS, the Software Quality Experts, provided this groundbreaking webinar, based upon our many years of experience in providing software quality solutions for more than 400 companies.
You will learn the answer to the following questions:
• What is Big Data and what does it mean to me?
• What are the business reasons for a building a Data Warehouse and for using Business Intelligence software?
• How do Data Warehouses, Business Intelligence tools and ETL work from a technical perspective?
• Who are the primary players in this software space?
• How do I test these environments?
• What tools should I use?
This slide deck is geared towards:
QA Testers
Data Architects
Business Analysts
ETL Developers
Operations Teams
Project Managers
...and anyone else who is (a) new to the EDW space, (b) wants to be educated in the business and technical sides and (c) wants to understand how to test them.
MariaDB AX: Solución analítica con ColumnStoreMariaDB plc
MariaDB ColumnStore is a high performance columnar storage engine that provides fast and efficient analytics on large datasets in distributed environments. It stores data column-by-column for high compression and read performance. Queries are processed in parallel across nodes for scalability. MariaDB ColumnStore is used for real-time analytics use cases in industries like healthcare, life sciences, and telecommunications to gain insights from large datasets for applications like customer behavior analysis, genome research, and call data monitoring.
MariaDB AX: Analytics with MariaDB ColumnStoreMariaDB plc
MariaDB ColumnStore is a high performance columnar storage engine that provides fast and efficient analytics on large datasets in distributed environments. It stores data column-by-column for high compression and read performance. Queries are processed in parallel across nodes for scalability. MariaDB ColumnStore is used for real-time analytics use cases in industries like healthcare, life sciences, and telecommunications to gain insights from large datasets.
Data warehousing is an architectural model that gathers data from various sources into a single unified data model for analysis purposes. It consists of extracting data from operational systems, transforming it, and loading it into a database optimized for querying and analysis. This allows organizations to integrate data from different sources, provide historical views of data, and perform flexible analysis without impacting transaction systems. While implementation and maintenance of a data warehouse requires significant costs, the benefits include a single access point for all organizational data and optimized systems for analysis and decision making.
In computing, a data warehouse (DW, DWH), or an enterprise data warehouse (EDW), is a database used for reporting (1) and data analysis (2). Integrating data from one or more disparate sources creates a central repository of data, a data warehouse (DW). Data warehouses store current and historical data and are used for creating trending reports for senior management reporting such as annual and quarterly comparisons.
1) Data warehousing aims to bring together information from multiple sources to provide a consistent database for decision support queries and analytical applications, offloading these tasks from operational transaction systems.
2) OLAP is focused on efficient multidimensional analysis of large data volumes for decision making, while OLTP is aimed at reliable processing of high-volume transactions.
3) A data warehouse is a subject-oriented, integrated collection of historical and summarized data used for analysis and decision making, separate from operational databases.
- Data warehousing aims to help knowledge workers make better decisions by integrating data from multiple sources and providing historical and aggregated data views. It separates analytical processing from operational processing for improved performance.
- A data warehouse contains subject-oriented, integrated, time-variant, and non-volatile data to support analysis. It is maintained separately from operational databases. Common schemas include star schemas and snowflake schemas.
- Online analytical processing (OLAP) supports ad-hoc querying of data warehouses for analysis. It uses multidimensional views of aggregated measures and dimensions. Relational and multidimensional OLAP are common architectures. Measures are metrics like sales, and dimensions provide context like products and time periods.
What is a Data Warehouse and How Do I Test It?RTTS
ETL Testing: A primer for Testers on Data Warehouses, ETL, Business Intelligence and how to test them.
Are you hearing and reading about Big Data, Enterprise Data Warehouses (EDW), the ETL Process and Business Intelligence (BI)? The software markets for EDW and BI are quickly approaching $22 billion, according to Gartner, and Big Data is growing at an exponential pace.
Are you being tasked to test these environments or would you like to learn about them and be prepared for when you are asked to test them?
RTTS, the Software Quality Experts, provided this groundbreaking webinar, based upon our many years of experience in providing software quality solutions for more than 400 companies.
You will learn the answer to the following questions:
• What is Big Data and what does it mean to me?
• What are the business reasons for a building a Data Warehouse and for using Business Intelligence software?
• How do Data Warehouses, Business Intelligence tools and ETL work from a technical perspective?
• Who are the primary players in this software space?
• How do I test these environments?
• What tools should I use?
This slide deck is geared towards:
QA Testers
Data Architects
Business Analysts
ETL Developers
Operations Teams
Project Managers
...and anyone else who is (a) new to the EDW space, (b) wants to be educated in the business and technical sides and (c) wants to understand how to test them.
MariaDB AX: Solución analítica con ColumnStoreMariaDB plc
MariaDB ColumnStore is a high performance columnar storage engine that provides fast and efficient analytics on large datasets in distributed environments. It stores data column-by-column for high compression and read performance. Queries are processed in parallel across nodes for scalability. MariaDB ColumnStore is used for real-time analytics use cases in industries like healthcare, life sciences, and telecommunications to gain insights from large datasets for applications like customer behavior analysis, genome research, and call data monitoring.
MariaDB AX: Analytics with MariaDB ColumnStoreMariaDB plc
MariaDB ColumnStore is a high performance columnar storage engine that provides fast and efficient analytics on large datasets in distributed environments. It stores data column-by-column for high compression and read performance. Queries are processed in parallel across nodes for scalability. MariaDB ColumnStore is used for real-time analytics use cases in industries like healthcare, life sciences, and telecommunications to gain insights from large datasets.
Data warehousing is an architectural model that gathers data from various sources into a single unified data model for analysis purposes. It consists of extracting data from operational systems, transforming it, and loading it into a database optimized for querying and analysis. This allows organizations to integrate data from different sources, provide historical views of data, and perform flexible analysis without impacting transaction systems. While implementation and maintenance of a data warehouse requires significant costs, the benefits include a single access point for all organizational data and optimized systems for analysis and decision making.
The document provides an overview of data warehousing concepts including:
1) A data warehouse is a subject-oriented collection of integrated data used to support management decisions. It contains current and historical data.
2) A data warehouse architecture typically includes source systems, a staging area, and presentation layer for querying and reporting.
3) Data marts are focused subsets of a data warehouse tailored for specific business units or departments. There are dependent, independent, and hybrid approaches to building data marts.
This document discusses online analytical processing (OLAP) and related concepts. It defines data mining, data warehousing, OLTP, and OLAP. It explains that a data warehouse integrates data from multiple sources and stores historical data for analysis. OLAP allows users to easily extract and view data from different perspectives. The document also discusses OLAP cube operations like slicing, dicing, drilling, and pivoting. It describes different OLAP architectures like MOLAP, ROLAP, and HOLAP and data warehouse schemas and architecture.
Bill Hayduk is the founder and CEO of QuerySurge, a software division that provides data integration and analytics solutions, with headquarters in New York; QuerySurge was founded in 1996 and has grown to serve Fortune 1000 customers through partnerships with technology companies and consulting firms. The document discusses the data and analytics marketplace and provides an overview of concepts like data warehousing, ETL, BI, data quality, data testing, big data, Hadoop, and NoSQL.
Various Applications of Data Warehouse.pptRafiulHasan19
The document discusses various applications of data warehousing. It begins by describing problems with traditional transactional systems and how data warehouses address these issues. It then defines key components of a data warehouse including the extraction, transformation, and loading of data from various sources. The document outlines how online analytical processing (OLAP) tools, metadata repositories, and data mining techniques analyze and explore the collected data. Finally, it weighs the benefits of a data warehouse against the costs of implementation and maintenance.
Types of database processing,OLTP VS Data Warehouses(OLAP), Subject-oriented
Integrated
Time-variant
Non-volatile,
Functionalities of Data Warehouse,Roll-Up(Consolidation),
Drill-down,
Slicing,
Dicing,
Pivot,
KDD Process,Application of Data Mining
This document discusses data warehousing concepts and technologies. It defines a data warehouse as a subject-oriented, integrated, non-volatile, and time-variant collection of data used to support management decision making. It describes the data warehouse architecture including extract-transform-load processes, OLAP servers, and metadata repositories. Finally, it outlines common data warehouse applications like reporting, querying, and data mining.
5 Things that Make Hadoop a Game Changer
Webinar by Elliott Cordo, Caserta Concepts
There is much hype and mystery surrounding Hadoop's role in analytic architecture. In this webinar, Elliott presented, in detail, the services and concepts that makes Hadoop a truly unique solution - a game changer for the enterprise. He talked about the real benefits of a distributed file system, the multi workload processing capabilities enabled by YARN, and the 3 other important things you need to know about Hadoop.
To access the recorded webinar, visit the event site: https://www.brighttalk.com/webcast/9061/131029
For more information the services and solutions that Caserta Concepts offers, please visit http://casertaconcepts.com/
The document provides an overview of database, big data, and data science concepts. It discusses topics such as database management systems (DBMS), data warehousing, OLTP vs OLAP, data mining, and the data science process. Key points include:
- DBMS are used to store and manage data in an organized way for use by multiple users. Data warehousing is used to consolidate data from different sources.
- OLTP systems are for real-time transactional systems, while OLAP systems are used for analysis and reporting of historical data.
- Data mining involves applying algorithms to large datasets to discover patterns and relationships. The data science process involves business understanding, data preparation, modeling, evaluation, and deployment
This document provides an overview of data warehousing and related concepts. It defines a data warehouse as a centralized database for analysis and reporting that stores current and historical data from multiple sources. The document describes key elements of data warehousing including Extract-Transform-Load (ETL) processes, multidimensional data models, online analytical processing (OLAP), and data marts. It also outlines advantages such as enhanced access and consistency, and disadvantages like time required for data extraction and loading.
OLAP (online analytical processing) allows users to easily extract and analyze data from different perspectives. It stores data in multidimensional databases to allow for complex queries. There are three main types of OLAP - relational, multidimensional, and hybrid. OLAP is used with data warehouses to enable analytics like data mining and decision making. It provides benefits over transactional systems by facilitating flexible analysis of integrated data over time.
Are you confused by Big Data? Get in touch with this new "black gold" and familiarize yourself with undiscovered insights through our complimentary introductory lesson on Big Data and Hadoop!
The document provides an overview of data warehousing, decision support, online analytical processing (OLAP), and data mining. It discusses what data warehousing is, how it can help organizations make better decisions by integrating data from various sources and making it available for analysis. It also describes OLAP as a way to transform warehouse data into meaningful information for interactive analysis, and lists some common OLAP operations like roll-up, drill-down, slice and dice, and pivot. Finally, it gives a brief introduction to data mining as the process of extracting patterns and relationships from data.
The New Frontier: Optimizing Big Data ExplorationInside Analysis
The Briefing Room with Dr. Robin Bloor and Cirro
Live Webcast on February 11, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=0ec1fa381886313cc06d841015c65898
As information ecosystems continue to expand, businesses are searching for ways to combine traditional analytics with a new source of insight: Big Data. But with data flooding in from all kinds of sources, fast access and performance at scale can easily become an issue. One effective approach for solving this challenge is data federation, a method that involves taking the analytical processing to the data, allowing streamlined access to multiple data sources without the expensive ETL overhead or building of semantic layers.
Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor as he explains how the prevalence of distributed data calls for a new approach to Big Data. He will be briefed by Mark Theissen of Cirro, who will tout his company’s Data Hub, a data federation solution that provides a single point of access to all enterprise data assets without excessive data movements, preprocessing or staging. He will discuss how data federation differs from virtualization and ETL approaches, and demonstrate how a Cirro deployment solves the analytics challenge of integrating data silos across the data center – and the cloud – using the BI tools you already have on your desktop for real-time distributed analytics.
Visit InsideAnlaysis.com for more information.
MSBI online training offered by Quontra Solutions with special features having Extensive Training will be in both MSBI Online Training and Placement. We help you in resume preparation and conducting Mock Interviews.
Emphasis is given on important topics that were required and mostly used in real time projects. Quontra Solutions is an Online Training Leader when it comes to high-end effective and efficient IT Training. We have always been and still are focusing on the key aspect which is providing utmost effective and competent training to both students and professionals who are eager to enrich their technical skills.
The document provides an overview of key concepts in data warehousing and business intelligence, including:
1) It defines data warehousing concepts such as the characteristics of a data warehouse (subject-oriented, integrated, time-variant, non-volatile), grain/granularity, and the differences between OLTP and data warehouse systems.
2) It discusses the evolution of business intelligence and key components of a data warehouse such as the source systems, staging area, presentation area, and access tools.
3) It covers dimensional modeling concepts like star schemas, snowflake schemas, and slowly and rapidly changing dimensions.
1. Data warehousing is a process of transforming data from operational systems into a centralized data repository optimized for analysis and reporting.
2. A data warehouse contains integrated, subject-oriented, non-volatile data used to support organizational decision making through queries and analysis.
3. The data warehouse is designed to provide historical, summarized views of data to business users to help answer questions and make decisions.
The Data Lake and Getting Buisnesses the Big Data Insights They NeedDunn Solutions Group
Do terms like "Data Lake" confuse you? You’re not alone. With all of the technology buzzwords flying around today, it can become a task to keep up with and clearly understand each of them. However a data lake is definitely something to dedicate the time to understand. Leveraging data lake technology, companies are finally able to keep all of their disparate information and streams of data in one secure location ready for consumption at any time – this includes structured, unstructured, and semi-structured data. For more information on our Big Data Consulting Services, don’t hesitate to visit us online at: http://bit.ly/2fvV5rR
Sentiment analysis aims to identify the orientation of opinions in text, whether positive, negative, or neutral. It draws from fields like cognitive science, natural language processing, and machine learning. Challenges include domain dependence, sarcasm, and contrasting with standard text categorization where word presence indicates category. Approaches include subjectivity detection using graph algorithms, sentiment lexicons capturing word sentiment, and scoring adjective-adverb combinations. Applications include review analysis, question answering, and developing "hate mail" filters. Future work includes exploring the cognitive perspective on sentiment analysis.
The document describes the Naive Bayes classifier. It begins with background on probabilistic classification and generative models. It then covers probability basics and the probabilistic classification rule. It introduces Naive Bayes, making the independence assumption between attributes. The Naive Bayes algorithm is described for both the learning and testing phases. An example of classifying whether to play tennis is provided. Issues with Naive Bayes like violating independence and zero probabilities are discussed. Continuous attributes modeled with normal distributions are also covered. It concludes that Naive Bayes training is easy and fast while testing is straightforward, and it often performs competitively despite its independence assumption.
The document provides an overview of data warehousing concepts including:
1) A data warehouse is a subject-oriented collection of integrated data used to support management decisions. It contains current and historical data.
2) A data warehouse architecture typically includes source systems, a staging area, and presentation layer for querying and reporting.
3) Data marts are focused subsets of a data warehouse tailored for specific business units or departments. There are dependent, independent, and hybrid approaches to building data marts.
This document discusses online analytical processing (OLAP) and related concepts. It defines data mining, data warehousing, OLTP, and OLAP. It explains that a data warehouse integrates data from multiple sources and stores historical data for analysis. OLAP allows users to easily extract and view data from different perspectives. The document also discusses OLAP cube operations like slicing, dicing, drilling, and pivoting. It describes different OLAP architectures like MOLAP, ROLAP, and HOLAP and data warehouse schemas and architecture.
Bill Hayduk is the founder and CEO of QuerySurge, a software division that provides data integration and analytics solutions, with headquarters in New York; QuerySurge was founded in 1996 and has grown to serve Fortune 1000 customers through partnerships with technology companies and consulting firms. The document discusses the data and analytics marketplace and provides an overview of concepts like data warehousing, ETL, BI, data quality, data testing, big data, Hadoop, and NoSQL.
Various Applications of Data Warehouse.pptRafiulHasan19
The document discusses various applications of data warehousing. It begins by describing problems with traditional transactional systems and how data warehouses address these issues. It then defines key components of a data warehouse including the extraction, transformation, and loading of data from various sources. The document outlines how online analytical processing (OLAP) tools, metadata repositories, and data mining techniques analyze and explore the collected data. Finally, it weighs the benefits of a data warehouse against the costs of implementation and maintenance.
Types of database processing,OLTP VS Data Warehouses(OLAP), Subject-oriented
Integrated
Time-variant
Non-volatile,
Functionalities of Data Warehouse,Roll-Up(Consolidation),
Drill-down,
Slicing,
Dicing,
Pivot,
KDD Process,Application of Data Mining
This document discusses data warehousing concepts and technologies. It defines a data warehouse as a subject-oriented, integrated, non-volatile, and time-variant collection of data used to support management decision making. It describes the data warehouse architecture including extract-transform-load processes, OLAP servers, and metadata repositories. Finally, it outlines common data warehouse applications like reporting, querying, and data mining.
5 Things that Make Hadoop a Game Changer
Webinar by Elliott Cordo, Caserta Concepts
There is much hype and mystery surrounding Hadoop's role in analytic architecture. In this webinar, Elliott presented, in detail, the services and concepts that makes Hadoop a truly unique solution - a game changer for the enterprise. He talked about the real benefits of a distributed file system, the multi workload processing capabilities enabled by YARN, and the 3 other important things you need to know about Hadoop.
To access the recorded webinar, visit the event site: https://www.brighttalk.com/webcast/9061/131029
For more information the services and solutions that Caserta Concepts offers, please visit http://casertaconcepts.com/
The document provides an overview of database, big data, and data science concepts. It discusses topics such as database management systems (DBMS), data warehousing, OLTP vs OLAP, data mining, and the data science process. Key points include:
- DBMS are used to store and manage data in an organized way for use by multiple users. Data warehousing is used to consolidate data from different sources.
- OLTP systems are for real-time transactional systems, while OLAP systems are used for analysis and reporting of historical data.
- Data mining involves applying algorithms to large datasets to discover patterns and relationships. The data science process involves business understanding, data preparation, modeling, evaluation, and deployment
This document provides an overview of data warehousing and related concepts. It defines a data warehouse as a centralized database for analysis and reporting that stores current and historical data from multiple sources. The document describes key elements of data warehousing including Extract-Transform-Load (ETL) processes, multidimensional data models, online analytical processing (OLAP), and data marts. It also outlines advantages such as enhanced access and consistency, and disadvantages like time required for data extraction and loading.
OLAP (online analytical processing) allows users to easily extract and analyze data from different perspectives. It stores data in multidimensional databases to allow for complex queries. There are three main types of OLAP - relational, multidimensional, and hybrid. OLAP is used with data warehouses to enable analytics like data mining and decision making. It provides benefits over transactional systems by facilitating flexible analysis of integrated data over time.
Are you confused by Big Data? Get in touch with this new "black gold" and familiarize yourself with undiscovered insights through our complimentary introductory lesson on Big Data and Hadoop!
The document provides an overview of data warehousing, decision support, online analytical processing (OLAP), and data mining. It discusses what data warehousing is, how it can help organizations make better decisions by integrating data from various sources and making it available for analysis. It also describes OLAP as a way to transform warehouse data into meaningful information for interactive analysis, and lists some common OLAP operations like roll-up, drill-down, slice and dice, and pivot. Finally, it gives a brief introduction to data mining as the process of extracting patterns and relationships from data.
The New Frontier: Optimizing Big Data ExplorationInside Analysis
The Briefing Room with Dr. Robin Bloor and Cirro
Live Webcast on February 11, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=0ec1fa381886313cc06d841015c65898
As information ecosystems continue to expand, businesses are searching for ways to combine traditional analytics with a new source of insight: Big Data. But with data flooding in from all kinds of sources, fast access and performance at scale can easily become an issue. One effective approach for solving this challenge is data federation, a method that involves taking the analytical processing to the data, allowing streamlined access to multiple data sources without the expensive ETL overhead or building of semantic layers.
Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor as he explains how the prevalence of distributed data calls for a new approach to Big Data. He will be briefed by Mark Theissen of Cirro, who will tout his company’s Data Hub, a data federation solution that provides a single point of access to all enterprise data assets without excessive data movements, preprocessing or staging. He will discuss how data federation differs from virtualization and ETL approaches, and demonstrate how a Cirro deployment solves the analytics challenge of integrating data silos across the data center – and the cloud – using the BI tools you already have on your desktop for real-time distributed analytics.
Visit InsideAnlaysis.com for more information.
MSBI online training offered by Quontra Solutions with special features having Extensive Training will be in both MSBI Online Training and Placement. We help you in resume preparation and conducting Mock Interviews.
Emphasis is given on important topics that were required and mostly used in real time projects. Quontra Solutions is an Online Training Leader when it comes to high-end effective and efficient IT Training. We have always been and still are focusing on the key aspect which is providing utmost effective and competent training to both students and professionals who are eager to enrich their technical skills.
The document provides an overview of key concepts in data warehousing and business intelligence, including:
1) It defines data warehousing concepts such as the characteristics of a data warehouse (subject-oriented, integrated, time-variant, non-volatile), grain/granularity, and the differences between OLTP and data warehouse systems.
2) It discusses the evolution of business intelligence and key components of a data warehouse such as the source systems, staging area, presentation area, and access tools.
3) It covers dimensional modeling concepts like star schemas, snowflake schemas, and slowly and rapidly changing dimensions.
1. Data warehousing is a process of transforming data from operational systems into a centralized data repository optimized for analysis and reporting.
2. A data warehouse contains integrated, subject-oriented, non-volatile data used to support organizational decision making through queries and analysis.
3. The data warehouse is designed to provide historical, summarized views of data to business users to help answer questions and make decisions.
The Data Lake and Getting Buisnesses the Big Data Insights They NeedDunn Solutions Group
Do terms like "Data Lake" confuse you? You’re not alone. With all of the technology buzzwords flying around today, it can become a task to keep up with and clearly understand each of them. However a data lake is definitely something to dedicate the time to understand. Leveraging data lake technology, companies are finally able to keep all of their disparate information and streams of data in one secure location ready for consumption at any time – this includes structured, unstructured, and semi-structured data. For more information on our Big Data Consulting Services, don’t hesitate to visit us online at: http://bit.ly/2fvV5rR
Sentiment analysis aims to identify the orientation of opinions in text, whether positive, negative, or neutral. It draws from fields like cognitive science, natural language processing, and machine learning. Challenges include domain dependence, sarcasm, and contrasting with standard text categorization where word presence indicates category. Approaches include subjectivity detection using graph algorithms, sentiment lexicons capturing word sentiment, and scoring adjective-adverb combinations. Applications include review analysis, question answering, and developing "hate mail" filters. Future work includes exploring the cognitive perspective on sentiment analysis.
The document describes the Naive Bayes classifier. It begins with background on probabilistic classification and generative models. It then covers probability basics and the probabilistic classification rule. It introduces Naive Bayes, making the independence assumption between attributes. The Naive Bayes algorithm is described for both the learning and testing phases. An example of classifying whether to play tennis is provided. Issues with Naive Bayes like violating independence and zero probabilities are discussed. Continuous attributes modeled with normal distributions are also covered. It concludes that Naive Bayes training is easy and fast while testing is straightforward, and it often performs competitively despite its independence assumption.
Here are the key steps in building a spoken dialogue system:
1. Speech recognition: Convert speech to text using acoustic and language models.
2. Natural language understanding: Parse text and extract meaning using NLP techniques like parsing.
3. Dialogue management: Maintain context of conversation and decide system responses using dialogue state tracking and policy models.
4. Natural language generation: Convert system responses to text using templates and/or data-to-text models.
5. Speech synthesis: Convert text to speech using text-to-speech models.
The goal is smooth turn-taking conversation where the system understands the user's intent and responds appropriately through speech. Challenges include speech recognition errors,
The document discusses the ETL (Extract-Transform-Load) process used in data warehousing. The key steps of ETL are: 1) Capture/Extract which obtains data from source systems, 2) Scrub/Cleanse which cleans and formats the data, 3) Transform which converts the data to the structure of the data warehouse, and 4) Load/Index which loads the transformed data into the warehouse where it can be indexed. The ETL process results in data in the warehouse that is detailed, historical, normalized, comprehensive, timely, and quality controlled.
This lecture introduces a graduate-level course on machine learning for natural language processing (NLP). It covers the course overview, what is NLP, applications of NLP, and challenges in the field. Key NLP tasks like part-of-speech tagging, syntactic and dependency parsing, word sense disambiguation, and coreference resolution are discussed. Machine learning algorithms commonly used for NLP like classification, regression, and neural networks are also introduced. The lecture outlines the course content, assignments including paper summaries, presentations and a final project, and expectations for students.
The document summarizes a lecture on decision trees for classification. It introduces decision trees as a non-linear classification approach that learns directly from data representations. It then describes the basic ID3 algorithm for learning decision trees in a greedy top-down manner by choosing attributes that best split the data based on information gain at each step, recursively building the tree until reaching leaf nodes of single target classifications. The goal is to learn a compact tree representation of the training data.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
artificial intelligence and data science contents.pptxGauravCar
What is artificial intelligence? Artificial intelligence is the ability of a computer or computer-controlled robot to perform tasks that are commonly associated with the intellectual processes characteristic of humans, such as the ability to reason.
› ...
Artificial intelligence (AI) | Definitio
artificial intelligence and data science contents.pptx
05_Decision Support and OLAP.pdf
1. Decision Support,
Decision Support, Data
Data
Warehousing,
Warehousing, and OLAP
and OLAP
Data Warehousing
Data Warehousing
Warehousing,
Warehousing, and OLAP
and OLAP
2. Decision Support
Decision Support and OLAP
and OLAP
Information technology to help the knowledge
Information technology to help the knowledge
Information technology to help the knowledge
Information technology to help the knowledge
worker (executive, manager, analyst) make faster
worker (executive, manager, analyst) make faster
and better decisions
and better decisions.
.
3. Decision Support
Decision Support and OLAP
and OLAP
•
• What
What were the sales volumes by region and product
were the sales volumes by region and product
•
• What
What were the sales volumes by region and product
were the sales volumes by region and product
category for the last year?
category for the last year?
•
• How did the share price of computer manufacturers
How did the share price of computer manufacturers
correlate with quarterly profits over the past 10 years?
correlate with quarterly profits over the past 10 years?
•
• Which orders should we fill to maximize revenues?
Which orders should we fill to maximize revenues?
•
• Will a 10% discount increase sales volume sufficiently?
Will a 10% discount increase sales volume sufficiently?
•
• Which of two new medications will result in the best
Which of two new medications will result in the best
•
• Which of two new medications will result in the best
Which of two new medications will result in the best
outcome: higher recovery rate shorter hospital stay?
outcome: higher recovery rate shorter hospital stay?
On
On-
-Line
Line Analytical Processing (OLAP) is an element of
Analytical Processing (OLAP) is an element of
decision support
decision support systems
systems (DSS).
(DSS).
4. Evolution
Evolution
60’s: Batch reports
60’s: Batch reports
60’s: Batch reports
60’s: Batch reports
•
• hard to find and analyze information
hard to find and analyze information
•
• inflexible and expensive, reprogram every new request
inflexible and expensive, reprogram every new request
70’s: Terminal
70’s: Terminal-
-based DSS and EIS (executive information
based DSS and EIS (executive information
systems)
systems)
•
• still inflexible, not integrated with desktop tools
still inflexible, not integrated with desktop tools
80’s: Desktop data access and analysis tools
80’s: Desktop data access and analysis tools
80’s: Desktop data access and analysis tools
80’s: Desktop data access and analysis tools
•
• query tools, spreadsheets, GUIs
query tools, spreadsheets, GUIs
•
• easier to use, but only access operational databases
easier to use, but only access operational databases
90’s: Data warehousing with integrated OLAP engines
90’s: Data warehousing with integrated OLAP engines
and tools
and tools
5. OLTP vs. OLAP
OLTP vs. OLAP
OLTP (
OLTP (Online Transaction Processing
Online Transaction Processing
OLTP (
OLTP (Online Transaction Processing
Online Transaction Processing
Systems
Systems) is a direct transactional processing
) is a direct transactional processing
system (
system (insert, update, delete)
insert, update, delete) through the
through the
network.
network.
OLAP
OLAP (
(Online Analytical Processing
Online Analytical Processing
OLAP
OLAP (
(Online Analytical Processing
Online Analytical Processing
Systems
Systems)
) is a system built to help in
is a system built to help in
planning, problem solving, and decision
planning, problem solving, and decision
support.
support.
6. OLTP vs. OLAP
OLTP vs. OLAP
Item OLTP OLAP
Item OLTP OLAP
Data source Operational Data,
OLTP as the original
data.
Consolidated Data,
OLAP data from
OLTP.
Data function Controlling and
running the main tasks.
Planning, problem
solving and supporting
the decision.
Data showed Sustainable business From varies of
Data showed Sustainable business
process.
From varies of
business activities.
Query used Simple Query. Complex Queries.
7. OLTP vs. OLAP
OLTP vs. OLAP
Item OLTP OLAP
Item OLTP OLAP
Speed of access Faster. Depends on the data
involved, it could be faster
using indexes.
Space required Smaller. Larger, needs more
indexing other than OLTP.
Database Design Normalized with
many tables.
De-normalized with less
table and using star /
snowflakes schemas.
snowflakes schemas.
User IT Professional Knowledge worker
8. Data Warehouse
Data Warehouse
A decision support database that is maintained
A decision support database that is maintained
A decision support database that is maintained
A decision support database that is maintained
separately from the organization’s operational
separately from the organization’s operational
databases.
databases.
A data warehouse is a
A data warehouse is a
•
• subject
subject-
-oriented,
oriented,
•
• integrated,
integrated,
•
• integrated,
integrated,
•
• time
time-
-varying,
varying,
•
• non
non-
-volatile
volatile
collection of data that is used primarily in
collection of data that is used primarily in
organizational decision making
organizational decision making
9. Why Separate Data Warehouse?
Why Separate Data Warehouse?
Performance
Performance
Performance
Performance
•
• Operational database designed
Operational database designed tuned for
tuned for
known
known transactions
transactions workloads.
workloads.
•
• Complex OLAP queries would degrade
Complex OLAP queries would degrade
performance.
performance. For
For operational transactions.
operational transactions.
•
• Special data organization, access
Special data organization, access
•
• Special data organization, access
Special data organization, access
implementation methods needed for
implementation methods needed for
multidimensional views queries.
multidimensional views queries.
10. Why Separate Data Warehouse?
Why Separate Data Warehouse?
Function
Function
Function
Function
•
• Missing data: Decision support requires historical
Missing data: Decision support requires historical
data, which
data, which operational DB do
operational DB do not typically
not typically
maintain.
maintain.
•
• Data consolidation: Decision support requires
Data consolidation: Decision support requires
consolidation (aggregation, summarization) of data
consolidation (aggregation, summarization) of data
consolidation (aggregation, summarization) of data
consolidation (aggregation, summarization) of data
from many heterogeneous sources:
from many heterogeneous sources: operational DB,
operational DB,
external sources.
external sources.
•
• Data quality: Different sources typically use
Data quality: Different sources typically use
inconsistent data representations, codes, and
inconsistent data representations, codes, and
formats which have to be reconciled.
formats which have to be reconciled.
11. Data Warehousing Market
Data Warehousing Market
Hardware: servers, storage, clients
Hardware: servers, storage, clients
Hardware: servers, storage, clients
Hardware: servers, storage, clients
Warehouse DBMs
Warehouse DBMs
Tools
Tools
Market growing from
Market growing from
•
• $2B in 1995 to $8 B in 1998 (Meta Group)
$2B in 1995 to $8 B in 1998 (Meta Group)
•
• 1.5B today to $6.9B in 1999 (Gartner Group)
1.5B today to $6.9B in 1999 (Gartner Group)
Systems integration Consulting
Systems integration Consulting
Already deployed in many industries: manufacturing,
Already deployed in many industries: manufacturing,
retail, financial, insurance, transportation, telecom.,
retail, financial, insurance, transportation, telecom.,
utilities, healthcare.
utilities, healthcare.
12. Data Warehousing Architecture
Data Warehousing Architecture
OLAP servers
OLAP servers
Monitoring
Monitoring Administration
Administration
Extract
Extract
Transform
Transform
Load
Load
Refresh
Refresh
External
External
Sources
Sources
Serve
Serve
OLAP servers
OLAP servers
Analysis
Analysis
Query/
Query/
Reporting
Reporting
Metadata
Metadata
Repository
Repository
Refresh
Refresh
Data Marts
Data Marts
Serve
Serve
Data
Data
Mining
Mining
Operational
Operational
DB
DB
13. Three
Three-
-Tier Architecture
Tier Architecture
Warehouse database server
Warehouse database server
Warehouse database server
Warehouse database server
•
• Almost always a relational DBMS; rarely flat files
Almost always a relational DBMS; rarely flat files
OLAP servers
OLAP servers
•
• Relational OLAP (ROLAP): extended relational DBMS that maps
Relational OLAP (ROLAP): extended relational DBMS that maps
operations on multidimensional data to standard relational
operations on multidimensional data to standard relational
operations.
operations.
•
• Multidimensional OLAP (MOLAP): special purpose server that
Multidimensional OLAP (MOLAP): special purpose server that
•
• Multidimensional OLAP (MOLAP): special purpose server that
Multidimensional OLAP (MOLAP): special purpose server that
directly implements multidimensional data and operations.
directly implements multidimensional data and operations.
Clients
Clients
•
• Query and reporting tools.
Query and reporting tools.
•
• Analysis tools
Analysis tools
•
• Data mining tools (e.g., trend analysis, prediction)
Data mining tools (e.g., trend analysis, prediction)
14. Design Operational Process
Design Operational Process
Define architecture. Do capacity planning.
Define architecture. Do capacity planning.
Define architecture. Do capacity planning.
Define architecture. Do capacity planning.
Integrate db and OLAP servers, storage and client tools.
Integrate db and OLAP servers, storage and client tools.
Design warehouse schema, views.
Design warehouse schema, views.
Design physical warehouse organization: data placement,
Design physical warehouse organization: data placement,
partitioning, access methods.
partitioning, access methods.
Connect sources: gateways, ODBC drivers, wrappers.
Connect sources: gateways, ODBC drivers, wrappers.
Connect sources: gateways, ODBC drivers, wrappers.
Connect sources: gateways, ODBC drivers, wrappers.
Design implement scripts for data extract, load refresh.
Design implement scripts for data extract, load refresh.
Define metadata and populate repository.
Define metadata and populate repository.
Design implement end
Design implement end-
-user applications.
user applications.
Roll out warehouse and applications.
Roll out warehouse and applications.
Monitor the warehouse.
Monitor the warehouse.
15. OLAP for Decision Support
OLAP for Decision Support
Goal of OLAP is to support ad
Goal of OLAP is to support ad-
-hoc querying for the
hoc querying for the
Goal of OLAP is to support ad
Goal of OLAP is to support ad-
-hoc querying for the
hoc querying for the
business analyst
business analyst
Business analysts are familiar with spreadsheets
Business analysts are familiar with spreadsheets
Extend spreadsheet analysis model to work with
Extend spreadsheet analysis model to work with
warehouse data
warehouse data
•
• Large data set
Large data set
•
• Semantically enriched to understand business terms (e.g., time,
Semantically enriched to understand business terms (e.g., time,
•
• Semantically enriched to understand business terms (e.g., time,
Semantically enriched to understand business terms (e.g., time,
geography)
geography)
•
• Combined with reporting features
Combined with reporting features
Multidimensional
Multidimensional view of data is the foundation of OLAP
view of data is the foundation of OLAP
16. Multidimensional Data Model
Multidimensional Data Model
Database is a set of
Database is a set of facts
facts (points) in a
(points) in a
Database is a set of
Database is a set of facts
facts (points) in a
(points) in a
multidimensional space
multidimensional space
A fact has a
A fact has a measure
measure dimension
dimension
•
• quantity that is analyzed, e.g., sale, budget
quantity that is analyzed, e.g., sale, budget
A set of
A set of dimensions
dimensions on which data is
on which data is
A set of
A set of dimensions
dimensions on which data is
on which data is
analyzed
analyzed
•
• e.g. , store, product, date associated with a sale
e.g. , store, product, date associated with a sale
amount
amount
17. Multidimensional Data Model
Multidimensional Data Model
Dimensions
Dimensions form a sparsely populated
form a sparsely populated
Dimensions
Dimensions form a sparsely populated
form a sparsely populated
coordinate system
coordinate system
Each dimension has a set of
Each dimension has a set of attributes
attributes
•
• e.g., owner city and county of store
e.g., owner city and county of store
Attributes of a dimension may be related by
Attributes of a dimension may be related by
Attributes of a dimension may be related by
Attributes of a dimension may be related by
partial order
partial order
•
• Hierarchy
Hierarchy: e.g., street county city
: e.g., street county city
•
• Lattice
Lattice: e.g., date monthyear,
: e.g., date monthyear,
dateweekyear
dateweekyear
19. Operations in Multidimensional
Operations in Multidimensional
Data Model
Data Model
Aggregation (
Aggregation (roll
roll-
-up
up)
)
Aggregation (
Aggregation (roll
roll-
-up
up)
)
•
• dimension reduction: e.g., total sales by city
dimension reduction: e.g., total sales by city
•
• summarization over aggregate hierarchy: e.g., total
summarization over aggregate hierarchy: e.g., total
sales by city and year
sales by city and year -
- total sales by region and by
total sales by region and by
year
year
Selection (
Selection (slice
slice) defines a subcube
) defines a subcube
•
• e.g., sales where city = Palo Alto and date = 1/15/96
e.g., sales where city = Palo Alto and date = 1/15/96
Navigation to detailed data (
Navigation to detailed data (drill
drill-
-down
down)
)
•
• e.g., (sales
e.g., (sales -
- expense) by city, top 3% of cities by
expense) by city, top 3% of cities by
average income
average income
Visualization Operations (e.g., Pivot)
Visualization Operations (e.g., Pivot)
20. A Visual Operation: Pivot (Rotate)
A Visual Operation: Pivot (Rotate)
10
10
47
47
30
30
Juice
Juice
Cola
Cola
Milk
Milk 30
30
12
12
Milk
Milk
Cream
Cream
3/1
3/1 3/2 3/3
3/2 3/3 3/4
3/4
Date
Date
Product
Product
21. Approaches to OLAP Servers
Approaches to OLAP Servers
Relational OLAP (ROLAP)
Relational OLAP (ROLAP)
Relational OLAP (ROLAP)
Relational OLAP (ROLAP)
•
• Relational and Specialized Relational DBMS
Relational and Specialized Relational DBMS
to store and manage warehouse data
to store and manage warehouse data
•
• OLAP middleware to support missing pieces
OLAP middleware to support missing pieces
–
– Optimize for each DBMS backend
Optimize for each DBMS backend
–
– Aggregation Navigation Logic
Aggregation Navigation Logic
–
– Aggregation Navigation Logic
Aggregation Navigation Logic
–
– Additional tools and services
Additional tools and services
•
• Example:
Example: Microstrategy
Microstrategy,
, MetaCube
MetaCube
(Informix
(Informix)
)
22. Approaches to OLAP Servers
Approaches to OLAP Servers
Multidimensional
Multidimensional OLAP (MOLAP)
OLAP (MOLAP)
Multidimensional
Multidimensional OLAP (MOLAP)
OLAP (MOLAP)
•
• Array
Array-
-based storage structures
based storage structures
•
• Direct access to array data structures
Direct access to array data structures
•
• Example:
Example: Essbase
Essbase (Arbor),
(Arbor), Accumate
Accumate
(
(Kenan
Kenan)
)
(
(Kenan
Kenan)
)
Domain
Domain-
-specific enrichment
specific enrichment
23. Relational DBMS as Warehouse
Relational DBMS as Warehouse
Server
Server
Schema design
Schema design
Schema design
Schema design
Specialized scan, indexing and join techniques
Specialized scan, indexing and join techniques
Handling of aggregate views (querying and
Handling of aggregate views (querying and
materialization)
materialization)
Supporting query language extensions beyond
Supporting query language extensions beyond
SQL
SQL
SQL
SQL
Complex query processing and optimization
Complex query processing and optimization
Data partitioning and parallelism
Data partitioning and parallelism
24. Warehouse Database Schema
Warehouse Database Schema
ER design techniques not appropriate
ER design techniques not appropriate
ER design techniques not appropriate
ER design techniques not appropriate
Design should reflect multidimensional
Design should reflect multidimensional
view
view
•
• Star Schema
Star Schema
•
• Snowflake Schema
Snowflake Schema
•
• Snowflake Schema
Snowflake Schema
•
• Fact Constellation Schema
Fact Constellation Schema
25. Example of a Star Schema
Example of a Star Schema
Order
Order
Product
Product
Order No
Order No
Order Date
Order Date
SalespersonID
SalespersonID
SalespersonName
SalespersonName
City
City
Quota
Quota
OrderNO
OrderNO
SalespersonID
SalespersonID
CustomerNO
CustomerNO
ProdNo
ProdNo
ProductNO
ProductNO
ProdName
ProdName
ProdDescr
ProdDescr
Category
Category
CategoryDescription
CategoryDescription
UnitPrice
UnitPrice
Salesperson
Salesperson
Date
Date
Product
Product
Fact Table
Fact Table
Customer No
Customer No
Customer Name
Customer Name
Customer Address
Customer Address
City
City
ProdNo
ProdNo
DateKey
DateKey
CityName
CityName
Quantity
Quantity
TotalPrice
DateKey
DateKey
Date
Date
CityName
CityName
State
State
Country
Country
Customer
Customer
City
City
Date
Date
26. Star Schema
Star Schema
A single fact table and a single table for each dimension
A single fact table and a single table for each dimension
A single fact table and a single table for each dimension
A single fact table and a single table for each dimension
Every fact points to one tuple in each of the dimensions
Every fact points to one tuple in each of the dimensions
and has additional attributes
and has additional attributes
Does not capture hierarchies directly
Does not capture hierarchies directly
Generated keys are used for performance and maintenance
Generated keys are used for performance and maintenance
reasons
reasons
Fact constellation: Multiple Fact tables that share many
Fact constellation: Multiple Fact tables that share many
dimension tables
dimension tables
•
• Example: Projected expense and the actual expense may share
Example: Projected expense and the actual expense may share
dimensional tables
dimensional tables
27. Example of a Snowflake Schema
Example of a Snowflake Schema
Order
Order
Product
Product
Category
Category
Order No
Order No
Order Date
Order Date
Customer No
Customer No
Customer Name
Customer Name
Customer
Customer
Address
Address
OrderNO
OrderNO
SalespersonID
SalespersonID
CustomerNO
CustomerNO
ProdNo
ProdNo
ProductNO
ProductNO
ProdName
ProdName
ProdDescr
ProdDescr
CategoryName
CategoryName
Category
Category
UnitPrice
UnitPrice
Customer
Customer
Date
Date
Product
Product
Fact Table
Fact Table
CategoryName
CategoryName
CategoryDescr
CategoryDescr
Category
Category
Month
Month
City
City
SalespersonID
SalespersonID
SalespersonName
SalespersonName
City
City
Quota
Quota
ProdNo
ProdNo
DateKey
DateKey
CityName
CityName
Quantity
Quantity
TotalPrice
DateKey
DateKey
Date
Date
Month
Month
CityName
CityName
StateName
StateName
Salesperson
Salesperson
City
City
Date
Date
Month
Month
Year
Year
Year
Year
StateName
StateName
Country
Country
State
State
Month
Month
Year
Year
28. Snowflake Schema
Snowflake Schema
Represent dimensional hierarchy directly by
Represent dimensional hierarchy directly by
Represent dimensional hierarchy directly by
Represent dimensional hierarchy directly by
normalizing the dimension tables
normalizing the dimension tables
Easy to maintain
Easy to maintain
Saves storage, but is alleged that it reduces
Saves storage, but is alleged that it reduces
effectiveness of browsing (Kimball)
effectiveness of browsing (Kimball)
effectiveness of browsing (Kimball)
effectiveness of browsing (Kimball)