This document presents a case study on data warehousing for an ABC Retail chain operating in the US, Canada, and Mexico. It discusses setting up a data warehouse to analyze key performance indicators related to sales, promotions, and customer preferences. Dimensions such as customers, products, dates, and stores are defined. Fact tables for sales and procurement transactions are created. A star schema links the dimensions and facts. Visualizations and analyses are proposed to measure metrics like sales by store type, product sales by store, the effect of promotions on sales, and the impact of customer professions on buying patterns.
Implemented Data warehouse on “Retail Stores of five states of USA” by using 3 different data sources including structured and unstructured using SSIS, SSAS and Power BI.
The document describes the process of building dimensional data warehouses for three different companies - ZAGI Retail Company, City Police Department, and Big Z Inc. It provides details of the source data, dimensional models created with star schemas, and SQL insert statements to populate the fact and dimension tables. The dimensional models are analyzed by date, product, customer, and other attributes. Aggregated fact tables are also created to summarize daily sales or revenue amounts.
Here are the steps to create the RFQ:
1. Enter your purchase requisition number
2. Select all items
3. Click "Adopt" to copy item details to RFQ
4. Click "Save" to save the RFQ
This will create the RFQ with the item details copied from the purchase requisition. Now you can generate quotations by adding vendors.
The document discusses setting up forward and reverse pricing scenarios in SAP. It provides details on:
1) The sender's client needs both forward and reverse pricing scenarios, and the forward scenario is working but they are stuck on replicating it in reverse.
2) An overview of the standard forward pricing procedure including condition types and pricing sequence.
3) The request is to build a reverse procedure where the user enters the total price of Rs. 106.20 and the system calculates the base price, VAT, and service tax values.
4) Suggestions are requested on how to set up the reverse pricing scenario.
What specific role of people using developing and managing is seleact an appr...Muhammad Tahir Mehmood
This document discusses the roles of people in developing and managing information systems, using FrontAccounting as an example. It provides an overview of FrontAccounting, describing it as open source web-based accounting software that allows for double-entry accounting and integrated business processes. It then discusses the various modules in FrontAccounting for transactions, inquiries and reports, and maintenance in areas like sales, purchases, inventory, and manufacturing. Finally, it outlines some of the key roles people may have in using FrontAccounting, such as end users, data entry clerks, managers, programmers and database administrators.
The document outlines several key concepts in SAP Sales and Distribution including:
1) Sales organizations, distribution channels, divisions, and sales areas are the primary organizational units used to define responsibilities and group products. Each document is assigned to a specific sales area.
2) Master data such as customer, material, pricing, and output masters are critical for sales documents. Customer masters contain detailed contact and account information.
3) The sales process in SAP begins with inquiries and quotations and progresses through orders, deliveries, and billing. Inventory availability, shipping, picking, and billing are managed through this process.
The document discusses purchasing info records (PIRs) in SAP, which contain vendor and material information to help buyers determine pricing and supplier options. PIRs can be created at the plant or purchasing organization level and include fields for vendor details, material descriptions, pricing conditions, order quantities and delivery terms. The document outlines how to create, change, display and report on PIRs within SAP.
The document provides an overview of key concepts in SAP SD (Sales and Distribution) module including:
1) Organizational structure configuration such as company code, business area, plant, division, sales organization, distribution channel, and storage location.
2) The differences between cash sales and rush orders including order types, delivery timing, billing, and availability checks.
3) Master data such as customer, material, and condition masters that are used in sales documents.
4) The standard sales process flow from inquiry to billing.
Implemented Data warehouse on “Retail Stores of five states of USA” by using 3 different data sources including structured and unstructured using SSIS, SSAS and Power BI.
The document describes the process of building dimensional data warehouses for three different companies - ZAGI Retail Company, City Police Department, and Big Z Inc. It provides details of the source data, dimensional models created with star schemas, and SQL insert statements to populate the fact and dimension tables. The dimensional models are analyzed by date, product, customer, and other attributes. Aggregated fact tables are also created to summarize daily sales or revenue amounts.
Here are the steps to create the RFQ:
1. Enter your purchase requisition number
2. Select all items
3. Click "Adopt" to copy item details to RFQ
4. Click "Save" to save the RFQ
This will create the RFQ with the item details copied from the purchase requisition. Now you can generate quotations by adding vendors.
The document discusses setting up forward and reverse pricing scenarios in SAP. It provides details on:
1) The sender's client needs both forward and reverse pricing scenarios, and the forward scenario is working but they are stuck on replicating it in reverse.
2) An overview of the standard forward pricing procedure including condition types and pricing sequence.
3) The request is to build a reverse procedure where the user enters the total price of Rs. 106.20 and the system calculates the base price, VAT, and service tax values.
4) Suggestions are requested on how to set up the reverse pricing scenario.
What specific role of people using developing and managing is seleact an appr...Muhammad Tahir Mehmood
This document discusses the roles of people in developing and managing information systems, using FrontAccounting as an example. It provides an overview of FrontAccounting, describing it as open source web-based accounting software that allows for double-entry accounting and integrated business processes. It then discusses the various modules in FrontAccounting for transactions, inquiries and reports, and maintenance in areas like sales, purchases, inventory, and manufacturing. Finally, it outlines some of the key roles people may have in using FrontAccounting, such as end users, data entry clerks, managers, programmers and database administrators.
The document outlines several key concepts in SAP Sales and Distribution including:
1) Sales organizations, distribution channels, divisions, and sales areas are the primary organizational units used to define responsibilities and group products. Each document is assigned to a specific sales area.
2) Master data such as customer, material, pricing, and output masters are critical for sales documents. Customer masters contain detailed contact and account information.
3) The sales process in SAP begins with inquiries and quotations and progresses through orders, deliveries, and billing. Inventory availability, shipping, picking, and billing are managed through this process.
The document discusses purchasing info records (PIRs) in SAP, which contain vendor and material information to help buyers determine pricing and supplier options. PIRs can be created at the plant or purchasing organization level and include fields for vendor details, material descriptions, pricing conditions, order quantities and delivery terms. The document outlines how to create, change, display and report on PIRs within SAP.
The document provides an overview of key concepts in SAP SD (Sales and Distribution) module including:
1) Organizational structure configuration such as company code, business area, plant, division, sales organization, distribution channel, and storage location.
2) The differences between cash sales and rush orders including order types, delivery timing, billing, and availability checks.
3) Master data such as customer, material, and condition masters that are used in sales documents.
4) The standard sales process flow from inquiry to billing.
• Developed and Analysed Data warehouse Using SSIS ETL tool, SSDT, SQL server
• Provided Analysed Quarterly Report Using SSRS of Total sales, Total Revenue, Predicted Future sales, topmost selling products, top discounted product.
• Used Performance tuning to fetch rows faster from database and performed data visualization using R-studio and Neo-4j.
Pivot Tables and Beyond Data Analysis in Excel 2013 - Course Technology Compu...Cengage Learning
Pivot Tables and Beyond Data Analysis in Excel 2013 - Course Technology Computing Conference
Presenter: Patrick Carey, Cengage Learning Author
Excel is sometimes called the most popular "database" in the world, not because it's a database but because it makes data so accessible that users often turn to spreadsheets for data entry. Yet for all that, Excel's tools for data analysis and modeling remain largely untapped by the average user. In this, pivot tables may be the most powerful and least utilized tool for data exploration. In this presentation we'll examine some of the new enhancements to pivot tables introduced in Excel 2013. We'll examine how to set up relationships using the Excel Data Model to summarize information across multiple data tables. And then we'll go beyond, exploring the data modeling and data visualizing tools provided by the PowerPivot and Power View add-ins, interpreting data not just numerically but through visual imagery, charts, and interactive maps.
Webinar: Designing a schema for a Data WarehouseFederico Razzoli
Are you new to data warehouses (DWH)? Do you need to check whether your data warehouse follows the best practices for a good design? In both cases, this webinar is for you.
A data warehouse is a central relational database that contains all measurements about a business or an organisation. This data comes from a variety of heterogeneous data sources, which includes databases of any type that back the applications used by the company, data files exported by some applications, or APIs provided by internal or external services.
But designing a data warehouse correctly is a hard task, which requires gathering information about the business processes that need to be analysed in the first place. These processes must be translated into so-called star schemas, which means, denormalised databases where each table represents a dimension or facts.
We will discuss these topics:
- How to gather information about a business;
- Understanding dictionaries and how to identify business entities;
- Dimensions and facts;
- Setting a table granularity;
- Types of facts;
- Types of dimensions;
- Snowflakes and how to avoid them;
- Expanding existing dimensions and facts.
The document discusses the use of business intelligence (BI) in the fast-moving consumer goods (FMCG) and retail industries. It outlines key performance indicators and frameworks for BI systems in retail. It also describes data modeling approaches for retail scenarios and discusses how BI can help address challenges and questions in areas like inventory, pricing, promotions and customer analytics. Major trends include increased competition and expectations, and the need for greater organizational alignment through BI.
The document discusses dimensional modeling best practices for a retail sales case study. It outlines a four-step process: 1) select the business process, 2) declare the grain, 3) choose dimensions, 4) identify facts. For the retail case, the process modeled is point-of-sale sales, with a grain of individual transactions. Key dimensions are date, product, store, and promotion. Facts include sales quantity, price, amount, and costs. The document also discusses design considerations like degenerate dimensions, extensibility, and surrogate keys.
This document summarizes key aspects of using business intelligence (BI) systems in the fast-moving consumer goods (FMCG) and retail industries. It outlines major changes in these industries, key performance indicators (KPIs) used, an example data model for a retail scenario, and how BI can help address challenges around customer analytics, supply chain management, and operations optimization. Major trends are discussed, including the need for agile infrastructure to support dynamic business demands and enhanced customer experiences.
ales forecasting is the process of predicting future sales for a business or organization. Accurate sales forecasting can help businesses make informed decisions about inventory management, staffing, and overall strategy. Sparkflows provides a range of tools and frameworks for building machine learning models that can be used for sales forecasting.
To build a sales forecasting model using Sparkflows, you would typically follow these steps:
Data preparation: Gather and preprocess your sales data, which typically includes historical sales data, seasonality data, and other relevant data such as pricing and promotions.
Feature selection: Select the most relevant features that are likely to impact sales, such as time of year, marketing campaigns, and changes in pricing.
Model selection: Choose a machine learning algorithm to use for predicting sales, such as linear regression, time series forecasting, or neural networks.
Model training: Use historical sales data to train your model on examples of past sales trends.
Model evaluation: Evaluate the performance of your model using a set of metrics such as mean absolute error (MAE), mean squared error (MSE), and root mean squared error (RMSE).
Deployment: Once you have a model that performs well, you can deploy it to your production environment to make predictions on new sales data.
Sparkflows provides a variety of tools and frameworks that can help with each of these steps, including data preprocessing and transformation tools, feature selection algorithms, and machine learning libraries such as Apache Spark MLlib. Additionally, Sparkflows allows you to automate the entire workflow using pipelines, which can be useful for large-scale, complex projects.
One of the key challenges in sales forecasting is accounting for seasonality, or the regular patterns of sales that occur over time. Sparkflows provides several tools that can help with this, such as time series forecasting models and seasonal decomposition algorithms that can identify and account for seasonal trends in your sales data.
Another important aspect of sales forecasting is feature selection, or identifying which variables are likely to have the most impact on sales. Sparkflows provides a range of feature selection algorithms that can help you identify the most relevant variables, such as correlation-based feature selection and recursive feature elimination.
Overall, Sparkflows provides a comprehensive set of tools and frameworks for building accurate sales forecasting models, and can be a powerful tool for businesses looking to make informed decisions based on their sales data.
For more details visit - https://www.sparkflows.io/
This document discusses using Ruby to perform multidimensional data analysis on relational databases. It introduces Mondrian, an open-source OLAP engine that allows for multidimensional analysis on top of SQL databases using the MDX query language. A new Ruby gem called mondrian-olap will integrate Mondrian and provide a Ruby DSL and ActiveRecord-like query interface for defining OLAP schemas and performing analytical queries on relational data in a simpler way than SQL. Examples show how to write multidimensional queries in MDX and the Ruby interface to analyze sales data across dimensions like time, products, and customers.
Dimensional data modeling is a technique for database design intended to support analysis and reporting. It contains dimension tables that provide context about the business and fact tables that contain measures. Dimension tables describe attributes and may include hierarchies, while fact tables contain measurable events linked to dimensions. When designing a dimensional model, the business process, grain, dimensions, and facts are identified. Star and snowflake schemas are common types that differ in normalization of the dimensions. Slowly changing dimensions also must be accounted for.
The document describes the development and implementation of an Integrated Analytical Model (IAM) at Konica Minolta Ukraine to improve business decision making. Key points:
- The IAM integrated data from various ERP systems and financial reports into a multidimensional model to allow comprehensive analysis of KPIs across business areas over time.
- The model linked dimensions like products, customers, and time to financial and operational metrics, enabling analysis of factors like customer profitability by segment.
- Results are communicated through SharePoint reports and Excel pivot tables, providing managers targeted insights while allowing power users flexible ad-hoc analysis.
- The IAM created a "single version of truth" and
This document discusses using machine learning for demand forecasting in supply chain management. It begins by outlining problems with traditional forecasting methods and high errors affecting business decisions. It then proposes using machine learning algorithms that can learn from large datasets to more accurately model demand. Key steps discussed include collecting internal and external data, pre-processing data, building and comparing regression models, and developing a technical architecture to provide ongoing demand forecasting capabilities. The goals are to reduce errors, optimize inventory levels and pricing, and improve profits.
The document discusses data warehousing and the star schema. It defines a data warehouse as a repository of integrated information available for queries and analysis. The data comes from heterogeneous sources and can be queried together. It describes how a star schema organizes data into a central fact table surrounded by dimension tables. The fact table contains keys linking to attributes in the dimension tables. Star queries are processed by first using bitmap indexes on the fact table keys to retrieve relevant rows, then joining the results to the dimension tables.
This document describes a data warehouse and business intelligence project for analyzing Starbucks store data. It discusses extracting data from various structured, semi-structured, and unstructured sources, transforming the data using SQL and R, and loading it into a star schema data warehouse with fact and dimension tables. The data warehouse is then used for business queries and analysis in Tableau, with case studies examining city revenue, visitor and beverage sales by city, and city ratings based on food and beverage counts. The analysis finds that New York City generally has the highest revenue, visitor counts, and ratings.
Applying machine learning to Kaggle data set to predict which customers are most likely to become customers. Random Forest column importance graph is helpful to prioritize the best segments to target.
Data Warehousing and Business Intelligence is one of the hottest skills today, and is the cornerstone for reporting, data science, and analytics. This course teaches the fundamentals with examples plus a project to fully illustrate the concepts.
This document discusses questions to ask when preparing to analyze a company's data from multiple sources. It addresses determining data sources and structure, cleaning data, identifying key performance indicators and metrics to analyze, understanding user needs, and planning how analysis results will be used and reported. The goal is to map the data lifecycle, ensure quality, and select the right analysis approach and tools to provide insights for business decisions.
The document summarizes key concepts in dimensional data warehouse design based on Ralph Kimball's approach. It discusses the four-step design process of selecting a business process, declaring the grain, choosing dimensions, and identifying facts. It then provides an example applying these steps to design a dimensional schema for a retail sales data warehouse using data from a point-of-sale system, with dimensions for date, product, store, and promotion. It also covers topics like additive/non-additive facts, degenerate dimensions, extensibility, and avoiding normalization of dimension tables.
• Developed and Analysed Data warehouse Using SSIS ETL tool, SSDT, SQL server
• Provided Analysed Quarterly Report Using SSRS of Total sales, Total Revenue, Predicted Future sales, topmost selling products, top discounted product.
• Used Performance tuning to fetch rows faster from database and performed data visualization using R-studio and Neo-4j.
Pivot Tables and Beyond Data Analysis in Excel 2013 - Course Technology Compu...Cengage Learning
Pivot Tables and Beyond Data Analysis in Excel 2013 - Course Technology Computing Conference
Presenter: Patrick Carey, Cengage Learning Author
Excel is sometimes called the most popular "database" in the world, not because it's a database but because it makes data so accessible that users often turn to spreadsheets for data entry. Yet for all that, Excel's tools for data analysis and modeling remain largely untapped by the average user. In this, pivot tables may be the most powerful and least utilized tool for data exploration. In this presentation we'll examine some of the new enhancements to pivot tables introduced in Excel 2013. We'll examine how to set up relationships using the Excel Data Model to summarize information across multiple data tables. And then we'll go beyond, exploring the data modeling and data visualizing tools provided by the PowerPivot and Power View add-ins, interpreting data not just numerically but through visual imagery, charts, and interactive maps.
Webinar: Designing a schema for a Data WarehouseFederico Razzoli
Are you new to data warehouses (DWH)? Do you need to check whether your data warehouse follows the best practices for a good design? In both cases, this webinar is for you.
A data warehouse is a central relational database that contains all measurements about a business or an organisation. This data comes from a variety of heterogeneous data sources, which includes databases of any type that back the applications used by the company, data files exported by some applications, or APIs provided by internal or external services.
But designing a data warehouse correctly is a hard task, which requires gathering information about the business processes that need to be analysed in the first place. These processes must be translated into so-called star schemas, which means, denormalised databases where each table represents a dimension or facts.
We will discuss these topics:
- How to gather information about a business;
- Understanding dictionaries and how to identify business entities;
- Dimensions and facts;
- Setting a table granularity;
- Types of facts;
- Types of dimensions;
- Snowflakes and how to avoid them;
- Expanding existing dimensions and facts.
The document discusses the use of business intelligence (BI) in the fast-moving consumer goods (FMCG) and retail industries. It outlines key performance indicators and frameworks for BI systems in retail. It also describes data modeling approaches for retail scenarios and discusses how BI can help address challenges and questions in areas like inventory, pricing, promotions and customer analytics. Major trends include increased competition and expectations, and the need for greater organizational alignment through BI.
The document discusses dimensional modeling best practices for a retail sales case study. It outlines a four-step process: 1) select the business process, 2) declare the grain, 3) choose dimensions, 4) identify facts. For the retail case, the process modeled is point-of-sale sales, with a grain of individual transactions. Key dimensions are date, product, store, and promotion. Facts include sales quantity, price, amount, and costs. The document also discusses design considerations like degenerate dimensions, extensibility, and surrogate keys.
This document summarizes key aspects of using business intelligence (BI) systems in the fast-moving consumer goods (FMCG) and retail industries. It outlines major changes in these industries, key performance indicators (KPIs) used, an example data model for a retail scenario, and how BI can help address challenges around customer analytics, supply chain management, and operations optimization. Major trends are discussed, including the need for agile infrastructure to support dynamic business demands and enhanced customer experiences.
ales forecasting is the process of predicting future sales for a business or organization. Accurate sales forecasting can help businesses make informed decisions about inventory management, staffing, and overall strategy. Sparkflows provides a range of tools and frameworks for building machine learning models that can be used for sales forecasting.
To build a sales forecasting model using Sparkflows, you would typically follow these steps:
Data preparation: Gather and preprocess your sales data, which typically includes historical sales data, seasonality data, and other relevant data such as pricing and promotions.
Feature selection: Select the most relevant features that are likely to impact sales, such as time of year, marketing campaigns, and changes in pricing.
Model selection: Choose a machine learning algorithm to use for predicting sales, such as linear regression, time series forecasting, or neural networks.
Model training: Use historical sales data to train your model on examples of past sales trends.
Model evaluation: Evaluate the performance of your model using a set of metrics such as mean absolute error (MAE), mean squared error (MSE), and root mean squared error (RMSE).
Deployment: Once you have a model that performs well, you can deploy it to your production environment to make predictions on new sales data.
Sparkflows provides a variety of tools and frameworks that can help with each of these steps, including data preprocessing and transformation tools, feature selection algorithms, and machine learning libraries such as Apache Spark MLlib. Additionally, Sparkflows allows you to automate the entire workflow using pipelines, which can be useful for large-scale, complex projects.
One of the key challenges in sales forecasting is accounting for seasonality, or the regular patterns of sales that occur over time. Sparkflows provides several tools that can help with this, such as time series forecasting models and seasonal decomposition algorithms that can identify and account for seasonal trends in your sales data.
Another important aspect of sales forecasting is feature selection, or identifying which variables are likely to have the most impact on sales. Sparkflows provides a range of feature selection algorithms that can help you identify the most relevant variables, such as correlation-based feature selection and recursive feature elimination.
Overall, Sparkflows provides a comprehensive set of tools and frameworks for building accurate sales forecasting models, and can be a powerful tool for businesses looking to make informed decisions based on their sales data.
For more details visit - https://www.sparkflows.io/
This document discusses using Ruby to perform multidimensional data analysis on relational databases. It introduces Mondrian, an open-source OLAP engine that allows for multidimensional analysis on top of SQL databases using the MDX query language. A new Ruby gem called mondrian-olap will integrate Mondrian and provide a Ruby DSL and ActiveRecord-like query interface for defining OLAP schemas and performing analytical queries on relational data in a simpler way than SQL. Examples show how to write multidimensional queries in MDX and the Ruby interface to analyze sales data across dimensions like time, products, and customers.
Dimensional data modeling is a technique for database design intended to support analysis and reporting. It contains dimension tables that provide context about the business and fact tables that contain measures. Dimension tables describe attributes and may include hierarchies, while fact tables contain measurable events linked to dimensions. When designing a dimensional model, the business process, grain, dimensions, and facts are identified. Star and snowflake schemas are common types that differ in normalization of the dimensions. Slowly changing dimensions also must be accounted for.
The document describes the development and implementation of an Integrated Analytical Model (IAM) at Konica Minolta Ukraine to improve business decision making. Key points:
- The IAM integrated data from various ERP systems and financial reports into a multidimensional model to allow comprehensive analysis of KPIs across business areas over time.
- The model linked dimensions like products, customers, and time to financial and operational metrics, enabling analysis of factors like customer profitability by segment.
- Results are communicated through SharePoint reports and Excel pivot tables, providing managers targeted insights while allowing power users flexible ad-hoc analysis.
- The IAM created a "single version of truth" and
This document discusses using machine learning for demand forecasting in supply chain management. It begins by outlining problems with traditional forecasting methods and high errors affecting business decisions. It then proposes using machine learning algorithms that can learn from large datasets to more accurately model demand. Key steps discussed include collecting internal and external data, pre-processing data, building and comparing regression models, and developing a technical architecture to provide ongoing demand forecasting capabilities. The goals are to reduce errors, optimize inventory levels and pricing, and improve profits.
The document discusses data warehousing and the star schema. It defines a data warehouse as a repository of integrated information available for queries and analysis. The data comes from heterogeneous sources and can be queried together. It describes how a star schema organizes data into a central fact table surrounded by dimension tables. The fact table contains keys linking to attributes in the dimension tables. Star queries are processed by first using bitmap indexes on the fact table keys to retrieve relevant rows, then joining the results to the dimension tables.
This document describes a data warehouse and business intelligence project for analyzing Starbucks store data. It discusses extracting data from various structured, semi-structured, and unstructured sources, transforming the data using SQL and R, and loading it into a star schema data warehouse with fact and dimension tables. The data warehouse is then used for business queries and analysis in Tableau, with case studies examining city revenue, visitor and beverage sales by city, and city ratings based on food and beverage counts. The analysis finds that New York City generally has the highest revenue, visitor counts, and ratings.
Applying machine learning to Kaggle data set to predict which customers are most likely to become customers. Random Forest column importance graph is helpful to prioritize the best segments to target.
Data Warehousing and Business Intelligence is one of the hottest skills today, and is the cornerstone for reporting, data science, and analytics. This course teaches the fundamentals with examples plus a project to fully illustrate the concepts.
This document discusses questions to ask when preparing to analyze a company's data from multiple sources. It addresses determining data sources and structure, cleaning data, identifying key performance indicators and metrics to analyze, understanding user needs, and planning how analysis results will be used and reported. The goal is to map the data lifecycle, ensure quality, and select the right analysis approach and tools to provide insights for business decisions.
The document summarizes key concepts in dimensional data warehouse design based on Ralph Kimball's approach. It discusses the four-step design process of selecting a business process, declaring the grain, choosing dimensions, and identifying facts. It then provides an example applying these steps to design a dimensional schema for a retail sales data warehouse using data from a point-of-sale system, with dimensions for date, product, store, and promotion. It also covers topics like additive/non-additive facts, degenerate dimensions, extensibility, and avoiding normalization of dimension tables.
Similar to Tn shaw 107 data warehousing problem set (20)
We are pleased to share with you the latest VCOSA statistical report on the cotton and yarn industry for the month of March 2024.
Starting from January 2024, the full weekly and monthly reports will only be available for free to VCOSA members. To access the complete weekly report with figures, charts, and detailed analysis of the cotton fiber market in the past week, interested parties are kindly requested to contact VCOSA to subscribe to the newsletter.
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataKiwi Creative
Harness the power of AI-backed reports, benchmarking and data analysis to predict trends and detect anomalies in your marketing efforts.
Peter Caputa, CEO at Databox, reveals how you can discover the strategies and tools to increase your growth rate (and margins!).
From metrics to track to data habits to pick up, enhance your reporting for powerful insights to improve your B2B tech company's marketing.
- - -
This is the webinar recording from the June 2024 HubSpot User Group (HUG) for B2B Technology USA.
Watch the video recording at https://youtu.be/5vjwGfPN9lw
Sign up for future HUG events at https://events.hubspot.com/b2b-technology-usa/
1. Case Study on Data
Warehousing
Name: Tej Narayan Shaw
Roll: 107/MBA/191107
MBA-Day (BASM), 2019-2021 Batch
Paper & Code: Business Intelligence & Data Warehousing (B30)
2. Acknowledgment
I would like to express my heartfelt gratitude towards our
institute’s Director Shri. Dipankar Das Gupta, our Head of the
Department Prof. Dr. Tanima Ray, our esteemed faculty Professor
Subhasis Ray and all my other teachers at the Indian Institute of
Social Welfare and Business Management for their guidance and
invaluable advice throughout the course of this project. I would
also like to acknowledge my family, friends and each and every
one who contributed to this project work either directly or
indirectly.
Regards,
Tej Narayan Shaw
3. Index
Problem Set: Background 4
Problem Set: Issues discussed in the presentation 5
The Business Process 6
Process to solve the problem set 7
Dimensions 8, 9, 10, 11, 12, 13
Fact Tables 14, 15
Star Schema 16
Key Performance Indicators in Sales Performance 17, 18, 19, 20
Key Performance Indicators in Promotion Performance 21
Key Performance Indicators in Customer Preferences & Buying Patterns 22, 23
Conclusion 24
Bibliography 25
4. Problem Set: Background
● ABC Retail chain operating
in USA, Canada and Mexico.
● Foodmart has following
types of stores:
○ Supermarket
○ Small Grocery
○ Gourmet Supermarket
○ Deluxe Supermarket
○ Mid-size Grocery
● Operation started in 1990.
● Total of 24 different
outlets and 10 warehouses.
5. Problem Set: Issues discussed in the presentation
● Sales turnover from each type of store
● Product selling as per store
● Effect of promotion on sales of products
Customer Preferences
& Buying Patterns
● Effect of professions on the buying
patterns
Promotion
Performance
Sales Performance
6. The Business Process
● Indents for exhausted SKUs are raised
at store levels
● Indents goes to corresponding
warehouse in the network
● Warehouse raises PO to
suppliers/manufacturers/farmers
proportional to indents from various
stores
● Suppliers supply against PO and raises
invoice
● Payments cleared as per purchase
invoice
● Stocks transferred to stores from
warehouse as per indents
● At the end, stocks sold out to
customers through various means of
sale from the stores
● Sales are done using digital and brick
and mortar medium
7. Process to solve the problem set
● We are using Kimball’s approach of dimensional
modelling for our data warehouse
● The modelling consists of dimensions and facts
● Fact consists of consists of data which could be
measured or used for logical processioning, such
as clustering, classification, trending, etc.
● Dimensions consists of the metadata about the data
used in facts
● In context of relational schema, dimensions are
tables with primary keys and other attributes, and
fact(s) consists of measurable attributes and uses
the primary keys of dimensions as function keys
● Before creating the fact(s) table, dimensions must
pass through ETL process, so that data has single
version in the data warehouse
● Using Star Schema, the relationships between
dimensions and fact(s) are shown
8. Dimensions
CREATE TABLE Customer
(
Customer_ID INT NOT NULL,
Customer_Name CHAR(50) NOT NULL,
Customer_Location CHAR(50) NOT NULL,
Cust_Mob INT NOT NULL,
Cust_email CHAR(50) NOT NULL,
Profession_ID INT NOT NULL,
PRIMARY KEY (Customer_ID),
FOREIGN KEY (Profession_ID) REFERENCES
Customer_Profession(Cust_Profession_ID)
)
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Customer_Profession
(
Cust_Profession_ID INT NOT NULL,
Profession_Name INT NOT NULL,
PRIMARY KEY (Cust_Profession_ID)
);
9. Dimensions
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Warehouse
(
Warehouse_ID INT NOT
NULL,
Warehouse_address
CHAR(50) NOT NULL,
Telephone INT NOT NULL,
Email CHAR(20) NOT NULL,
PRIMARY KEY
(Warehouse_ID)
);
CREATE TABLE Promotion
(
Sales_Promotion_ID INT NOT NULL,
Promotion/Campaign_Name INT NOT
NULL,
Description INT NOT NULL,
Cost_of_Promotion INT NOT NULL,
PRIMARY KEY (Sales_Promotion_ID)
);
10. Dimensions
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Product_Type
(
PD_Type_Id INT NOT NULL,
Product_Type_Name INT NOT NULL,
PRIMARY KEY (PD_Type_Id)
);
CREATE TABLE Date
(
Date_ DATE NOT NULL,
Day INT NOT NULL,
Month INT NOT NULL,
Year INT NOT NULL,
Quarter INT NOT NULL,
PRIMARY KEY (Date_)
);
11. Dimensions
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Warehouse
(
Warehouse_ID INT NOT NULL,
Warehouse_address CHAR(50) NOT NULL,
Telephone INT NOT NULL,
Email CHAR(20) NOT NULL,
PRIMARY KEY (Warehouse_ID)
);
CREATE TABLE Product_
(
Product_ID INT NOT NULL,
Product_Name INT NOT NULL,
PD_Type_Id INT NOT NULL,
PRIMARY KEY (Product_ID),
FOREIGN KEY (PD_Type_Id) REFERENCES
Product_Type(PD_Type_Id)
);
12. Dimensions
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Store
(
Store_ID INT NOT NULL,
Store_Location CHAR(24) NOT NULL,
Country CHAR(20) NOT NULL,
Phone INT NOT NULL,
email_ID INT NOT NULL,
Store_Type_ID INT NOT NULL,
PRIMARY KEY (Store_ID),
FOREIGN KEY (Store_Type_ID) REFERENCES
Store_Type(Store_Type_ID)
);
CREATE TABLE Store_Type
(
Store_Type_ID INT NOT NULL,
Type_Name CHAR(50) NOT NULL,
PRIMARY KEY (Store_Type_ID)
);
13. Dimensions
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Supplier
(
Vendor_ID INT NOT NULL,
Vendor_Name CHAR(50) NOT NULL,
Telephone INT NOT NULL,
Email CHAR(50) NOT NULL,
PRIMARY KEY (Vendor_ID)
);
CREATE TABLE Consignment_Details
(
Venhicle_Details INT NOT NULL,
From INT NOT NULL,
To INT NOT NULL,
Expense INT NOT NULL,
Description VARCHAR(50) NOT NULL,
Consignment_ID INT NOT NULL,
PRIMARY KEY (Consignment_ID)
);
14. Fact Tables
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Sales_Trans
(
Stock_before_Sale INT NOT NULL,
Stock_After_Sale INT NOT NULL,
Quantity_Enquired INT NOT NULL,
Quantity_Supplied INT NOT NULL,
MRP_per_unit INT NOT NULL,
Cost_per_unit INT NOT NULL,
Discount INT NOT NULL,
Revenue INT NOT NULL,
Expense INT NOT NULL,
Income INT NOT NULL,
Store_Txn_No. INT NOT NULL,
Warehouse_ID INT NOT NULL,
Customer_ID INT NOT NULL,
Profession_ID INT NOT NULL,
Date_ DATE NOT NULL,
Sales_Promotion_ID INT NOT NULL,
PD_Type_Id INT NOT NULL,
Product_ID INT NOT NULL,
Unit_Annotation CHAR(10) NOT NULL,
Store_Type_ID INT NOT NULL,
Store_ID INT NOT NULL,
Consignment_ID INT NOT NULL,
PRIMARY KEY (Store_Txn_No.),
FOREIGN KEY (Warehouse_ID) REFERENCES Warehouse(Warehouse_ID),
FOREIGN KEY (Customer_ID) REFERENCES Customer(Customer_ID),
FOREIGN KEY (Profession_ID) REFERENCES Customer_Profession(Cust_Profession_ID),
FOREIGN KEY (Date_) REFERENCES Date(Date_),
FOREIGN KEY (Sales_Promotion_ID) REFERENCES Promotion(Sales_Promotion_ID),
FOREIGN KEY (PD_Type_Id) REFERENCES Product_Type(PD_Type_Id),
FOREIGN KEY (Product_ID) REFERENCES Product_(Product_ID),
FOREIGN KEY (Unit_Annotation) REFERENCES Scaling_Unit_of_Material(Unit_Annotation),
FOREIGN KEY (Store_Type_ID) REFERENCES Store_Type(Store_Type_ID),
FOREIGN KEY (Store_ID) REFERENCES Store(Store_ID),
FOREIGN KEY (Consignment_ID) REFERENCES Consignment_Details(Consignment_ID)
);
15. Fact Tables
ETL queries & relational diagram from OLTP to Dimensional Model:
CREATE TABLE Procurement_Trans
(
WH_Txn_No. INT NOT NULL,
In_stock INT NOT NULL,
Out_Stock INT NOT NULL,
In_Stock_Value INT NOT NULL,
Out_Stock_Value INT NOT NULL,
Warehouse_ID INT NOT NULL,
Date_ DATE NOT NULL,
PD_Type_Id INT NOT NULL,
Product_ID INT NOT NULL,
Unit_Annotation CHAR(10) NOT NULL,
Store_ID INT NOT NULL,
Vendor_ID INT NOT NULL,
Consignment_ID INT NOT NULL,
PRIMARY KEY (WH_Txn_No.),
FOREIGN KEY (Warehouse_ID) REFERENCES Warehouse(Warehouse_ID),
FOREIGN KEY (Date_) REFERENCES Date(Date_),
FOREIGN KEY (PD_Type_Id) REFERENCES Product_Type(PD_Type_Id),
FOREIGN KEY (Product_ID) REFERENCES Product_(Product_ID),
FOREIGN KEY (Unit_Annotation) REFERENCES Scaling_Unit_of_Material(Unit_Annotation),
FOREIGN KEY (Store_ID) REFERENCES Store(Store_ID),
FOREIGN KEY (Vendor_ID) REFERENCES Supplier(Vendor_ID),
FOREIGN KEY (Consignment_ID) REFERENCES Consignment_Details(Consignment_ID)
);
17. Key Performance Indicators in Sales Performance
● Sales turnover from each type of store:
Flowchart and illustrative output to achieve the above KPI
Store Type
Revenue (in $
millions)
Supermarket 76
Small Grocery 44
Gourmet Supermarket 60
Deluxe Supermarket 34
Mid-Size Grocery 88
Data
Visualization
of Illustration
18. Key Performance Indicators in Sales Performance
● Sales turnover from each type of store:
Key aspects of the KPI:
● Cause and effect analysis of the best and
worst performing store type
● Areas of study could be:
○ Location of stores
○ Mode of payments
○ Forecast of goods in demand
○ Staff behaviour
○ Competitor analysis in the particular
type of store
○ Discounts
● Total number of stores in each store type vs
revenue, i.e, revenue of sales per store in a
particular store type
● Total stock capacity of various stock types
● Average proximity of store location from
warehouse in each store type
19. Key Performance Indicators in Sales Performance
● Product selling as per store:
Flowchart and illustrative output to achieve the above KPI
Data Visualization of Illustration
Store ID
Revenue (in
$ '000)
1 2
2 7
3 9
4 3
5 7
6 7
7 7
8 6
9 4
10 8
11 2
12 1
13 7
14 6
15 2
16 2
17 8
18 1
19 7
20 8
21 7
22 2
23 1
24 3
20. Key Performance Indicators in Sales Performance
● Product selling as per store:
Key aspects of the KPI:
● Results hypothesis testing of the chart in
previous slide, could be used to identify
stores with high and low demand of particular
product(s)
● Comparison of different products of the same
product type could be done to map product
distribution
● Issues related to low stock-turnovers could be
resolved
● Efficient shelf management in small grocery
category or smaller outlets could be achieved
● Helps to study the customer behavior
hierarchy for various stores
● Helps in bulk procurements of goods by the
warehouses
21. Key Performance Indicators in Promotion Performance
● Effect of promotion on sales of products: Flowchart, illustrative example and inference
Key aspects of the KPI:
● Impact of promotion on the sales could be compared
● Best promotion strategy could be deduced by comparing sales of
products after various types of promotions in different platforms
● Trending methods of promotions could be implemented across
product range for various locations
● Further,relevance of various kind of promotions for different cohorts
of customers could be deduced
Promotion
Revenue
(in $ '000)
Newspaper 1
Television 2
Radio 2
Leafelet 3
Facebok 6
Instagram 9
Free
Samples 9
22. Key Performance Indicators in Customer Preferences & Buying Patterns
● Effect of professions on the buying patterns:
Flowchart and inference to achieve the above KPI Inference of the report under the KPI:
● The pie chart output would help to identify
professions of the regular customers
● Identification of contribution on the revenue by
each cohort based on customers’ profession
● Further slicing and dicing could be done to
identify store type or store preferred by each
cohort
● Slicing and dicing could be done to identify
products in various product types preferred by
each cohort
● For queue management, average invoicing value
of each cohort can be used to allocate billing
counters in the corresponding stores dominated
by each cohort
● Particular days in a week could be identified,
where most of transactions take place by specific
cohort
● Value chain for each such cohort could be further
planned
23. Key Performance Indicators in Customer Preferences & Buying Patterns
● Effect of professions on the buying patterns: Few more flowcharts as per the above KPI
24. Conclusion
● The fact table of the data warehouse should have
necessary metadata and metrics to address all the KPIs
● Data Marts are generated from the data warehouse
● KPIs could be better understood using interfaces which
can visualise the data
● ETL should should extract raw data and transform it for
loading in singular format, which is also used in the
data warehouse, i.e., if date in warehouse bears format
dd-mm-yyyy but the raw data is provided in the format
mm-dd-yyyy, then ETL tool should transform the date of
raw data into dd-mm-yyyy format