Analysis of the article "A Predictive Analytics Primer" by Thomas H. DavenportVaibhav Srivastav
This presentation gives analysis of the article "A Predictive Analytics Primer" by Thomas H. Davenport
Slide 1: A Predictive Analytics Primer by Thomas H. Davenport
Slide 2: Thomas H. Davenport
Slide 3: Powers of Predictive analytics
Slide 4: Predictive analytics refers to predicting future from the data of the past.
Slide 5: The quantitative analysis isn’t magic—but it is normally done with a lot of past data, a little statistical wizardry, and some important assumptions.
Slide 6: The Data: Lack of good data is the most common barrier to organizations seeking to employ predictive analytics.
Slide 7: The Statistics: Regression analysis in its various forms is the primary tool that organizations use for predictive analytics.
Slide 8: An analyst hypothesizes that a set of independent variables (say, gender, income, visits to a website) are statistically correlated with the purchase of a product for a sample of customers. The analyst performs a regression analysis to see just how correlated each variable is; this usually requires some iteration to find the right combination of variables and the best model.
Slide 9: The Assumptions: That brings us to the other key factor in any predictive model—the assumptions that underlie it. Every model has them, and it’s important to know what they are and monitor whether they are still true. The big assumption in predictive analytics is that the future will continue to be like the past.
Slide 10: What can make assumptions invalid?
Slide 11: The most common reason is time. If your model was created several years ago, it may no longer accurately predict current behavior. The greater the elapsed time, the more likely customer behavior has changed.
Slide 12: Another reason a predictive model’s assumptions may no longer be valid is if the analyst didn’t include a key variable in the model, and that variable has changed substantially over time.
Slide 13: Managers should always ask analysts what the key assumptions are, and what would have to happen for them to no longer be valid. And both managers and analysts should continually monitor the world to see if key factors involved in assumptions might have changed over time.
Slide 14: With these fundamentals in mind, here are a few good questions to ask your analysts:
Can you tell me something about the source of data you used in your analysis?
Are you sure the sample data are representative of the population?
Are there any outliers in your data distribution? How did they affect the results?
What assumptions are behind your analysis?
Are there any conditions that would make your assumptions invalid?
Slide 15: Thank You!
ML Drift - How to find issues before they become problemsAmy Hodler
Over time, our AI predictions degrade. Full Stop.
Whether it's concept drift where the relationships of our data to what we're trying to predict as changed or data drift where our production data no longer resembles the historical training data, identifying meaningful ML drift versus spurious or acceptable drift is tedious. Not to mention the difficulty of uncovering which ML features are the source of poorer accuracy.
This session looked at the key types of machine learning drift and how to catch them before they become a problem.
Presentation on " ANALYSIS OF TED TALK BY MONA CHALABI ON 3 WAYS TO SPOT A BAD STATISTIC" made as a task for the internship on "DATA ANALYTICS WITH MANAGERIAL APPLICATIONS" under Professor Sameer Mathur, IIM Lucknow. Submitted by TARANG JAIN,DTU
Analysis of the article "A Predictive Analytics Primer" by Thomas H. DavenportVaibhav Srivastav
This presentation gives analysis of the article "A Predictive Analytics Primer" by Thomas H. Davenport
Slide 1: A Predictive Analytics Primer by Thomas H. Davenport
Slide 2: Thomas H. Davenport
Slide 3: Powers of Predictive analytics
Slide 4: Predictive analytics refers to predicting future from the data of the past.
Slide 5: The quantitative analysis isn’t magic—but it is normally done with a lot of past data, a little statistical wizardry, and some important assumptions.
Slide 6: The Data: Lack of good data is the most common barrier to organizations seeking to employ predictive analytics.
Slide 7: The Statistics: Regression analysis in its various forms is the primary tool that organizations use for predictive analytics.
Slide 8: An analyst hypothesizes that a set of independent variables (say, gender, income, visits to a website) are statistically correlated with the purchase of a product for a sample of customers. The analyst performs a regression analysis to see just how correlated each variable is; this usually requires some iteration to find the right combination of variables and the best model.
Slide 9: The Assumptions: That brings us to the other key factor in any predictive model—the assumptions that underlie it. Every model has them, and it’s important to know what they are and monitor whether they are still true. The big assumption in predictive analytics is that the future will continue to be like the past.
Slide 10: What can make assumptions invalid?
Slide 11: The most common reason is time. If your model was created several years ago, it may no longer accurately predict current behavior. The greater the elapsed time, the more likely customer behavior has changed.
Slide 12: Another reason a predictive model’s assumptions may no longer be valid is if the analyst didn’t include a key variable in the model, and that variable has changed substantially over time.
Slide 13: Managers should always ask analysts what the key assumptions are, and what would have to happen for them to no longer be valid. And both managers and analysts should continually monitor the world to see if key factors involved in assumptions might have changed over time.
Slide 14: With these fundamentals in mind, here are a few good questions to ask your analysts:
Can you tell me something about the source of data you used in your analysis?
Are you sure the sample data are representative of the population?
Are there any outliers in your data distribution? How did they affect the results?
What assumptions are behind your analysis?
Are there any conditions that would make your assumptions invalid?
Slide 15: Thank You!
ML Drift - How to find issues before they become problemsAmy Hodler
Over time, our AI predictions degrade. Full Stop.
Whether it's concept drift where the relationships of our data to what we're trying to predict as changed or data drift where our production data no longer resembles the historical training data, identifying meaningful ML drift versus spurious or acceptable drift is tedious. Not to mention the difficulty of uncovering which ML features are the source of poorer accuracy.
This session looked at the key types of machine learning drift and how to catch them before they become a problem.
Presentation on " ANALYSIS OF TED TALK BY MONA CHALABI ON 3 WAYS TO SPOT A BAD STATISTIC" made as a task for the internship on "DATA ANALYTICS WITH MANAGERIAL APPLICATIONS" under Professor Sameer Mathur, IIM Lucknow. Submitted by TARANG JAIN,DTU
This infographic tells the story of how leaps in techniques/capabilities net big results for supply chain professionals. For example, did you know that the combination of demand modeling and machine learning could reduce forecast errors by 33%?
infographics and data visualizations show that small businesses face significant cyber security threats yet still lack affordable and adequate solutions.
Predictive Analytics - Get Skilled or Die TryingKienco
Slides from the Predictive Analytics session at the Australasian Talent Conference 2014.
Topic Overview: Alex Hagan is an expert at using predictive analytics to empower decision-making at a strategic level and has advised numerous clients throughout his career. In this unconference discussion, delegates will discuss how and why talent analytics is changing the recruiting paradigm to deliver competitive advantage. In this session you’ll learn more about the best analytics structures and models that can help an organisation win in the modern age of recruiting and learn from the examples of industry peers.
Answer to the most commonly used terminology Data Science with their areas of crucial roles in solving issues with case studies.
Likewise, let me know if anything is required. Ping me at google #bobrupakroy
This presentation analyzes the HBR Article on "Big Data Hype (and Reality)" by Gregory Piatetsky-Shapiro. It emphasizes on the slow improvement of the technology, but in the end provides the areas where big data is useful.
This infographic tells the story of how leaps in techniques/capabilities net big results for supply chain professionals. For example, did you know that the combination of demand modeling and machine learning could reduce forecast errors by 33%?
infographics and data visualizations show that small businesses face significant cyber security threats yet still lack affordable and adequate solutions.
Predictive Analytics - Get Skilled or Die TryingKienco
Slides from the Predictive Analytics session at the Australasian Talent Conference 2014.
Topic Overview: Alex Hagan is an expert at using predictive analytics to empower decision-making at a strategic level and has advised numerous clients throughout his career. In this unconference discussion, delegates will discuss how and why talent analytics is changing the recruiting paradigm to deliver competitive advantage. In this session you’ll learn more about the best analytics structures and models that can help an organisation win in the modern age of recruiting and learn from the examples of industry peers.
Answer to the most commonly used terminology Data Science with their areas of crucial roles in solving issues with case studies.
Likewise, let me know if anything is required. Ping me at google #bobrupakroy
This presentation analyzes the HBR Article on "Big Data Hype (and Reality)" by Gregory Piatetsky-Shapiro. It emphasizes on the slow improvement of the technology, but in the end provides the areas where big data is useful.
Data Analysis Methods 101 - Turning Raw Data Into Actionable InsightsDataSpace Academy
Data analytics is powerful for organisations. It can help companies improve their overall efficiency and effectiveness. The blog offers a step-by-step narration of the data analysis methods that will help you to comprehend the fundamentals of an analytics project.
Economics & Statistics Insights in Data Science by DataPerts TechnologiesRavindra Panwar
DATA is an inevitable part of our life today. These tiny pieces of information from which we derive valuable insights
have their genesis in the domain of ECONOMICS and STATISTICS.
Data Analyst Interview Questions & AnswersSatyam Jaiswal
Practice Best Data Analyst Interview Questions for the best preparation of the data analyst interview. these interview questions are very popular and asked various times in data analyst interview.
📊 Dive into the world of #DataAnalytics to unlock the secrets of information! 🚀 Understanding the basics is your gateway to data-driven success. 🌐 Explore foundational concepts, from data collection to interpretation, demystifying the data landscape. 📈 Master key techniques, empowering you to extract valuable insights and make informed decisions. 💡 Enhance your analytical skills and stay ahead in the fast-paced digital era. 🧠 Whether you're a beginner or looking for a refresher, this journey into data understanding is your stepping stone to a data-savvy future!
Unveiling the Power of Data Analytics Transforming Insights into Action.pdfKajal Digital
Data analytics is the process of examining raw data to discover patterns, correlations, trends, and other valuable information. Its significance lies in its ability to transform data into actionable insights, ultimately leading to informed decision-making and improved business outcomes. From optimizing operational processes to enhancing customer experiences, data analytics offers a plethora of benefits across various sectors.
This presentation is an Introduction to the importance of Data Analytics in Product Management. During this talk Etugo Nwokah, former Chief Product Officer for WellMatch, covered how to define Data Analytics why it should be a first class citizen in any software organization
This ebook is all about data analysis, what are the steps involved in data analysis and what are the techniques. We will bring out a detailed course very soon. pls register https://excelfinanceacademy.zenler.com/ to save over 80% cost
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
3. PROCESS INVOLVED IN DATA ANALYSIS
The first step is to determine the data requirements or how
the data is grouped. Data may be separated by age,
demographic, income, or gender. Data values may be
numerical or be divided by category.
1.
4. PROCESS INVOLVED IN DATA ANALYSIS
The second step in data analytics is the process of
collecting it. This can be done through a variety of sources
such as computers, online sources, cameras,
environmental sources, or through personnel.
2.
5. PROCESS INVOLVED IN DATA ANALYSIS
Once the data is collected, it must be organized so it can be
analyzed. Organization may take place on a spreadsheet or
other form of software that can take statistical data.
3.
6. PROCESS INVOLVED IN DATA ANALYSIS
The data is then cleaned up before analysis. This means it
is scrubbed and checked to ensure there is no duplication
or error, and that it is not incomplete. This step helps
correct any errors before it goes on to a data analyst to be
analyzed.
4.
8. TYPES OF DATA ANALYTICS
Descriptive analytics: describes what has happened over a
given period of time. Have the number of views gone up? Are
sales stronger this month than last?
Diagnostic analytics: focuses more on why something
happened. This involves more diverse data inputs and a bit
of hypothesizing. Did the weather affect beer sales? Did that
latest marketing campaign impact sales?
9. TYPES OF DATA ANALYTICS
Predictive analytics: moves to what is likely going to
happen in the near term. What happened to sales the last
time we had a hot summer? How many weather models
predict a hot summer this year?
Prescriptive analytics: suggests a course of action. If the
likelihood of a hot summer is measured as an average of
these five weather models is above 58%, we should add an
evening shift to the brewery and rent an additional tank to
increase output.
13. Key Takeaways
● Data analytics is the science of
analyzing raw data in order to
make conclusions about that
information.
● The techniques and processes
of data analytics have been
automated into mechanical
processes and algorithms that
work over raw data for human
consumption.
● Data analytics help a business
optimize its performance.
15. Thank you
Datacamp Instructor |
Lead Instructor, School
of Data at Gitgirl
Co- Organiser Pydata
Port Harcourt |
Phschoolofai / Drunk in
Open Source,
IbmChampion 2019
@emekaboris
@Dataknight