The objective of this project is to discuss the importance of Machine Learning in different sectors and how does it solve the problems in the Marketing Analytics field. We have discussed Marketing Segmentation, Advertisement, and Fraud detection in our project. We used different Machine Learning algorithms and used R and Python library to predict and solve these problems. After making models and running test data on those models we got following results:
• We trained a Decision tree and Random Forest classifier model which has 73% accuracy to predict whether a person will be a defaulter or not based on credit history, income, job type, dependents etc.
• We segmented the Social networking profiles based on the likes and dislikes of a person using K-Means Clustering.
• We made a predictive model of the messages a customer receives and determined whether a message will be a Spam or not a spam with an accuracy of 97%. We used Naïve Bayes classifier for this model.
Marketing analytics a practical guide to improving consumer insights using da...MarketingForum
Available at a lower price from other sellers that may not offer free Prime shipping.
Who is most likely to buy and what is the best way to target them? How can businesses improve strategy without identifying the key influencing factors? The second edition of Marketing Analytics enables marketers and business analysts to leverage predictive techniques to measure and improve marketing performance. By exploring real-world marketing challenges, it provides clear, jargon-free explanations on how to apply different analytical models for each purpose. From targeted list creation and data segmentation, to testing campaign effectiveness, pricing structures and forecasting demand, this book offers a welcome handbook on how statistics, consumer analytics and modelling can be put to optimal use.
The fully revised second edition of Marketing Analytics includes three new chapters on big data analytics, insights and panel regression, including how to collect, separate and analyze big data. All of the advanced tools and techniques for predictive analytics have been updated, translating models such as tobit analysis for customer lifetime value into everyday use. Whether an experienced practitioner or having no prior knowledge, methodologies are simplified to ensure the more complex aspects of data and analytics are fully accessible for any level of application. Complete with downloadable data sets and test bank resources, this book supplies a concrete foundation to optimize marketing analytics for day-to-day business advantage.
What Is Prescriptive Analytics? Your 5-Minute OverviewShannon Kearns
This slide deck walks you through the basis of understanding prescriptive analytics. Understand the different kinds of prescriptive analytics, how it works, its value, where to find use cases and more!
Data science is an integrative field that uses scientific methods, processes, algorithms, and systems to extract, knowledge and awareness from data in various forms
Marketing analytics a practical guide to improving consumer insights using da...MarketingForum
Available at a lower price from other sellers that may not offer free Prime shipping.
Who is most likely to buy and what is the best way to target them? How can businesses improve strategy without identifying the key influencing factors? The second edition of Marketing Analytics enables marketers and business analysts to leverage predictive techniques to measure and improve marketing performance. By exploring real-world marketing challenges, it provides clear, jargon-free explanations on how to apply different analytical models for each purpose. From targeted list creation and data segmentation, to testing campaign effectiveness, pricing structures and forecasting demand, this book offers a welcome handbook on how statistics, consumer analytics and modelling can be put to optimal use.
The fully revised second edition of Marketing Analytics includes three new chapters on big data analytics, insights and panel regression, including how to collect, separate and analyze big data. All of the advanced tools and techniques for predictive analytics have been updated, translating models such as tobit analysis for customer lifetime value into everyday use. Whether an experienced practitioner or having no prior knowledge, methodologies are simplified to ensure the more complex aspects of data and analytics are fully accessible for any level of application. Complete with downloadable data sets and test bank resources, this book supplies a concrete foundation to optimize marketing analytics for day-to-day business advantage.
What Is Prescriptive Analytics? Your 5-Minute OverviewShannon Kearns
This slide deck walks you through the basis of understanding prescriptive analytics. Understand the different kinds of prescriptive analytics, how it works, its value, where to find use cases and more!
Data science is an integrative field that uses scientific methods, processes, algorithms, and systems to extract, knowledge and awareness from data in various forms
Large amounts of heterogeneous medical data have become available in various healthcare organizations (payers, providers, pharmaceuticals). Those data could be an enabling resource for deriving insights for improving care delivery and reducing waste. The enormity and complexity of these datasets present great challenges in analyses and subsequent applications to a practical clinical environment. More details are available here http://dmkd.cs.wayne.edu/TUTORIAL/Healthcare/
This presentation introduces some concepts of Data Analytics including: Data Science, Big Data, Social Network Analysis, Process Mining, Market Basket Analysis, and Pattern Recognition
Data 101: Introduction to Data VisualizationDavid Newbury
Do you want to make pictures using data but don't know where to start? Would you like to learn how data visualization works, and how to tell stories with data?
This workshop by David Newbury explores the history of data visualization from the first maps to the latest interactive tools from the New York Times.The workshop will also discuss the hows and whys of storytelling with data. It finshes with a collaborative exploration of data visualization using Sharpies, Post-It notes, and things that begin with "S".
No computers will be used in this class, and there are no prerequisites. As a result of this workshop, you'll have a stronger foundation in understanding how to communicate information more-effectively.
We’re excited to partner with the Carnegie Library of Pittsburgh on a “Data 101” training series designed to build information literacy, mapping, and data visualization skills for people looking to get started in using data, or more-experienced users looking to brush-up on their skills. The training sessions will be offered monthly at one of the Library’s branches, and will be followed by ample time to practice what you’ve learned.
This first class on data visualization was offered on the morning of May 10, 2016 at the East Liberty Branch.
Data Visualization is widely used in industries in info-graphics design, business analytics, data analytics, advanced analytics, business intelligence dashboards, content marketing. It is the 1st part of 3 part series on data visualization. These techniques will enable you to create a good design UI/UX. It contains r codes useful for programmers to create good visual charts and depict a story to clients, customer, senior management, etc ...
Business Analytics, "Second Edition teaches the fundamental concepts of the emerging field of business analytics and provides vital tools in understanding how data analysis works in today s organizations. Students will learn to apply basic business analytics principles, communicate with analytics professionals, and effectively use and interpret analytic models to make better business decisions. Included access to commercial grade analytics software gives students real-world experience and career-focused value. Author James Evans takes a balanced, holistic approach and looks at business analytics from descriptive, and predictive perspectives.
Data analytics presentation- Management career institute PoojaPatidar11
1. The basic definition of Data, Analytics, and Data Analytics
2. Definition: Data: Data is a set of values of qualitative or quantitative variables. It is information in the raw or unorganized form. It may be a fact, figure, characters, symbols etc
Analytics: Analytics is the discovery, interpretation, and communication of meaningful patterns in data and applying those patterns towards effective decision making.
Data Analytics: Data analytics refers to qualitative and quantitative techniques and processes used to enhance productivity and business gain.
3.Types of analytics: Predictive Analytics (What could happen?)
Prescriptive Analytics (What should we do)
Descriptive Analytics (What has happened?)
4.Why Data analytics? Data Analytics is needed in Business to Consumer applications (B2C)
5.The process of Data analytics: Data requirements,
Data collection, Data processing, Data cleaning, Exploratory data analysis,
Modeling and algorithms, Data product, Communication
6.The scope of Data Analytics: Bright future of data analytics, many professionals and students are interested in a career in data analytics.
7.Importance of data analytics:1. Predict customer trends and behaviors
Analyze,
2 interpret and deliver data in meaningful ways
3.Increase business productivity
4.Drive effective decision-making
8.why become a data analyst? talented gaps of skill candidates, good salaries for freshers, great future growth path
9. What recruiters look for in applicants: Problem-Solving Skills, Analytical Mind, Maths and Statistic Skills, Communication (both oral and written), Teamwork Abilities
10. Skill is required for Data analytics?
1.) Analytical Skills
2.) Numeracy Skills
3.) Technical and Computer Skills
4.) Attention to Details
5.) Business Skills
6.) Communication Skills
11. Data analytics tools
1.SAS: SAS (Statistical Analysis System) is a software suite developed by SAS Institute. sas language can be defined as a programming language in the computing field. This language is generally used for the purpose of statistical analysis. The language has the ability to read data from databases and common spreadsheets.
2. R: R is a programming language and software environment for statistical analysis, graphics representation and reporting.R is freely available under the GNU General Public License, and pre-compiled binary versions are provided for various operating systems like Linux, Windows, and Mac.
3.PYTHON: Python is a popular programming language Python is a powerful, flexible, open-sources language that is easy to use,
and has a powerful library for data manipulation and analysis.
4.TABLEAU: Tableau Software is a software company that produces interactive data visualization products focused on business intelligence.
Top Data Mining Techniques and Their ApplicationsPromptCloud
In this presentation we have covered why data mining is important and various techniques used for data mining. Apart from that, examples of applications have been given for each technique. This presentation also explains how an enterprise can source web data via crawling services to bolster data mining models.
Top Big data Analytics tools: Emerging trends and Best practicesSpringPeople
For many IT experts, big data analytics tools and technologies are now a top priority. Let's find out the top big data analytics tools in this slide to initialize and advance the process of big data analysis.
Data science is the practice of extracting, analyzing, and interpreting large amounts of data to identify trends, correlations, and patterns. It combines machine learning, statistics, programming, and data engineering tools to uncover insights that can inform business decisions. Data scientists collect, organize, and analyze large amounts of data to find valuable insights and make predictions. Data science can be used in various industries, from finance and health care to retail and advertising. By leveraging data-driven decision-making, companies are able to gain a better understanding of their customers, identify new growth opportunities, and optimize their operations.
Large amounts of heterogeneous medical data have become available in various healthcare organizations (payers, providers, pharmaceuticals). Those data could be an enabling resource for deriving insights for improving care delivery and reducing waste. The enormity and complexity of these datasets present great challenges in analyses and subsequent applications to a practical clinical environment. More details are available here http://dmkd.cs.wayne.edu/TUTORIAL/Healthcare/
This presentation introduces some concepts of Data Analytics including: Data Science, Big Data, Social Network Analysis, Process Mining, Market Basket Analysis, and Pattern Recognition
Data 101: Introduction to Data VisualizationDavid Newbury
Do you want to make pictures using data but don't know where to start? Would you like to learn how data visualization works, and how to tell stories with data?
This workshop by David Newbury explores the history of data visualization from the first maps to the latest interactive tools from the New York Times.The workshop will also discuss the hows and whys of storytelling with data. It finshes with a collaborative exploration of data visualization using Sharpies, Post-It notes, and things that begin with "S".
No computers will be used in this class, and there are no prerequisites. As a result of this workshop, you'll have a stronger foundation in understanding how to communicate information more-effectively.
We’re excited to partner with the Carnegie Library of Pittsburgh on a “Data 101” training series designed to build information literacy, mapping, and data visualization skills for people looking to get started in using data, or more-experienced users looking to brush-up on their skills. The training sessions will be offered monthly at one of the Library’s branches, and will be followed by ample time to practice what you’ve learned.
This first class on data visualization was offered on the morning of May 10, 2016 at the East Liberty Branch.
Data Visualization is widely used in industries in info-graphics design, business analytics, data analytics, advanced analytics, business intelligence dashboards, content marketing. It is the 1st part of 3 part series on data visualization. These techniques will enable you to create a good design UI/UX. It contains r codes useful for programmers to create good visual charts and depict a story to clients, customer, senior management, etc ...
Business Analytics, "Second Edition teaches the fundamental concepts of the emerging field of business analytics and provides vital tools in understanding how data analysis works in today s organizations. Students will learn to apply basic business analytics principles, communicate with analytics professionals, and effectively use and interpret analytic models to make better business decisions. Included access to commercial grade analytics software gives students real-world experience and career-focused value. Author James Evans takes a balanced, holistic approach and looks at business analytics from descriptive, and predictive perspectives.
Data analytics presentation- Management career institute PoojaPatidar11
1. The basic definition of Data, Analytics, and Data Analytics
2. Definition: Data: Data is a set of values of qualitative or quantitative variables. It is information in the raw or unorganized form. It may be a fact, figure, characters, symbols etc
Analytics: Analytics is the discovery, interpretation, and communication of meaningful patterns in data and applying those patterns towards effective decision making.
Data Analytics: Data analytics refers to qualitative and quantitative techniques and processes used to enhance productivity and business gain.
3.Types of analytics: Predictive Analytics (What could happen?)
Prescriptive Analytics (What should we do)
Descriptive Analytics (What has happened?)
4.Why Data analytics? Data Analytics is needed in Business to Consumer applications (B2C)
5.The process of Data analytics: Data requirements,
Data collection, Data processing, Data cleaning, Exploratory data analysis,
Modeling and algorithms, Data product, Communication
6.The scope of Data Analytics: Bright future of data analytics, many professionals and students are interested in a career in data analytics.
7.Importance of data analytics:1. Predict customer trends and behaviors
Analyze,
2 interpret and deliver data in meaningful ways
3.Increase business productivity
4.Drive effective decision-making
8.why become a data analyst? talented gaps of skill candidates, good salaries for freshers, great future growth path
9. What recruiters look for in applicants: Problem-Solving Skills, Analytical Mind, Maths and Statistic Skills, Communication (both oral and written), Teamwork Abilities
10. Skill is required for Data analytics?
1.) Analytical Skills
2.) Numeracy Skills
3.) Technical and Computer Skills
4.) Attention to Details
5.) Business Skills
6.) Communication Skills
11. Data analytics tools
1.SAS: SAS (Statistical Analysis System) is a software suite developed by SAS Institute. sas language can be defined as a programming language in the computing field. This language is generally used for the purpose of statistical analysis. The language has the ability to read data from databases and common spreadsheets.
2. R: R is a programming language and software environment for statistical analysis, graphics representation and reporting.R is freely available under the GNU General Public License, and pre-compiled binary versions are provided for various operating systems like Linux, Windows, and Mac.
3.PYTHON: Python is a popular programming language Python is a powerful, flexible, open-sources language that is easy to use,
and has a powerful library for data manipulation and analysis.
4.TABLEAU: Tableau Software is a software company that produces interactive data visualization products focused on business intelligence.
Top Data Mining Techniques and Their ApplicationsPromptCloud
In this presentation we have covered why data mining is important and various techniques used for data mining. Apart from that, examples of applications have been given for each technique. This presentation also explains how an enterprise can source web data via crawling services to bolster data mining models.
Top Big data Analytics tools: Emerging trends and Best practicesSpringPeople
For many IT experts, big data analytics tools and technologies are now a top priority. Let's find out the top big data analytics tools in this slide to initialize and advance the process of big data analysis.
Data science is the practice of extracting, analyzing, and interpreting large amounts of data to identify trends, correlations, and patterns. It combines machine learning, statistics, programming, and data engineering tools to uncover insights that can inform business decisions. Data scientists collect, organize, and analyze large amounts of data to find valuable insights and make predictions. Data science can be used in various industries, from finance and health care to retail and advertising. By leveraging data-driven decision-making, companies are able to gain a better understanding of their customers, identify new growth opportunities, and optimize their operations.
As per Gartner, global revenue in the business intelligence (BI) and analytics software market is forecast to reach $18.3 billion in 2017, an increase of 7.3 percent from 2016, according to the latest forecast from Gartner, Inc. By the end of 2020, the market is forecast to grow to $22.8 billion.
Data Science has become one of the most demanded jobs of the 21st century. It has become a buzzword that almost everyone talks about these days. But what is Data Science? In this article, we will demystify Data Science, the role of a Data Scientist and have a look at the tools required to master Data Science.
Data science is an interdisciplinary field that uses algorithms, procedures, and processes to examine large amounts of data in order to uncover hidden patterns, generate insights, and direct decision making.
Big Data Courses In Mumbai at Asterix Solution is designed to scale up from single servers to thousands of machines, each offering local computation and storage. With the rate at which memory cost decreased the processing speed of data never increased and hence loading the large set of data is still a big headache and here comes Hadoop as the solution for it.
http://www.asterixsolution.com/big-data-hadoop-training-in-mumbai.html
Data science is an interdisciplinary field that uses algorithms, procedures, and processes to examine large amounts of data in order to uncover hidden patterns, generate insights, and direct decision making.
Data Science - Part I - Sustaining Predictive Analytics CapabilitiesDerek Kane
This is the first lecture in a series of data analytics topics and geared to individuals and business professionals who have no understand of building modern analytics approaches. This lecture provides an overview of the models and techniques we will address throughout the lecture series, we will discuss Business Intelligence topics, predictive analytics, and big data technologies. Finally, we will walk through a simple yet effective example which showcases the potential of predictive analytics in a business context.
What is Data analytics? How is data analytics a better career option?Aspire Techsoft Academy
Are you looking for the Best Data analytics Training Institute in Pune Aspire Techsoft offers you the best SAS Data Analytics Certification Training in Pune with Certified expert faculties.
Now companies are in the middle of a renovation that forces them to be analytics-driven to
continue being competitive. Data analysis provides a complete insight about their business. It
also gives noteworthy advantages over their competitors. Analytics-driven insights compel
businesses to take action on service innovation, enhance client experience, detect irregularities in
process and provide extra time for product or service marketing. To work on analytics driven
activities, companies require to gather, analyse and store information from all possible sources.
Companies should bring appropriate tools and workflows in practice to analyse data rapidly and
unceasingly. They should obtain insight from data analysis result and make changes in their
business process and practice on the basis of gained result. It would help to be more agile than
their previous process and function.
Similar to Marketing Analytics using R/Python (20)
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
Marketing Analytics using R/Python
1. Capstone Project - IS 6596
Project Supervisor:
Dr. Rohit Aggarwal
Project Contributors:
Mayank Badjatya - u1085897
Sagar Singh - u1088202
MARKETING ANALYTICS USING
R/PYTHON
2. 1
Capstone Project – IS 6596
Contents
Executive Summary.......................................................................................................................................2
Book Description...........................................................................................................................................3
Why Data Science?........................................................................................................................................5
Skill sets required for a Data Science............................................................................................................6
7 Steps to effective Predictive Modelling.....................................................................................................7
Marketing Analysis........................................................................................................................................9
Fraud Detection ......................................................................................................................................10
Market Segmentation.............................................................................................................................13
Advertising..............................................................................................................................................16
Lessons Learned..........................................................................................................................................19
Next Steps...................................................................................................................................................19
3. 2
Capstone Project – IS 6596
Executive Summary
The objective of this project is to discuss the importance of Machine Learning in different sectors and how
does it solve the problems in the Marketing Analytics field. We have discussed Marketing Segmentation,
Advertisement, and Fraud detection in our project. We used different Machine Learning algorithms and
used R and Python library to predict and solve these problems. After making models and running test data
on those models we got following results:
• We trained a Decision tree and Random Forest classifier model which has 73% accuracy to predict
whether a person will be a defaulter or not based on credit history, income, job type, dependents
etc.
• We segmented the Social networking profiles based on the likes and dislikes of a person using K-
Means Clustering.
• We made a predictive model on the messages a customer receives and determined whether a
message will be a Spam or not a spam with an accuracy of 97%. We used Naïve Bayes classifier
for this model.
• We created several other models using different algorithms, but these are beyond the scope of
this report.
4. 3
Capstone Project – IS 6596
Book Description
An Introduction to Statistical Learning provides an accessible overview of the field of statistical learning,
an essential toolset for making sense of the vast and complex data sets that have emerged in fields ranging
from biology to finance to marketing to astrophysics in the past twenty years. This book presents some of
the most important modeling and prediction techniques, along with relevant applications. Topics include
linear regression, classification, resampling methods, shrinkage approaches, tree-based methods, support
vector machines, clustering, and more. Color graphics and real-world examples are used to illustrate the
methods presented. Since the goal of this textbook is to facilitate the use of these statistical learning
techniques by practitioners in science, industry, and other fields, each chapter contains a tutorial on
implementing the analyses and methods presented in R, an extremely popular open source statistical
software platform. An Introduction to Statistical Learning covers many of the same topics, but at a level
accessible to a much broader audience. This book is targeted at statisticians and non-statisticians alike
who wish to use innovative statistical learning techniques to analyze their data. The text assumes only a
previous course in linear regression and no knowledge of matrix algebra.
Machine Learning with R: This book is intended for anybody hoping to use data for action. Perhaps you
already know a bit about machine learning, but have never used R; or perhaps you know a little about R,
but are new to machine learning. In any case, this book will get you up and running quickly. It would be
helpful to have a bit of familiarity with basic math and programming concepts, but no prior experience is
required. All you need is curiosity.
Machine learning, at its core, is concerned with the algorithms that transform information into actionable
intelligence. This fact makes machine learning well-suited to the present-day era of big data. Without
machine learning, it would be nearly impossible to keep up with the massive stream of information. Given
the growing prominence of R—a cross-platform, zero-cost statistical programming environment—there
has never been a better time to start using machine learning. R offers a powerful but easy-to-learn set of
tools that can assist you with finding data insights. By combining hands-on case studies with the essential
5. 4
Capstone Project – IS 6596
theory that you need to understand how things work under the hood, this book provides all the knowledge
that you will need to start applying machine learning to your own projects.
Marketing Analytics Data Driven Techniques: This book helps tech-savvy marketers and data analysts
solve real-world business problems with Excel.
Using data-driven business analytics to understand customers and improve results is a great idea in
theory, but in today's busy offices, marketers and analysts need simple, low-cost ways to process and
make the most of all that data. This expert book offers the perfect solution. Written by data analysis expert
Wayne L. Winston, this practical resource shows you how to tap a simple and cost-effective tool, Microsoft
Excel, to solve specific business problems using powerful analytic techniques—and achieve optimum
results. Practical exercises in each chapter helped us to apply and reinforce techniques as you learn.
Shows you how to perform sophisticated business analyses using the cost-effective and widely available
Microsoft Excel instead of expensive, proprietary analytical tools
• Reveals how to target and retain profitable customers and avoid high-risk customers
• Helps you forecast sales and improve response rates for marketing campaigns
• Explores how to optimize price points for products and services, optimize store layouts, and
improve online advertising
• Covers social media, viral marketing, and how to exploit both effectively.
6. 5
Capstone Project – IS 6596
Why Data Science?
Data Science is a field, which can be implemented anywhere. Here is the list of people who uses data
science as a tool in their field and are not from IT background.
• Politics: We may have heard how statistical wizard Nate Silver predicted the electoral votes for
each state in the 2012 presidential election, showing that raw data crunching of polls is much
more reliable than traditional punditry.
• Healthcare: The role of big data in medicine is one where we can build better health profiles and
better predictive models around individual patients so that we can better diagnose and treat
disease. Big data comes into play around aggregating increasingly information around multiple
scales for what constitutes a disease—from the DNA, proteins, and metabolites to cells, tissues,
organs, organisms, and ecosystems.
• Automotive Industry: Areas in the automotive industry impacted by Big Data include:
a. Conceptual Design: Real-world data collected from billions of miles driven will undoubtedly
influence safety, aerodynamics, power algorithms and other fundamental elements of the vehicle.
b. Drawing Boards: Efficiency gained in design, production volumes and manufacturing through
Big Data in the auto industry will make it economically feasible to make today’s options
tomorrow’s standard equipment.
c. Procurement: Supply chain management optimized by Big Data will help manufacturers
continue to wring new efficiency from the procurement process.
d. Manufacturing: On the assembly line, data gathered throughout the building process will be
used in predictive analytics to improve manufacturing simulations and watch machine
performance, making the next assembly line even more efficient and flexible.
• Marketing: Big Data is already having a major influence on vehicle marketing. Social sentiment
will play a growing role in manufacturers’ plans to design new vehicles. Customer feedback on
current models also helps marketing experts identify key themes and messages for new
campaigns.
• Finance: Understanding consumer habits, preferences and buying power across market segments
gives manufacturers insights needed to develop more-effective financing programs. But that’s just
the first step. New insights from Big Data analyses of sales and in-field use data will help captive
financing companies develop new services and new revenue streams.
• Services: Like performance, service will benefit as both a contributor and a user of Big Data in the
automotive industry. Information gathered through millions of service events will provide
feedback to designers.
7. 6
Capstone Project – IS 6596
Skill sets required for a Data Science
Technical Skills:
Python Coding – Python is the most common coding language I typically see required in data science roles,
along with Java, Perl, or C/C++.
Hadoop Platform – Although this isn’t always a requirement, it is heavily preferred in many cases. Having
experience with Hive or Pig is also a strong selling point. Familiarity with cloud tools such as Amazon S3
can also be beneficial.
SQL Database/Coding – Even though NoSQL and Hadoop have become a large component of data science,
it is still expected that a candidate will be able to write and execute complex queries in SQL.
Unstructured data – It is critical that a data scientist be able to work with unstructured data, whether it is
from social media, video feeds or audio.
Non-Technical Skills
Intellectual curiosity – No doubt we have seen this phrase everywhere lately, especially as it relates to
data scientists. Frank Lo describes what it means, and talks about other necessary “soft skills” in his guest
blog posted a few months ago.
Business acumen – To be a data scientist we’ll need a solid understanding of the industry we’re working
in, and know what business problems your company is trying to solve. In terms of data science, being able
to discern which problems are important to solve for the business is critical, in addition to identifying new
ways the business should be leveraging its data.
Communication skills – Companies searching for a strong data scientist are looking for someone who can
clearly and fluently translate their technical findings to a non-technical team, such as the Marketing or
Sales departments. A data scientist must enable the business to make decisions by arming them with
quantified insights, in addition to understanding the needs of their non-technical colleagues to wrangle
the data appropriately.
8. 7
Capstone Project – IS 6596
7 Steps to effective Predictive Modelling
Step 1: Defining the Objective
The first step in any modeling process is defining the objective. We see in what field does the problem fall
in. There are many fields like Target Marketing, Risk & Fraud Management, Strategy Implementation and
Change Management, Operational Efficiency, Increase Customer Experience, Manage Marketing,
Campaigns Forecast, Revenue or Loss, Workforce Management, Financial Modeling, Churn Management,
and Social Media Influencers
Step 2: Gathering the Data
Accurate, actionable, accessible data is the lifeblood of any successful model. So we collect enough data
to make a predictive model on it.
Step 3: Preparing the Data for Modeling
The average modeler spends 70% of his or her time preparing data. In this step we need to prepare data
into right format for analysis and the tool we may want use.
1. Do initial cleaning up
2. Define Variables and Create Data Dictionary
3. Joining/Appending multiple datasets
4. Validate for correctness
5. Produce Basic Summary Reports
Step 4: Selecting and Transforming the Variables
Determining the best fit is essential to good model performance. The underlying structure of the
independent variables in relation to the dependent variable, determines the power and longevity of a
model.
Special consideration is given to the fact that marketing data can have hundreds or even thousands of
variables. We apply methods for identifying the best candidate variables. Programs are introduced that
automatically segment and transform the most powerful variables, to ensure the best fit.
Step 5: Processing and Evaluating the Model
All the preparation works up to this point makes this next step run smoothly. Weights of Evidence and
Information Values are calculated. For our main case study, we used various options within PROC LOGISTIC
to determine the model with the best fit. Validation data are scored, tabulated, and compared using both
SAS® & MSExcel®.
Step 6: Validating the Model
Models should perform well on the development data. Plus, if the hold-out sample is randomly selected,
the model performance should score the validation data with similar results. A true test of model
performance is how well it performs on data from a different time or market area. So, we used three
powerful methods for ensuring model fit. 1) Scoring alternate data is the best way to tell if our model will
9. 8
Capstone Project – IS 6596
perform in a real campaign; 2) Bootstrapping uses simple resampling techniques to find confidence
intervals around our estimates; 3) Key Variable Analysis calculates important market factors as they are
affected by the model, thus ensuring reasonable results.
Step 7: Implementing and Maintaining the Model
Effective implementation is a combination of business intelligence and well-designed procedures. So, we
score a new data set with the new model. Several auditing procedures are done and tracking, and model
maintenance are emphasized as best practices.
Figure 1 7 Steps of Predictive Model
10. 9
Capstone Project – IS 6596
Marketing Analysis
Figure 2 : Facets of Marketing Analysis
An accurate customer risk assessment will help us acquire the most profitable consumers while
minimizing risk. For business-to-consumer companies, Experian offers consumer credit information,
advanced scoring software, prescreening systems, and application decisioning tools. For companies
looking to acquire business customers, our business reports and public records, portfolio data and risk
modeling tools allow clients to create comprehensive profiles of business prospects. Determine which
businesses are well-capitalized and financially suited for customer acquisition.
11. 10
Capstone Project – IS 6596
Fraud Detection
Fraud is a billion-dollar business and it is increasing every year. The PwC global economic crime survey of
2016 suggests that more than one in three (36%) of organizations experienced economic crime.
Traditional methods of data analysis have long been used to detect fraud. They require complex and time-
consuming investigations that deal with different domains of knowledge like financial, economics,
business practices and law.
To know more about how Machine Learning algorithms, solve Fraud detection problem we took a dataset
from the “Machine Learning using R” credit data set.
The idea behind our credit model is to identify factors that make an applicant at higher risk of default.
Therefore, we need to obtain data on many past bank loans and whether the loan went into default, as
well as information about the applicant.
We can see that “job”, “phone”,
“checking_balance”,
“credit_history”, “purpose”,”
savings_balance”,
“employment_duration”,
“other_credit”, “housing” are the
categorical data so in Python we
use onehotencoder() to convert
the categorical data into 0s and 1s.
After applying the
onehotencoder() on all categorical
dataset we got 36 columns. The
credit dataset includes 1,000
examples of loans, plus a
combination of numeric and
nominal features indicating
characteristics of the loan and the
loan applicant. A class variable
indicates whether the loan went
into default.
Figure 3 Conversion of categorical data into 0s and 1s
12. 11
Capstone Project – IS 6596
We did the initial data exploration and plotted that using matplotlib library.
Figure 4 Exploratory Data Analysis
We used decision tree to determine whether a person is a defaulter or not depending on the features.
The core algorithm for building decision trees called ID3. The Decision tree classifiers uses greedy
approach hence an attribute chooses at first step can’t be used anymore which can give better
classification if used in later steps. Also, it overfits the training data which can give poor results for unseen
data. It uses two concepts to determine on which feature it needs to divide the dataset.
Information Gain
The information gain is based on the decrease in entropy after a dataset is split on an attribute.
Constructing a decision tree is all about finding attribute that returns the highest information gain (i.e.,
the most homogeneous branches).
Entropy
A decision tree is built top-down from a root node and involves partitioning the data into subsets that
contain instances with similar values (homogenous). ID3 algorithm uses entropy to calculate the
homogeneity of a sample. If the sample is completely homogeneous the entropy is zero and if the sample
is an equally divided it has entropy of one.
After applying the Decision tree model, we got the following classification report.
13. 12
Capstone Project – IS 6596
Figure 5 F1 Score for Decision Tree
F1 score is a measure of a test's accuracy. The F1 score is the harmonic average of the precision and recall,
where an F1 score reaches its best value at 1 (perfect precision and recall) and worst at 0.
Decision tree makes a model which is biased so to overcome this drawback we use Bagging.
Bagging is a way to decrease the variance of our prediction by generating additional data for training from
our original dataset using combinations with repetitions to produce multisets of the same cardinality/size
as our original data.
Random Forests is an ensemble classifier which uses many decision tree models to predict the result. A
different subset of training data is selected, with replacement to train each tree. A collection of trees is a
forest, and the trees are being trained on subsets which are being selected at random, hence random
forests. After applying Random Forest classifier, we got the following result.
Figure 6 F1 Score for Random Forest
We can clearly see the increase in the F1-score.
Now the next step in building model as discussed earlier is to fine tune the model. For this we use Grid
Search Cross Validation technique. After applying the GridSearchCV we got the following classification
report.
Figure 7 F1 Score after GridSearchCV
From this model we understand that the model will predict 73% of the time whether a person will be a
defaulter or not.
14. 13
Capstone Project – IS 6596
Market Segmentation
One of the most fundamental marketing activities is in market segmentation. As companies cannot
connect with all their potential customers, they must divide markets into groups (segments) of consumers,
customers, or clients with similar needs and wants. Firms can then target each of these segments by
positioning themselves in a unique segment (such as Ferrari in the high-end sports car market).
While market researchers often form market, segments based on
practical grounds, industry practice and wisdom, cluster analysis
allows segments to be formed that are based on data that are less
dependent on subjectivity.
Cluster analysis is a convenient method for identifying homogeneous
groups of objects called clusters. Objects (or cases, observations) in a
specific cluster share many characteristics, but are very dissimilar to
objects not belonging to that cluster.
Below we have tried try this process from start to finish.
For this analysis, we used a dataset representing a random sample of 30,000 U.S. high school students
who had profiles on a well-known SNS in 2006. To protect the users' anonymity, the SNS will remain
unnamed. However, at the time the data was collected, the SNS was a popular web destination for US
teenagers. Therefore, it is reasonable to assume that the profiles represent a wide cross section of
American adolescents in 2006.
Let's take a quick look at the specifics of the data.
Figure 8 Description of the data set
15. 14
Capstone Project – IS 6596
Figure 9 Min-Max of the Age Figure 10 Gender and Age anomaly
There is something strange around the gender row. On looking carefully, we noticed the NA value. We
see that 2,724 records (9 percent) have missing gender data.
Besides gender, only age has missing values. A total of 5,086 records (17 percent) have missing ages. Also
concerning is the fact that the minimum and maximum values seem to be unreasonable; it is unlikely that
a 3-year-old or a 106-year-old is attending high school. To ensure that these extreme values don't cause
problems for the analysis, we cleaned them up before moving on.
Figure 11 Box Plot for the age distribution
A more reasonable range of ages for the high school students includes those who are at least 13 years old
and not yet 20 years old. Any age value falling outside this range we treated the same as missing data.
An easy solution for handling the missing values is to exclude any record with a missing value. In this case,
we created dummy variables for female and unknown gender. We assigned teens$female the value 1 if
gender is equal to F and the gender is not equal to NA; otherwise, it assigns the value 0 .
Next, we eliminated the 5,523 missing age values. We have used a different strategy known as data
imputation, which involves filling in the missing data with a guess as to the true value. Most people in a
graduation cohort were born within a single calendar year. We have identified the typical age for each
cohort, we had a reasonable estimate of the age of a student in that graduation year.
16. 15
Capstone Project – IS 6596
To cluster the teenagers into marketing segments, we used an implementation of k-means clustering. We
started our cluster analysis by considering only the 36 features that represent the number of times various
interests appeared on the teen SNS profiles.
Evaluating clustering results can be somewhat subjective. Ultimately, the success or failure of the model
hinges on whether the clusters are useful for their intended purpose. As the goal of this analysis was to
identify clusters of teenagers with similar interests for marketing purposes, we largely measured our
success in qualitative terms. For other clustering applications, more quantitative measures of success may
be needed. By examining whether the clusters fall above or below the mean level for each interest
category, we can notice patterns that distinguish the clusters from each other. Cluster 3 is substantially
above the mean interest level on all the sports. This suggests that this may be a group of Athletes per The
Breakfast Club stereotype.
Figure 12 Cluster segmentation
Cluster 0 includes the most mentions of "cheerleading," the word "hot," and is above the average level of
football interest. Hence, these are the so-called Princesses. Similarly, we tried to cluster the different
groups, and this is what we found.
We now focused our effort on turning these insights into action. We applied the clusters back onto the
full dataset.
We looked at the demographic characteristics of the clusters. The mean age does not vary much by
cluster, which is not too surprising as these teen identities are often determined before high school. On
the other hand, there are some substantial differences in the proportion of females by cluster. This is a
very interesting finding as we didn't use gender data to create the clusters, yet the clusters are still
Cluster 0 (N =
872) Princess
cute
hair
shopping
clothes
dance
Cluster 1 (N =
21308) Basket
Cases
???
Cluster 2 (N =
1041) Criminals
drunk
deaths
drugs
die
music
Cluster 3 (N =
5971) Athletes
basketball
soccer
football
volleyball
soccer
Cluster 4 (N =
808) Brains
band
marching
music
rock
17. 16
Capstone Project – IS 6596
predictive of gender. Given our success in predicting gender, we also suspected that the clusters are
predictive of the number of friends the users have. This hypothesis seems to be supported by the data.
Our findings support the popular adage that "birds of a feather flock together." By using machine learning
methods to cluster teenagers with others who have similar interests, we were able to develop a typology
of teen identities that was predictive of personal characteristics, such as gender and the number of
friends. These same methods can be applied to other contexts with similar results.
Advertising
Compared to all the marketing techniques, email marketing is the cheapest way of sending a marketing
message to millions of people. Being so cheap, it is the tool of choice for marketing teams with a small
budget trying to sell cheap products. Most of the times, such products do not deliver what they promise.
Unfortunately, with email marketing, we run the risk of being exposed to malware and fraudulent emails.
Worms and viruses often make use of email and spam techniques to propagate. Phishing emails and
Nigerian 419 scams are examples of fraudulent emails which try to harvest either our money or our
personal information including credit card details. So, while email marketing is the tool of choice for most
marketing teams, it does require stringent regulations to ensure that it does not get abused. Below we
tried to build a model which predicts whether a composed message is spam or not.
The dataset included the text of SMS messages along with a label indicating whether the message is
unwanted. Junk messages are labeled spam, while legitimate messages are labeled ham. Since Naive
Bayes has been used successfully for e-mail spam filtering, it seems likely that it could also be applied to
SMS spam. However, relative to e-mail spam, SMS spam poses additional challenges for automated filters.
SMS messages are often limited to 160 characters, reducing the amount of text that can be used to identify
whether a message is junk.
Figure 13 Description of the data set
The first step towards constructing our classifier involves processing the raw data for analysis. SMS
messages are strings of text composed of words, spaces, numbers, and punctuation. Handling this type of
complex data takes a lot of thought and effort. One needs to consider how to remove numbers and
18. 17
Capstone Project – IS 6596
punctuation; handle uninteresting words such as and, but, and or; and how to break apart sentences into
individual words.
Figure 14 Description of length of the Ham messages Figure15 Description of length of the Spam messages
Our first order of business was to standardize the messages to use only lowercase characters. To this end,
we used tolower() function that returns a lowercase version of text strings. Continuing with our cleanup
process, we also eliminated any punctuation from the text messages. Our next task was to remove filler
words such as to, and, but, and or from our SMS messages. These terms are known as stop words and are
typically removed prior to text mining. This is due to the fact that although they appear very frequently,
they do not provide much useful information for machine learning.
Another common standardization for text data involves reducing words to their root form in a process
called stemming. The stemming process takes words like learned, learning, and learns, and strips the suffix
to transform them into the base form, learn. These are left with the blank spaces that previously separated
the now-missing pieces. The final step in our text cleanup process was to remove additional whitespace.
A word cloud is a way to visually depict the frequency at which words appear in text data. The cloud is
composed of words scattered somewhat randomly around the figure. The resulting word clouds are
shown in the following diagram:
19. 18
Capstone Project – IS 6596
Figure 16 Spam Word cloud Figure 17 Ham Word cloud
Now that the data are processed to our liking, the final step is to split the messages into individual
components through a process called vectorization. We took the corpus and created a data structure in
which rows indicate documents (SMS messages) and columns indicate terms (words). The final step in the
data preparation process was to transform the sparse matrix into a data structure that can be used to
train a Naive Bayes classifier. The sparse matrix included over 6,500 features; this is a feature for every
word that appears in at least one SMS message. It's unlikely that these are useful for classification. To
reduce the number of features, we eliminated any word that appear in less than five SMS messages, or in
less than about 0.1 percent of the records in the training data.
Figure 18 Vectorization
To evaluate the SMS classifier, we need to test its predictions on unseen messages in the test data. The
process of evaluating machine learning algorithms is very similar to the process of evaluating students.
Since algorithms have varying strengths and weaknesses, tests should distinguish among the learners.
Figure 19 Classification report
20. 19
Capstone Project – IS 6596
A confusion matrix is a table that categorizes predictions according to whether they match the actual
value. One of the table's dimensions indicates the possible categories of predicted values, while the other
dimension indicates the same for actual values. Although we have only seen 2 x 2 confusion matrices so
far, a matrix can be created for models that predict any number of class value.
Lessons Learned
Lesson 1: Marketing research is fun- We get to work with a wide variety of datasets, dive in and learn all
about the market their operating in and relay valuable insights back to stakeholders. We dig up everything
from why consumers make certain purchase decisions to what they’re passionate about and what makes
them tick.
Lesson 2: Collaboration is key- While doing this project we found out that while they might be tremendous
innovators, but collaboration is very important.
Lesson 3: Check, re-check and then check again Projects move quickly which means we don’t have time
to go back and re-collect data or make corrections to a report. Questionnaires, surveys, and reports must
be checked, checked by our coworker and checked again.
Next Steps
The next step would be to discover the other facets of Marketing Analysis like “Upsell and Cross Sell”,
“Recommendation System” etc. We can use algorithms like Principal Component Analysis(PCA), QDA, LDA
to reduce the number of features. Also, we can make analysis on the time series data using ARIMA
algorithm.