The document describes a statistical model created to predict NBA playoff results based on regular season statistics. Key steps included collecting NBA data, selecting important features through correlation analysis like net rating and true shooting percentage, training a logistic regression model on past seasons, and testing it on new seasons. The model successfully predicted 6 of 6 teams that reached conference finals, though missed some like the 2023 Miami Heat. Improving the model could include additional stats like all-NBA players. This type of predictive analytics could benefit NBA teams and be applied to other sports.
Quaderno 4. Tecnica individuale e programmazione dell'allenamento da 12 a 16 anni. Le tecniche difensive, il colpo di testa e il duello aereo. 40 sedute.
Jean Robert Faucher
http://www.calzetti-mariucci.it/shop/prodotti/tecnica-individuale-e-programmazione-dellallenamento-da-12-a-16-anni-quaderno-4
This is a tried and true 7 keys to winning football games. A head football coach that designs his practice and workouts around these key factors will have a competitive football program.
Quaderno 4. Tecnica individuale e programmazione dell'allenamento da 12 a 16 anni. Le tecniche difensive, il colpo di testa e il duello aereo. 40 sedute.
Jean Robert Faucher
http://www.calzetti-mariucci.it/shop/prodotti/tecnica-individuale-e-programmazione-dellallenamento-da-12-a-16-anni-quaderno-4
This is a tried and true 7 keys to winning football games. A head football coach that designs his practice and workouts around these key factors will have a competitive football program.
4 informe individual de los jugadores del equipo del nueva chicago (esp)Fútbol Formativo
Material de estudio complementario de la Clase 4 del Curso de Entrenamiento en Fútbol Juvenil, Semi Profesional y Profesional a cargo del Lic. Germán Cavalieri que tuvo de docente invitado al entrenador argentino Nazareno Vélez
A goal is scored when the whole of the ball passes over the goal line, between the goalposts and under the crossbar, provided that no offence has been committed by the team scoring the goal.
If a referee signals a goal before the ball has passed wholly over the goal line, play is restarted with a dropped ball.
4 informe individual de los jugadores del equipo del nueva chicago (esp)Fútbol Formativo
Material de estudio complementario de la Clase 4 del Curso de Entrenamiento en Fútbol Juvenil, Semi Profesional y Profesional a cargo del Lic. Germán Cavalieri que tuvo de docente invitado al entrenador argentino Nazareno Vélez
A goal is scored when the whole of the ball passes over the goal line, between the goalposts and under the crossbar, provided that no offence has been committed by the team scoring the goal.
If a referee signals a goal before the ball has passed wholly over the goal line, play is restarted with a dropped ball.
Data mining techniques are very effective and useful for forecasting in many domains or fields. In this
research, prediction of Spanish la liga football match outcomes is carried out using various data mining techniques
(Multilayer Perception, Decision Tables, Random Forest, Reptree and Meta. Bagging) to determine the most accurate
among these techniques.
Metulini, R., Manisera, M., Zuccolotto, P. (2017), Sensor Analytics in Basket...University of Salerno
A new approach in team sports analysis consists in studying positioning and movements of players during the game in relation to team performance. State of the art tracking systems produce spatio-temporal traces of players that have facilitated a variety of research aimed to extract insights from trajectories. Several methods borrowed from machine learning, network and complex systems, geographic information system, computer vision and statistics have been proposed. After having reviewed the state of the art in those niches of literature aiming to extract useful information to analysts and experts in terms of relation between players' trajectories and team performance, this paper presents preliminary results from analysing trajectories data and sheds light on potential future research in this eld of study. In particular, using convex hulls, we find interesting regularities in players' movement patterns.
Data analytics mostly involves studying data trends over a given period, and then extracting useful information from these trends.
Why Is Data Analytics Important?
More precise decision making process: Data analytics helps organizations make more accurate decisions based on the insights gotten from data trends over time.
For example, a company selling different products can figure out what time of the year different products sell higher. This will enable them boost production of such products at the required time.
A better decision making process will eliminate the need for guess work, and minimize losses and avoidable risks.
Improved customer satisfaction: When you're able to serve customers, you retain them and keep business going. Insights gotten from data analytics can help you understand exactly what your customers want and when to act.
Data analytics also enables businesses to identify their target audience easily.
Improved business strategy: Data analytics helps organizations channel their resources towards the most efficient strategies.
Performance evaluation: Data analytics can help organizations evaluate how well or badly they've performed over a specified period. This will enable them make important decisions for the future of the organization.
Although the points listed above seem to be from the business point of view, that's not the only industry where data analytics is important.
You can see data analytics being used in healthcare, education, agriculture, and so on.
Types of Data Analytics
There are mainly four different types of data analytics:
Descriptive analytics: This type of analytics has to do with what happened with analyzed data over a specified period of time.
Diagnostic analytics: Diagnostic data analytics shows the "why" in a data trend. This involves having a deeper look into why certain patterns were present in the data.
Predictive analytics: The goal here is to foretell what is expected to happen in the future based on the outcomes of analyzed data over time.
Prescriptive analytics: In prescriptive analytics, the results from data analysis is used to make recommendations on what to do next.
What Is the Difference Between Data Analysis and Data Analytics?
You'll come across different definitions of data analytics and data analysis.
Some sources would define data analytics and data analysis as the same. Others would use them interchangeably.
Although, they are closely related, these terms have slightly different meanings. They are similar because they aid in the decision making process.
What Is Data Analysis?
Data analysis is the process of studying what has happened in the past in a dataset. There is no need to extend this definition.
Data analysis studies the why and how of data trends. Yes, it involves data collection, organization, and "analysis".
"How did the users respond to a new feature?".
"Why did the rate of purchase of a product fall during a particular period?".
Data analysts can make use o
As an engineer, I want to work in a company which understands the challenges of today and tomorrow. I wish to innovate all the time as it means better understanding the world around me.
I am enthusiastic about machine learning and I am studying for data science as it regroups mathematics, creating algorithms and investigating data.
This portfolio shows the projects I have worked on.
link to my linkedin account : https://fr.linkedin.com/in/pierre-masse
NIT1201 Introduction to Database System Assignment by USA ExpertsJohnsmith5188
The objective of this assignment is for you to put into practice the many different skills that you are learning in this unit into a single cohesive database project.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
1. A Statistical Model to Predict NBA
Play-off Results
Rishikesh Ravi
Under the guidance of Divya Ma’am
2. Contents
No Title
1 Introduction
2 Specifications for Hardware and Software
3 Data Flow Diagram
4 Content
5 Validation
7 Output
11 Thank You
10 Questions
9 Bibliography
8 Conclusion
6 Implementation
3. Introduction
This research presentation is on a topic that I am interested in - Sports Analytics. The project can
be important in the future as this algorithm or a modified version can be used for prediction of
various sports events and the roster construction of NBA teams in the future.
In this project, data from various sources have been collected. Various programming and statistics
techniques have been used. These techniques have been used to determine the relevance and
correlation of different variables in the data set. In addition, various approaches for feature
selection have been used. Logistic regression has been used for the prediction model.
The final objective was to Predict the Success of an NBA Team. The model was trained using a
training set (14 seasons between 2000 and 2020) and the algorithm’s prediction capability was
tested on testing data set(3 seasons between 2021 and 2023).
4. Specifications (For the hardware and software)
Hardware Specifications
Processor- Intel(R) Core(TM) i5-8265U CPU @ 1.60GHz 1.80 GHz
System Type-64-bit operating system, x64-based processor
Installed RAM- 8 GB
Software Specifications
Operating System- Windows 11 , version 22H2
Programming Language- Python 3.10.9
Developer platform- conda 23.3.1
5. Data Flow Diagram
Data set of NBA data Feature selection Data Cleaning
Train the
data
Test the
model on
new
dataset
Assess accuracy of
the output
Non cleaned data
Use trained model and
test data
6. About NBA
• NBA is the National Basketball Association
• It is a USA based basketball league where the best players
from all over the world come and play
• There are 2 conferences within NBA - the East and the West
• Within each conference there are 15 teams
7. About NBA Play-offs
• Each NBA team's success in the regular season is tested in the playoffs
• After regular season, based on performance, only 16 teams qualify for play-offs
• A qualified team can play a maximum of 4 series in the playoffs- The 1st round, the conference
semi-finals, the conference finals and the finals.
• Each series consists of a maximum of 7 games and the first team to win 4 games is the winner
• Using this program, I have tried to determine which regular season stats are predictors of the
post season and have tried to predict the top 4 teams in the playoffs (the teams in the
conference finals).
8. About the Project
• The dataset is built from NBA data
• I have built a Logistic Regression based model to Predict success in the play-offs
based on regular season player statistics
• Success metric – The Team should finish in the top 4 in its conference after Play-offs.
• This model will help us better understand what are the factors that drive success for
a Team in NBA
• They can use this for their team building and recruitment
9. Key Statistics that were considered
• 3 groups of statistics were used- Per game statistics, Advanced statistics and shooting statistics
• The per game statistics include- Minutes played by the team’s players per game, Field
goals attempted, Field Goals Made, Field goal percentage, 2-point attempts, 2 pointers made,
2-point percentage, 3-point attempts, 3-pointers made, 3-point percentage, Free throws
attempted, free throws made, Free throw percentage, Offensive rebounds per game, defensive
rebounds per game, Total rebounds per game, blocks per game, steals per game, assists per
game, points per game and Personal fouls per game
• The advanced statistics include- Average age of players, Offensive Rating, Net Rating,
Defensive Rating, Pace of play, Free throw attempt rate, 3- point attempt rate, True shooting
percentage, Effective field goal percentage of the team and their opponents, Turn over
percentage of the team and their opponents, Offensive and defensive rebound percentage of
the team and their opponents.
10. Key Statistics that were considered (Page 2)
• The shooting statistics include- Field Goal percentage, Shooting percentages and
attempts from 0- 3 feet , 3-10 feet, 10-16 feet and 16 feet till the 3-point line away from the
basket, Number of half court shots made and attempted, number of dunks made and
attempted, number of layups made and attempted, number of corner 3-point shots made
and attempted and average distance of each shot
11. Exploratory Data Analysis (Correlation Graph)
• A graph can be plotted in python to analyze the relation
of each statistic with every other statistic in the data.
This is the correlation graph.
• On the right, less features have been used to show how
the graph works. In this graph, all the statistics have a
positive correlation with each other (correlation varies
from 0 to 1). The lightest color represents the highest
correlation
12. The correlation plot for the entire data
• For this graph, the darkest and the
lightest box represent the highest
correlation within the data members
• Here the correlations are both negative
and positive
Exploratory Data Analysis (Correlation Graph)
13. Feature Selection
• Feature selection is the method of
filtering out the most important
data features from the features that
hold less importance.
• In this program, we have used
feature selection to find the values
which have the highest correlation
with the target variable, ‘TOP4’
14. Most Important features after feature selection- NRtg
Net rating (NRtg) is the difference
between the number of points made
and defended for every 100
possessions.
15. True Shooting percentage(TS%) is a
measure of the number of points scored
by a team per every possession.
Most Important features after feature selection- TS%
PTS- Points per game
FGA- Field Goals attempted per game
FTA- Free throws attempted per game
16. Most Important features after feature selection- AST
and PPG
Assists per game (AST) and points
per game (PTS) are 2 interrelated
stats.
More assists lead to more points
Assists per game (AST)
Points per game (PTS)
17. Most Important features after feature selection- eFG%
Effective field goal percentage (eFG%)
is a statistic of the percentage of shots made
while providing extra weightage to the 3
pointers made.
FGM= Field Goals Made
3PM- 3 Pointers Made
FGA- Field Goals attempted
18. Methodology- Logistic Regression
• Logistic regression is a technique used for solving
classification problems (where the target variable is either 1 or
0, true or false). It divides a data set into a training set (the set
which gives the data to the program to help in its prediction)
and the test set (the set in which the target variable is predicted
using training data)
• Threshold- The threshold in logistic regression can be
explained with an example- In a school exam, if the passing
mark is 40%, then 0.4 is the threshold. If a student is predicted
to score above 0.4, then 1 will be returned, otherwise 0.
20. Implementation
Step 1- define the objective and the success criteria.
Step 2- understand and identify the data
Step 3- Collect and prepare the data set
Step 4- Determine the model’s features (Feature selection)
Step 5- Train the data set
Step 6- Evaluate the model’s performance
Step 7- Put the model in operation
Step 8- Check the results and adjust the model accordingly
21. Validation
The error that the program shows is that it will call out null values (inputs with no values entered) and
will not be able to execute them.
22. Output results
• The outputs on the right side of the page show the
accuracy of the prediction. The 2 most important terms in
it are PRECISION and RECALL.
• If there are 40 teams which are ‘1’, and only 20 have been
predicted to be ‘1’, then 20/40 is the recall.
• Whereas, if out of the 20 teams predicted to be ‘1’, only 16
of them are actually ‘1’, then the precision is 16/20.
• The teams predicted as 1 and are actually 1 are true
positives. The teams predicted as 0 and are actually 0
are true negatives. The teams predicted as 1 and are
actually 0 are false positives and the teams predicted as 0
and are actually 1 are false negatives.
• From the predicted data set, there were 78 true
negatives, 6 false negatives, 6 true positives , and
0 false positives.
Output of the train set
Prediction output for test set
23. Conclusion
• A logistic regression model was built for this project.
• Success in the model was predicting that a team reached the conference finals (Top 4).
• All the 6 teams that the model predicted to reach the conference finals did reach the top four. For
example, 2022 Golden State warriors, the 2022 Boston Celtics and 2023 Boston Celtics.
• On the other hand, teams like 2023 Miami heat and 2023 Los Angeles Lakers were teams which made the
conference finals but were not predicted by the model.
• In order to improve the model, new statistics like number of all NBA players from the previous season or
number of MVP level players should be taken into consideration.
• Other techniques like Random Forest could also be used to improve the model
• This type of model could be used for other sports like cricket, football, etc
24. Bibliography
• YUBI internship training for data science
• Basketball-reference.com
• www.nbastuffer.com
• Python for machine learning and Data science bootcamp – A Udemy Course