NEW DELHI: For those obsessed with the size-zero, here's a phone as thin as paper. Called PaperPhone, the smartphone is presently in prototype stage. It uses latest printing technologies to print copper circuits and wiring on to a 9.5-centimetre surface. A layer of E Ink, used in Amazon Inc's Kindle eReader, is applied to act as a display. As for OS, it is powered by Google Android. To be unveiled at the forthcoming Association of Computing Machinery's CHI conference in Canada, the PaperPhone, has been developed by a team of researchers from Arizona State University, Queen's University, and E Ink Corporation. The 'flexible' phone can store books, play music and make phone calls. According to the researchers, bend gestures are fed into a gesture-recognition engine and can associate certain movements with certain instructions. As creator Roel Vertegaal, the director of Queen's University Human Media Lab told the The Vancouver Sun, "So you can bend the top in order to page forward or make a bookmark, you can navigate left and right on your home screen in order to open an icon, and you can make a call by squeezing the paper so that it curves, and then if you want to stop the call you pop it back into shape." "This is the future. Everything is going to look and feel like this within five years," he said. This computer looks, feels and operates like a small sheet of interactive paper. You interact with it by bending it into a cell phone, flipping the corner to turn pages, or writing on it with a pen, Vertegaal reportedly added. As for the pricing, while the prototype costs as high as $6,000 to $7,000, the device is likely to be priced less than $100.
NEW DELHI: For those obsessed with the size-zero, here's a phone as thin as paper. Called PaperPhone, the smartphone is presently in prototype stage. It uses latest printing technologies to print copper circuits and wiring on to a 9.5-centimetre surface. A layer of E Ink, used in Amazon Inc's Kindle eReader, is applied to act as a display. As for OS, it is powered by Google Android. To be unveiled at the forthcoming Association of Computing Machinery's CHI conference in Canada, the PaperPhone, has been developed by a team of researchers from Arizona State University, Queen's University, and E Ink Corporation. The 'flexible' phone can store books, play music and make phone calls. According to the researchers, bend gestures are fed into a gesture-recognition engine and can associate certain movements with certain instructions. As creator Roel Vertegaal, the director of Queen's University Human Media Lab told the The Vancouver Sun, "So you can bend the top in order to page forward or make a bookmark, you can navigate left and right on your home screen in order to open an icon, and you can make a call by squeezing the paper so that it curves, and then if you want to stop the call you pop it back into shape." "This is the future. Everything is going to look and feel like this within five years," he said. This computer looks, feels and operates like a small sheet of interactive paper. You interact with it by bending it into a cell phone, flipping the corner to turn pages, or writing on it with a pen, Vertegaal reportedly added. As for the pricing, while the prototype costs as high as $6,000 to $7,000, the device is likely to be priced less than $100.
it will be helpfull to analysis of match field hockey,
football,handball,
Match analysis was primarily used by the coaching team , nowadays ,professionalization is increasingly taking place with an exchange between head coach , coach assistants and analysts .
The players themselves are increasingly involved in the preparation and follow - up ..
Game observation and game evaluation
Everyone is talking about Data Mesh architectures already - assuming that there is already a full-fledged self-service data platform in place. A reality check reveals, that most (data) platforms are not really working that well, and fail to deliver value at scale. And in contrast to the business notion of a platform, where network effects make platforms even more valuable, the more users and products are there, this does not hold true for data platforms in particular (at least I haven't seen a proof so far).
So where to start, when data-transforming an organization? One approach, inspired by the Lean framework, is outlined in this talk. It all starts with what is actually working - identify some (data) products that drive value already. These are the ones you can build a platform for. It's a myth that you just need to build a solid platform, and then everyone will come and build amazing data products. They will never come. But starting with what already works is a reasonable first step. Step two is about creating flow, supporting the value stream end-to-end. Co-creation is your main tool here, fostering collaboration and ownership. Then you can think of platformizing what is really, really needed, avoiding the "waste" that modern data systems / platforms / architectures tend to pile up. In the end, the "right" architecture for your organization will emerge, you cannot simply copy-paste "solutions" that are not addressing your specific challenges.
Long story short, there is a path to success, but it's not easy, it's not copying others, it's finding your own way. And as in all good strategies, you can specify the "qualities" you'd like to see in the end. And the concrete solutions need to emerge from the hard work of the motivated people, that are already driving value for your organization now.
Data teams are contributing to a variety of value streams, as they are delivering value to a variety of stakeholders. The value streams are often not well-supported and the involved teams are facing constant challenges like Data Quality and Data Ownership. Also, data products often rely on the same data points for building the product and for measuring its success - so a lack of data quality leads to poor product quality and weak measurability at the same time. These challenges become exponentially harder, the larger the organization has grown. We propose a way of conceptualizing and visualizing the process of building data products, using the concept of the data value chain. Applying the Five Principles of Lean, especially Defining Value and Mapping out Data Value Streams, to the way build data products and operate data systems at scale, we create a framework that allows to focus on value delivery, avoids "waste" and supports ownership.
Talk at MCubed London about Manifold Learning and ApplicationsStefan Kühn
How to make use of of Manifold Learning methods for Dimensionality Reduction, Data Visualization and Automated Feature Engineering, this time also with UMAP - most of the cool stuff is in the Jupyter notebooks
Talk at PyData Berlin about Manifold Learning and ApplicationsStefan Kühn
These are the slides from my talk at PyData Berlin about how to use Manifold Learning in the context of Data Visualization and Feature Engineering. There are several jupyter notebooks exporing this, you can find the on github under https://github.com/cc-skuehn/Manifold_Learning
My slides from the minds mastering machines conference 2018 in Cologne about Deep Learning and Mathematical Optimization, the methods that are used for training Neural Nets and how they perform with respect to Training and especially Learning, i.e. how well the trained predictors generalize
Manifold Learning and Data VisualizationStefan Kühn
Talk at PyData Hamburg 2018-03-01 about Manifold Learning and Data Visualization with Python and Scikit-learn plus Random Projections and PCA, includes links to all resources and the github repository with worked examples in form of jupyter notebooks - we recommend using jupyter lab
More Related Content
Similar to SKuehn_Talk_FootballAnalytics_data2day2015
it will be helpfull to analysis of match field hockey,
football,handball,
Match analysis was primarily used by the coaching team , nowadays ,professionalization is increasingly taking place with an exchange between head coach , coach assistants and analysts .
The players themselves are increasingly involved in the preparation and follow - up ..
Game observation and game evaluation
Everyone is talking about Data Mesh architectures already - assuming that there is already a full-fledged self-service data platform in place. A reality check reveals, that most (data) platforms are not really working that well, and fail to deliver value at scale. And in contrast to the business notion of a platform, where network effects make platforms even more valuable, the more users and products are there, this does not hold true for data platforms in particular (at least I haven't seen a proof so far).
So where to start, when data-transforming an organization? One approach, inspired by the Lean framework, is outlined in this talk. It all starts with what is actually working - identify some (data) products that drive value already. These are the ones you can build a platform for. It's a myth that you just need to build a solid platform, and then everyone will come and build amazing data products. They will never come. But starting with what already works is a reasonable first step. Step two is about creating flow, supporting the value stream end-to-end. Co-creation is your main tool here, fostering collaboration and ownership. Then you can think of platformizing what is really, really needed, avoiding the "waste" that modern data systems / platforms / architectures tend to pile up. In the end, the "right" architecture for your organization will emerge, you cannot simply copy-paste "solutions" that are not addressing your specific challenges.
Long story short, there is a path to success, but it's not easy, it's not copying others, it's finding your own way. And as in all good strategies, you can specify the "qualities" you'd like to see in the end. And the concrete solutions need to emerge from the hard work of the motivated people, that are already driving value for your organization now.
Data teams are contributing to a variety of value streams, as they are delivering value to a variety of stakeholders. The value streams are often not well-supported and the involved teams are facing constant challenges like Data Quality and Data Ownership. Also, data products often rely on the same data points for building the product and for measuring its success - so a lack of data quality leads to poor product quality and weak measurability at the same time. These challenges become exponentially harder, the larger the organization has grown. We propose a way of conceptualizing and visualizing the process of building data products, using the concept of the data value chain. Applying the Five Principles of Lean, especially Defining Value and Mapping out Data Value Streams, to the way build data products and operate data systems at scale, we create a framework that allows to focus on value delivery, avoids "waste" and supports ownership.
Talk at MCubed London about Manifold Learning and ApplicationsStefan Kühn
How to make use of of Manifold Learning methods for Dimensionality Reduction, Data Visualization and Automated Feature Engineering, this time also with UMAP - most of the cool stuff is in the Jupyter notebooks
Talk at PyData Berlin about Manifold Learning and ApplicationsStefan Kühn
These are the slides from my talk at PyData Berlin about how to use Manifold Learning in the context of Data Visualization and Feature Engineering. There are several jupyter notebooks exporing this, you can find the on github under https://github.com/cc-skuehn/Manifold_Learning
My slides from the minds mastering machines conference 2018 in Cologne about Deep Learning and Mathematical Optimization, the methods that are used for training Neural Nets and how they perform with respect to Training and especially Learning, i.e. how well the trained predictors generalize
Manifold Learning and Data VisualizationStefan Kühn
Talk at PyData Hamburg 2018-03-01 about Manifold Learning and Data Visualization with Python and Scikit-learn plus Random Projections and PCA, includes links to all resources and the github repository with worked examples in form of jupyter notebooks - we recommend using jupyter lab
Talk at the Data Science Meetup Hamburg about Deep Learning, the most important Optimization methods in this field and the relationship between training and learning
Visualizing and Communicating High-dimensional DataStefan Kühn
Slides from my talk at Data Natives, starting with the different Modes of Perception, the components of Visualization and Graphics and how to transport Information efficiently, then giving examples of how modern approximation techniques - manifold learning, principal curves - and visualization techniques - pair plots, correlation plots, parallel coordinates, grand tour - can be used in order to approach complex multi-dimensional data.
Data quality - The True Big Data ChallengeStefan Kühn
Data Quality is one of the most-overlooked key aspect in any Big Data project or approach. This talk adresses the problem from various perspectives, discusses the main challenges and identifies possible solutions.
In this talk we discuss the connections between (Supervised) Learning and Mathematical Optimization. Topics include iterative algorithm, search directions and stepsizes. The talk has been held at the Computer Science, Machine Learning and Statistics Meetup Hamburg.
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeWalaa Eldin Moustafa
Dynamic policy enforcement is becoming an increasingly important topic in today’s world where data privacy and compliance is a top priority for companies, individuals, and regulators alike. In these slides, we discuss how LinkedIn implements a powerful dynamic policy enforcement engine, called ViewShift, and integrates it within its data lake. We show the query engine architecture and how catalog implementations can automatically route table resolutions to compliance-enforcing SQL views. Such views have a set of very interesting properties: (1) They are auto-generated from declarative data annotations. (2) They respect user-level consent and preferences (3) They are context-aware, encoding a different set of transformations for different use cases (4) They are portable; while the SQL logic is only implemented in one SQL dialect, it is accessible in all engines.
#SQL #Views #Privacy #Compliance #DataLake
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...sameer shah
"Join us for STATATHON, a dynamic 2-day event dedicated to exploring statistical knowledge and its real-world applications. From theory to practice, participants engage in intensive learning sessions, workshops, and challenges, fostering a deeper understanding of statistical methodologies and their significance in various fields."
1. How smart is
Football Data
Analytics today?
Dr. Stefan Kühn
data2day - Karlsruhe
29.09.2015
2. Topic
Why Football Data Analytics?
• It’s about Football
• There is a lot of data out there
• There is a lot of ignorance out there
• Three examples
• Corners
• Marginal goals
• Substitutions
• Alternatives
2
3. Infos
Why Football is an interesting Use Case
• 209 FIFA federations - worldwide
• Most popular sport - 3.3-3.5 billion fans
• Monetary facts - revenue (Deloitte Money League)
• Real Madrid 2013/4: 549.5 Million € (Position 1)
• Bayern Munich 2013/4: 487.5 Million € (Position 3)
• Everton 2013/4: 144.1 Million € (Position 20)
• Social Media facts (Deloitte Money League)
• Facebook: FC Barcelona - 81.4 Million Likes
• Twitter: Real Madrid - 14.4 Million Followers
3
4. Some Stats
Why Football is a Data Use Case
• 306 Bundesliga matches per season
• 2000+ recorded events per match
• 512 Bundesliga players
• Live Statistics (Opta, Prozone etc.):
• Shots, Passes, Assists
• Tacklings, Blocks, intercepted Passes
• Saves and other actions of Goalkeepers
• Fouls and Foul types
• Position Data including time stamps
• 1.8 Million Amateur matches (Deutschland)
4
5. Some Remarks
Is there anything left to do?
• Big companies like SAP are involved
• Players are tracked in training and matches (and
sometimes at home as well)
• Physiological data, nutrition data, training plans
★ BUT:
Big data is not about the data.
(Gary King, Harvard University, 2013)
It’s about Analytics.
5
6. Some Remarks
Where is the ignorance?
• „The Number’s Game - Why Everything You
Know About Football Is Wrong“
• Book by Chris Anderson (former Cornell University
Prof) and David Sally (Economics and Behavioral
Game Theory)
• „Is it easier to score as a sub“?
• Blogpost by Dan Altman, founder of North Yard
Analytics
6
8. Corners
Claim: Long corners are overrated, short
corners are better, see e.g. Barca.
8
Long corners versus Short corners
9. Corners
Some useful stats
• Average number of goals per team per match: 1.3
• Average number of corners per team per match: 5
• Long corners account for ~8.5% of all goals
• Silly question: The average team scores once
every ten games from a penalty, shall they give
up on penalties as well?
• Lack of relevant context
• How efficient are the alternatives?
• How efficient is the average possession?
9
10. Corners
Average Possession
• Average number of possessions per team per match: 200
• Average number of goals per team per match: 1.3
• Expectation value per possession: 0.0065
• Normalized per match (200 possessions):
• All possessions are corners: 4.4 goals
• Half of the possessions are corner: 2.85 goals
• 10% of the possessions are corners: 1.46 goals
• The efficiency of long corners is more than three times
as high as the efficiency of the average possession.
• Still unknown:
• How efficient are the alternatives?
• Are there any negative counter effects?
10
15. Marginal goals
Why they should have bought a book on hypothesis testing
• How many second goals could have been scored without the first goal?
• Do the samples for matches with one (own) goal, two goals etc. differ,
and if yes (it’s a definite yes, selection bias): how?
• Is it more likely to score more against weaker teams and less against
stronger teams?
• And of course: The events considered here are not statistically
independent.
15
What they should have done
• Compute marginal goals per sample group (e.g. fixed number of own goals).
Here, the first goal cannot have less marginal points than the second goal etc.
which is the only reasonable result.
• Do not compare apples and pies. (In some sense Simpson’s paradox)
• Or: Hire the best striker for first goals and the best striker for second goals.
18. Substitutions and Scoring
Claim
Subs score more
than expected
• This is the first
correct claim!
• But still weak
effect, unknown
reason(s)
• Do opponents
score more as
well?
• Corrections needed
• 36% of subs are
forwards
• Individual Orders
• Tactical changes
• Lots of other things
18
20. Substitutions and Scoring
A closer look
Estimates for
the mean for
first and
second half
• Analysis:
No control for
fatigue
possible, only
control for
time spent on
the field.
20
From minute 60
on the share of
subs starts to
rise. Effect on
number of goals?
22. Summary
What are the commonalities in all cases?
• „New“ spectacular insights
• Preconceptions
• Confirmation Bias
• Lack of reflection
• Challenging own results?
• Alternative explanations?
• Do not mix up a variable and your interpretation
of this variable (fatigue vs. time on field)
• BUT: Data and Tools have been good!
22
24. What keeps Football Data Analytics from being smart?
24
Requirements
+ Scientific Method!
Reality
Tools Data
Money
???
+ Severe Time Constraint
+ Results must impress
25. What keeps Data Analytics from being smart?
25
Requirements
+ Scientific Method!
Reality
Tools Data
Money
???
+ Severe Time Constraint
+ Results must impress