Neural network Intuition
Neural network and vocabulary
Neural network algorithm
Math behind neural network algorithm
Building the neural networks
Validating the neural network model
Neural network applications
Image recognition using neural networks
Presentazione Tesi Laurea Triennale in InformaticaLuca Marignati
Università degli Studi di Torino
Dipartimento di Informatica
Titolo: Apprendimento per Rinforzo e Applicazione ai Problemi di Pianificazione del Percorso
Topic: Machine Learning
Neural network Intuition
Neural network and vocabulary
Neural network algorithm
Math behind neural network algorithm
Building the neural networks
Validating the neural network model
Neural network applications
Image recognition using neural networks
Presentazione Tesi Laurea Triennale in InformaticaLuca Marignati
Università degli Studi di Torino
Dipartimento di Informatica
Titolo: Apprendimento per Rinforzo e Applicazione ai Problemi di Pianificazione del Percorso
Topic: Machine Learning
Interactive Recommender Systems with Netflix and SpotifyChris Johnson
Interactive recommender systems enable the user to steer the received recommendations in the desired direction through explicit interaction with the system. In the larger ecosystem of recommender systems used on a website, it is positioned between a lean-back recommendation experience and an active search for a specific piece of content. Besides this aspect, we will discuss several parts that are especially important for interactive recommender systems, including the following: design of the user interface and its tight integration with the algorithm in the back-end; computational efficiency of the recommender algorithm; as well as choosing the right balance between exploiting the feedback from the user as to provide relevant recommendations, and enabling the user to explore the catalog and steer the recommendations in the desired direction.
In particular, we will explore the field of interactive video and music recommendations and their application at Netflix and Spotify. We outline some of the user-experiences built, and discuss the approaches followed to tackle the various aspects of interactive recommendations. We present our insights from user studies and A/B tests.
The tutorial targets researchers and practitioners in the field of recommender systems, and will give the participants a unique opportunity to learn about the various aspects of interactive recommender systems in the video and music domain. The tutorial assumes familiarity with the common methods of recommender systems.
Slides deck used by Praveen Devarao for Apache Spark Machine Learning session organized by Bangalore Spark enthusiasts meetup group @ IBM campus on 10th September 2016
Demo notebook used can be found at https://gist.github.com/praveend/fe9a0c5eacd6b43ee210e88a374eb230
Concept Location using Information Retrieval and Relevance FeedbackSonia Haiduc
Concept location is a critical activity during
software evolution as it produces the location where a change is to start in response to a modification
request, such as, a bug report or a new feature request. Lexical-based concept location techniques rely on matching the text embedded in the source code to queries formulated by the developers. The efficiency of such techniques is strongly dependent on the ability of the developer to write good queries. We propose an approach to augment information retrieval (IR) based concept location via an explicit relevance feedback (RF) mechanism. RF is a two-part process in which the developer judges existing results returned by a search and the IR system uses this information to perform a new search, returning more relevant information to the user. A set of case studies performed on open source software systems reveals the impact of RF on IR based concept location.
Inference & Learning in Linear-Chain Conditional Random Fields (CRFs)Anmol Dwivedi
Find the code on: https://github.com/anmold07/Graphical_Models/tree/master/CRF%20Learning
Probabilistic Graphical Models (PGMs) provides a general
framework to model dependencies among the output variables. Among the family of graphical models include Neural Networks, Markov Networks, Ising Models, factor graphs, Bayesian Networks etc, however, this project considers linear chain Conditional Random Fields to learn the inter-dependencies among the output variables for efficient classification of handwritten word recognition. Such models are capable of representing a complex distribution over multivariate distributions as a product of local factor functions.
Find all the relevant code on: https://github.com/anmold-07/Graphical_Models
Presentation made during the Intelligent User-Adapted Interfaces: Design and Multi-Modal Evaluation Workshop (IUadaptME) workshop conducted as part of UMAP 2018
Predicting Azure Churn with Deep Learning and Explaining Predictions with LIMEFeng Zhu
Although deep learning has proved to be very powerful, few results are reported on its application to business-focused problems. Feng Zhu and Val Fontama explore how Microsoft built a deep learning-based churn predictive model and demonstrate how to explain the predictions using LIME—a novel algorithm published in KDD 2016—to make the black box models more transparent and accessible.
Continuous Evaluation of Collaborative Recommender Systems in Data Stream Man...Dr. Cornelius Ludmann
Talk at the Data Streams and Event Processing Workshop at the 16. Fachtagung »Datenbanksysteme für Business, Technologie und Web« (BTW) of the Gesellschaft für Informatik (GI) in Hamburg, Germany. March 3, 2015
Interactive Recommender Systems with Netflix and SpotifyChris Johnson
Interactive recommender systems enable the user to steer the received recommendations in the desired direction through explicit interaction with the system. In the larger ecosystem of recommender systems used on a website, it is positioned between a lean-back recommendation experience and an active search for a specific piece of content. Besides this aspect, we will discuss several parts that are especially important for interactive recommender systems, including the following: design of the user interface and its tight integration with the algorithm in the back-end; computational efficiency of the recommender algorithm; as well as choosing the right balance between exploiting the feedback from the user as to provide relevant recommendations, and enabling the user to explore the catalog and steer the recommendations in the desired direction.
In particular, we will explore the field of interactive video and music recommendations and their application at Netflix and Spotify. We outline some of the user-experiences built, and discuss the approaches followed to tackle the various aspects of interactive recommendations. We present our insights from user studies and A/B tests.
The tutorial targets researchers and practitioners in the field of recommender systems, and will give the participants a unique opportunity to learn about the various aspects of interactive recommender systems in the video and music domain. The tutorial assumes familiarity with the common methods of recommender systems.
Slides deck used by Praveen Devarao for Apache Spark Machine Learning session organized by Bangalore Spark enthusiasts meetup group @ IBM campus on 10th September 2016
Demo notebook used can be found at https://gist.github.com/praveend/fe9a0c5eacd6b43ee210e88a374eb230
Concept Location using Information Retrieval and Relevance FeedbackSonia Haiduc
Concept location is a critical activity during
software evolution as it produces the location where a change is to start in response to a modification
request, such as, a bug report or a new feature request. Lexical-based concept location techniques rely on matching the text embedded in the source code to queries formulated by the developers. The efficiency of such techniques is strongly dependent on the ability of the developer to write good queries. We propose an approach to augment information retrieval (IR) based concept location via an explicit relevance feedback (RF) mechanism. RF is a two-part process in which the developer judges existing results returned by a search and the IR system uses this information to perform a new search, returning more relevant information to the user. A set of case studies performed on open source software systems reveals the impact of RF on IR based concept location.
Inference & Learning in Linear-Chain Conditional Random Fields (CRFs)Anmol Dwivedi
Find the code on: https://github.com/anmold07/Graphical_Models/tree/master/CRF%20Learning
Probabilistic Graphical Models (PGMs) provides a general
framework to model dependencies among the output variables. Among the family of graphical models include Neural Networks, Markov Networks, Ising Models, factor graphs, Bayesian Networks etc, however, this project considers linear chain Conditional Random Fields to learn the inter-dependencies among the output variables for efficient classification of handwritten word recognition. Such models are capable of representing a complex distribution over multivariate distributions as a product of local factor functions.
Find all the relevant code on: https://github.com/anmold-07/Graphical_Models
Presentation made during the Intelligent User-Adapted Interfaces: Design and Multi-Modal Evaluation Workshop (IUadaptME) workshop conducted as part of UMAP 2018
Predicting Azure Churn with Deep Learning and Explaining Predictions with LIMEFeng Zhu
Although deep learning has proved to be very powerful, few results are reported on its application to business-focused problems. Feng Zhu and Val Fontama explore how Microsoft built a deep learning-based churn predictive model and demonstrate how to explain the predictions using LIME—a novel algorithm published in KDD 2016—to make the black box models more transparent and accessible.
Continuous Evaluation of Collaborative Recommender Systems in Data Stream Man...Dr. Cornelius Ludmann
Talk at the Data Streams and Event Processing Workshop at the 16. Fachtagung »Datenbanksysteme für Business, Technologie und Web« (BTW) of the Gesellschaft für Informatik (GI) in Hamburg, Germany. March 3, 2015
A seminar in advanced Software Engineering concerning using models to guide the development process, and QVT to transfer a model into another model automatically
These slides are about KNN algorithm used in Machine Learning where a C++ made KNN algorithm is compared with an actual KNN running in WEKA (Machine Learning software).
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Gaussian Ranking by Matrix Factorization, ACM RecSys Conference 2015
1. Gaussian
Ranking
by
Matrix
Factoriza5on
Harald
Steck
hsteck@netflix.com
RecSys
2015
2. Overview
• Matrix
Factoriza<on
Model
• asymmetric
MF
• Objec<ve:
op<mize
various
Ranking
Metrics
•
exploit
proper<es
of
MF
model
&
implicit
data
• Training:
pointwise
&
listwise
• Related
Work
• Experiments
3.
Basic
Idea:
data
.
items
i
users
u
≈
users
u
Low-‐rank
Matrix
Factoriza<on
Model
4. Basic
Idea:
-‐
latent
user
vector:
-‐
by
[Paterek
07],
extended
to
SVD++
[Koren
08]
Asymmetric
Matrix
Factoriza<on
5. Overview
• Matrix
Factoriza<on
Model
• asymmetric
MF
• Objec5ve:
op5mize
various
Ranking
Metrics
•
exploit
proper5es
of
MF
model
&
implicit
data
• Training:
pointwise
&
listwise
• Related
Work
• Experiments
6. AMF
as
Neural
Network
rank
loss
=
f
(ranks)
items
i
…
click
history
…
user
vec.
…
scores
…
ranks
7. AMF
as
Neural
Network
rank
loss
=
f
(ranks)
items
i
…
click
history
…
user
vec.
…
scores
…
ranks
8. 1st
term:
Rank
Loss
example
1:
AUC
• pairwise
comparisons
!
(linear)
sum
of
ranks
9. example
2:
nDCG
(for
binary
relevance)
• emphasizes
top
of
ranked
list
• also
a
func<on
of
the
ranks
of
the
posi<ves
1st
term:
Rank
Loss
10. 2nd
term:
Ac<va<on
Func<on
T
Scores
!
Ranks:
+
+
+
-‐
binary
data:
nega<ves
and
posi<ves
-‐
sparse
data:
many
few
!
MF
scores:
Gaussian
distrib.
assumed
scores
i
14. Pueng
it
All
Together
training
objec<ve
func<on:
rank
prior
on
param’s
scores
of
loss
"
lambda
nega<ves
"gamma
-‐
minimized
by
stochas<c
gradient
descent
15. Overview
• Matrix
Factoriza<on
Model
• asymmetric
MF
• Objec<ve:
op<mize
various
Ranking
Metrics
•
exploit
proper<es
of
MF
model
&
data
• Training:
pointwise
&
listwise
• Related
Work
• Experiments
16. Listwise
Approach
• consider
ALL
items
for
each
user:
-‐ es<mate
standard
devia<on
of
scores
for
each
user
!
width
of
ac<va<on
func<on
17. Listwise
Approach
• consider
ALL
items
for
each
user:
-‐
sort
by
scores
!
exact
ranks
-‐
using
logis<c
ac<va<on
func<on:
2nd
term
in
chain
rule
18. AUC
nDCG
Listwise
Approach
deriva5ves
L’:
1st
&
2nd
terms
top
of
ranked
list
19. !
between
nDCG
and
AUC:
L’
=
constant
!
use
very
large
std.
for
ac<va<on
func<on
in
pointwise
approach
AUC
nDCG
Pointwise
Approach
deriva5ves
L’:
top
of
ranked
list
20. Overview
• Matrix
Factoriza<on
Model
• asymmetric
MF
• Objec<ve:
op<mize
various
Ranking
Metrics
•
exploit
proper<es
of
MF
model
&
data
• Training
• Related
Work
• Experiments
21. Related
Work
• various
learning-‐to-‐rank
approaches
exist
• ogen
tailored
to
specific
ranking
losses
• mostly
pairwise
approaches,
eg:
• AUC:
BPR
[Rendle
et
al.
’09]
• MRR:
CLiMF
[Shi
et
al.
’12]
used
as
• MAP:
TFMAP
[Shi
et
al.
‘12]
baselines
• listwise
approaches,
eg:
•
top-‐1
[Shi
et
al.
’10]
...
like
neural
network
• …
addi<onal
references
in
the
paper
22. Overview
• Matrix
Factoriza<on
Model
• basic
MF
!
asymmetric
MF
!
Neural
Network
• Objec<ve:
op<mize
various
Ranking
Metrics
•
exploit
proper<es
of
MF
model
&
data
• Training
• Related
Work
• Experiments
23. 10
m
MovieLens
Data
• 10k
movies
&
70k
users
• 1%
dense
data
• binarized:
3+
star
ra<ng
!
1,
otherwise
0
• 5-‐fold
cross-‐valida<on
24. 10
m
MovieLens
Data
5-‐fold
cross-‐valida<on
std
:
0.001
28. Conclusions
• learning-‐to-‐rank
approach:
– implicit
feedback
data
– proper<es
of
MF
model
! Gaussian
distribu<on
of
scores
! non-‐linear
ac<va<on
func<ons
derived
for
ranking
• pointwise
and
listwise
training
• various
ranking
metrics
can
be
used:
– compe<<ve
for
op<mizing
AUC
– par<cularly
effec<ve
at
head
of
ranked
list