Paradigmas de Programação - Imperativo, Orientado a Objetos e FuncionalGustavo Coutinho
1. A aula aborda os três principais paradigmas de programação: imperativo, orientado a objetos e funcional.
2. O paradigma imperativo é baseado na arquitetura de von Neumann e tem no coração a idéia de atribuição. Suporta declaração de variáveis, estruturas de controle e abstração procedural.
3. O paradigma orientado a objetos trata programas como coleções de objetos que se comunicam, concentrando responsabilidades em classes. Conceitos como herança, polimorfismo e interfaces são abordados.
4
Full Tutorial With Pictures: https://www.scienceez.com/build-recommender-system/
Macedonian Computer - Science Faculty (FCSE) Lecture by PhD. Andrea Kulakov. Topic: Recommender Systems
The document describes a generative model for networks called the Affiliation Graph Model (AGM). The AGM models how communities in a network "generate" the edges between nodes. It represents each node's membership in multiple communities as strengths in a membership matrix. The probability of an edge between two nodes depends on the product of their membership strengths in common communities. The maximum likelihood estimation technique can be used to estimate the community membership strengths matrix that best explains a given network.
The document discusses several collaborative filtering techniques for making recommendations:
1) Nearest neighbor techniques like k-NN make predictions based on the ratings of similar users. They require storing all user data but can be fast with appropriate data structures.
2) Naive Bayes classifiers treat each item's ratings independently; they make strong assumptions but require less data.
3) Dimensionality reduction techniques like SVD decompose the user-item rating matrix to find latent factors. Weighted SVD handles missing data.
4) Probabilistic models like mixtures of multinomials and aspect models represent additional user metadata but have more parameters.
The document discusses several collaborative filtering techniques for making recommendations, including k-nearest neighbors (kNN), naive Bayes classification, singular value decomposition (SVD), and probabilistic models. It provides examples of how these methods work, such as using ratings from similar users to predict a user's rating for an item (kNN), and decomposing a ratings matrix to capture relationships between users and items (SVD). The techniques vary in their assumptions, complexity, and ability to incorporate additional user/item metadata. Evaluation on new data is important to ensure the methods generalize well beyond the training data.
This document summarizes material from the book "Mining of Massive Datasets" by Jure Leskovec, Anand Rajaraman, and Jeff Ullman. It presents an algorithm called BigCLAM that can efficiently detect overlapping communities in large networks. BigCLAM models community membership using a strength matrix and optimizes the likelihood of the model to find communities. It scales to networks with millions of edges using techniques like caching neighbor sums. Experiments show BigCLAM can analyze networks orders of magnitude larger than previous methods in minutes instead of days.
This document describes a technique called MinHashing that can be used to efficiently find near-duplicate documents among a large collection. MinHashing works in three steps: 1) it converts documents to sets of shingles, 2) it computes signatures for the sets using MinHashing to preserve similarity, 3) it uses Locality-Sensitive Hashing to focus on signature pairs likely to be from similar documents, finding candidates efficiently. This avoids comparing all possible document pairs.
The document proposes a framework for recommendations based on analyzing relationships between users, items, tags, and ratings (quaternary relationships). It models these relationships using a 4-order tensor and applies Higher-Order Singular Value Decomposition (HOSVD) to reveal latent semantic associations. This allows generating recommendations for users, items, tags, and predicting ratings. Experimental results on a movie dataset show the proposed quaternary approach outperforms methods using only ternary relationships.
Paradigmas de Programação - Imperativo, Orientado a Objetos e FuncionalGustavo Coutinho
1. A aula aborda os três principais paradigmas de programação: imperativo, orientado a objetos e funcional.
2. O paradigma imperativo é baseado na arquitetura de von Neumann e tem no coração a idéia de atribuição. Suporta declaração de variáveis, estruturas de controle e abstração procedural.
3. O paradigma orientado a objetos trata programas como coleções de objetos que se comunicam, concentrando responsabilidades em classes. Conceitos como herança, polimorfismo e interfaces são abordados.
4
Full Tutorial With Pictures: https://www.scienceez.com/build-recommender-system/
Macedonian Computer - Science Faculty (FCSE) Lecture by PhD. Andrea Kulakov. Topic: Recommender Systems
The document describes a generative model for networks called the Affiliation Graph Model (AGM). The AGM models how communities in a network "generate" the edges between nodes. It represents each node's membership in multiple communities as strengths in a membership matrix. The probability of an edge between two nodes depends on the product of their membership strengths in common communities. The maximum likelihood estimation technique can be used to estimate the community membership strengths matrix that best explains a given network.
The document discusses several collaborative filtering techniques for making recommendations:
1) Nearest neighbor techniques like k-NN make predictions based on the ratings of similar users. They require storing all user data but can be fast with appropriate data structures.
2) Naive Bayes classifiers treat each item's ratings independently; they make strong assumptions but require less data.
3) Dimensionality reduction techniques like SVD decompose the user-item rating matrix to find latent factors. Weighted SVD handles missing data.
4) Probabilistic models like mixtures of multinomials and aspect models represent additional user metadata but have more parameters.
The document discusses several collaborative filtering techniques for making recommendations, including k-nearest neighbors (kNN), naive Bayes classification, singular value decomposition (SVD), and probabilistic models. It provides examples of how these methods work, such as using ratings from similar users to predict a user's rating for an item (kNN), and decomposing a ratings matrix to capture relationships between users and items (SVD). The techniques vary in their assumptions, complexity, and ability to incorporate additional user/item metadata. Evaluation on new data is important to ensure the methods generalize well beyond the training data.
This document summarizes material from the book "Mining of Massive Datasets" by Jure Leskovec, Anand Rajaraman, and Jeff Ullman. It presents an algorithm called BigCLAM that can efficiently detect overlapping communities in large networks. BigCLAM models community membership using a strength matrix and optimizes the likelihood of the model to find communities. It scales to networks with millions of edges using techniques like caching neighbor sums. Experiments show BigCLAM can analyze networks orders of magnitude larger than previous methods in minutes instead of days.
This document describes a technique called MinHashing that can be used to efficiently find near-duplicate documents among a large collection. MinHashing works in three steps: 1) it converts documents to sets of shingles, 2) it computes signatures for the sets using MinHashing to preserve similarity, 3) it uses Locality-Sensitive Hashing to focus on signature pairs likely to be from similar documents, finding candidates efficiently. This avoids comparing all possible document pairs.
The document proposes a framework for recommendations based on analyzing relationships between users, items, tags, and ratings (quaternary relationships). It models these relationships using a 4-order tensor and applies Higher-Order Singular Value Decomposition (HOSVD) to reveal latent semantic associations. This allows generating recommendations for users, items, tags, and predicting ratings. Experimental results on a movie dataset show the proposed quaternary approach outperforms methods using only ternary relationships.
Regression and Classification: An Artificial Neural Network ApproachKhulna University
This presentation introduces artificial neural networks (ANN) as a technique for regression and classification problems. It provides historical context on the development of ANN, describes common network structures and activation functions, and the backpropagation algorithm for training networks. Experimental results on 7 datasets show ANN outperformed other methods for both regression and classification across a variety of problem types and data characteristics. Limitations of ANN and areas for further research are also discussed.
Recommender Systems: Advances in Collaborative FilteringChangsung Moon
This document summarizes recommender systems, focusing on collaborative filtering techniques. It discusses how recommender systems help with information overload by matching users with relevant items. Collaborative filtering is introduced as a technique that seeks to predict user preferences based on other similar users' ratings. The document then covers various collaborative filtering algorithms like neighborhood models, latent factor models using matrix factorization, and extensions like adding biases and temporal dynamics. It concludes by discussing hybrid methods and providing references for further reading.
Computing Neighbourhoods with Language Models in a Collaborative Filtering Sc...Daniel Valcarce
Slides of the presentation given at IIR 2016 for the following extended abstract:
Daniel Valcarce, Javier Parapar, Alvaro Barreiro: Computing Neighbourhoods with Language Models in a Collaborative Filtering Scenario. IIR 2016, Venice, Italy.
http://dx.doi.org/10.1007/978-3-319-30671-1_45
Recommender systems aim to recommend items like books, movies, or products to users based on their preferences. There are two main approaches: collaborative filtering, which recommends items liked by similar users, and content-based filtering, which recommends items similar to those a user has liked based on item attributes. Both have strengths and weaknesses, so hybrid systems combining the approaches can provide the best recommendations.
Random Walk by User Trust and Temporal Issues toward Sparsity Problem in Soci...Sc Huang
Random Walk by User Trust and Temporal Issues toward Sparsity Problem in Social Tagging Recommender Systems
The document discusses using random walks and temporal factors to address sparsity problems in social tagging recommender systems. It introduces related work on item-based collaborative filtering, random walk recommendations, and models that learn influence probabilities. It then describes using random walks starting from users or items, and incorporating trust networks and influence powers to provide recommendations. Finally, it discusses addressing cold start problems, temporal decay issues, and experiment design.
A new similarity measurement based on hellinger distance for collaborating fi...Prabhu Kumar
This project proposed a similarity measurement which is focusing on recommendation performance under the cold start problem [The problem which occurs in the recommendation for newly comer items and users, which doesn't have any recognition in the system] and also perfectly suitable for sparse data set.
This technique solves the problem of the cold start in recommender system as well as improves the performance of recommendation to the users.
Recommender systems are software tools and techniques providing suggestions for items to be of interest to a user. Recommender systems have proved in recent years to be a valuable means of helping Web users by providing useful and effective recommendations or suggestions.
Matrix Factorization In Recommender SystemsYONG ZHENG
The document discusses matrix factorization techniques for recommender systems. It begins with an overview of recommender systems and their use of matrix factorization for dimensionality reduction. Principal component analysis and singular value decomposition are described as early linear algebra techniques used for this purpose. The document then focuses on how these techniques evolved into basic and extended matrix factorization methods in recommender systems, using the Netflix Prize competition as an example.
Deep Reinforcement Learning based Recommendation with Explicit User-ItemInter...Kishor Datta Gupta
—Recommendation is crucial in both academia andindustry, and various techniques are proposed such as content-based collaborative filtering, matrix factorization, logistic re-gression, factorization machines, neural networks and multi-armed bandits. However, most of the previous studies sufferfrom two limitations: (1) considering the recommendation asa static procedure and ignoring the dynamic interactive naturebetween users and the recommender systems; (2) focusing on theimmediate feedback of recommended items and neglecting thelong-term rewards. To address the two limitations, in this paperwe propose a novel recommendation framework based on deepreinforcement learning, called DRR. The DRR framework treatsrecommendation as a sequential decision making procedure andadopts an “Actor-Critic” reinforcement learning scheme to modelthe interactions between the users and recommender systems,which can consider both the dynamic adaptation and long-term rewards. Further more, a state representation module isincorporated into DRR, which can explicitly capture the interac-tions between items and users. Three instantiation structures aredeveloped. Extensive experiments on four real-world datasets areconducted under both the offline and online evaluation settings.The experimental results demonstrate the proposed DRR methodindeed outperforms the state-of-the-art competitors
The document summarizes recommender systems and different recommendation approaches. It introduces content-based recommendation which recommends items similar to those a user preferred. It also covers collaborative filtering which recommends items liked by similar users. Specific collaborative filtering techniques discussed include k-nearest neighbors, association rules, and matrix factorization using singular value decomposition as was done for the Netflix Prize contest.
This document outlines a proposed framework called TKmeans++ for identifying grey sheep users (GSU) and recommending items to them based on trust relations. The framework has two phases: a GSU identification phase that calculates user weights based on similarity, influence, and trust to assign users to clusters, and a recommendation phase that recommends top items to a GSU based on items positively rated by other GSU. An experimental study on the Epinions dataset shows TKmeans++ outperforms other clustering algorithms on MAE and coverage metrics for recommending to GSU. Future work could explore matrix factorization approaches or combining clustering and matrix factorization.
A Randomized Approach for Crowdsourcing in the Presence of Multiple Viewscollwe
The document presents a framework called M2VW for crowdsourcing in the presence of multiple views. M2VW models the multi-view learning problem using crowdsourced labels. It formulates an optimization problem to minimize prediction errors while encouraging worker consensus across views and feature sparsity. A randomized block coordinate descent algorithm is proposed to solve the optimization efficiently in linear time with respect to the number of workers. Experimental results on real and synthetic datasets demonstrate that M2VW achieves better performance than baseline methods and scales linearly with the number of workers.
This document provides an introduction to deep reinforcement learning. It begins with an overview of reinforcement learning and its key characteristics such as using reward signals rather than supervision and sequential decision making. The document then covers the formulation of reinforcement learning problems using Markov decision processes and the typical components of an RL agent including policies, value functions, and models. It discusses popular RL algorithms like Q-learning, deep Q-networks, and policy gradient methods. The document concludes by outlining some potential applications of deep reinforcement learning and recommending further educational resources.
Recommender systems help users deal with overwhelming choices by providing personalized recommendations. They are commonly used by websites like Amazon, Netflix, and YouTube. Research on recommender systems has grown significantly over the past 20 years. Common recommendation models include collaborative filtering, which predicts ratings based on similar users or items, and matrix factorization, which represents users and items as vectors in a latent space. Transfer learning techniques allow knowledge from related domains to improve recommendations for new users or items.
Recommender systems help users deal with large amounts of options by providing personalized recommendations. They are commonly used by websites like Amazon, Netflix, and YouTube. Twenty years of research on recommender systems has led to many different recommendation models, including collaborative filtering, content-based filtering, knowledge-based, and hybrid approaches. Collaborative filtering uses user ratings and preferences to find similar users or items and provide recommendations. It has been widely used by many companies and responsible for a large portion of their sales and views.
Low rank models for recommender systems with limited preference informationEvgeny Frolov
The document summarizes a PhD thesis focused on developing low rank approximation methods for recommender systems with limited preference information. The thesis consists of three parts: an overview of matrix and tensor approximation methods, proposed new methods (a higher order tensor model, SVD-based hybrid model, and combined hybrid tensor model), and software developed. The proposed methods aim to address problems with limited preference data by incorporating side information and more accurately representing user preferences. Evaluation shows the methods improve upon state-of-the-art results for recommender systems.
1. The document discusses various techniques for improving recommender systems, including incorporating trust and identifying expert recommenders.
2. It suggests identifying a subset of trusted users or "experts" who provide high quality recommendations to improve accuracy and coverage.
3. The techniques explored include modeling trust and reputation, identifying influential users in recommendation graphs, and adaptively selecting the best data sources depending on user characteristics.
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
Regression and Classification: An Artificial Neural Network ApproachKhulna University
This presentation introduces artificial neural networks (ANN) as a technique for regression and classification problems. It provides historical context on the development of ANN, describes common network structures and activation functions, and the backpropagation algorithm for training networks. Experimental results on 7 datasets show ANN outperformed other methods for both regression and classification across a variety of problem types and data characteristics. Limitations of ANN and areas for further research are also discussed.
Recommender Systems: Advances in Collaborative FilteringChangsung Moon
This document summarizes recommender systems, focusing on collaborative filtering techniques. It discusses how recommender systems help with information overload by matching users with relevant items. Collaborative filtering is introduced as a technique that seeks to predict user preferences based on other similar users' ratings. The document then covers various collaborative filtering algorithms like neighborhood models, latent factor models using matrix factorization, and extensions like adding biases and temporal dynamics. It concludes by discussing hybrid methods and providing references for further reading.
Computing Neighbourhoods with Language Models in a Collaborative Filtering Sc...Daniel Valcarce
Slides of the presentation given at IIR 2016 for the following extended abstract:
Daniel Valcarce, Javier Parapar, Alvaro Barreiro: Computing Neighbourhoods with Language Models in a Collaborative Filtering Scenario. IIR 2016, Venice, Italy.
http://dx.doi.org/10.1007/978-3-319-30671-1_45
Recommender systems aim to recommend items like books, movies, or products to users based on their preferences. There are two main approaches: collaborative filtering, which recommends items liked by similar users, and content-based filtering, which recommends items similar to those a user has liked based on item attributes. Both have strengths and weaknesses, so hybrid systems combining the approaches can provide the best recommendations.
Random Walk by User Trust and Temporal Issues toward Sparsity Problem in Soci...Sc Huang
Random Walk by User Trust and Temporal Issues toward Sparsity Problem in Social Tagging Recommender Systems
The document discusses using random walks and temporal factors to address sparsity problems in social tagging recommender systems. It introduces related work on item-based collaborative filtering, random walk recommendations, and models that learn influence probabilities. It then describes using random walks starting from users or items, and incorporating trust networks and influence powers to provide recommendations. Finally, it discusses addressing cold start problems, temporal decay issues, and experiment design.
A new similarity measurement based on hellinger distance for collaborating fi...Prabhu Kumar
This project proposed a similarity measurement which is focusing on recommendation performance under the cold start problem [The problem which occurs in the recommendation for newly comer items and users, which doesn't have any recognition in the system] and also perfectly suitable for sparse data set.
This technique solves the problem of the cold start in recommender system as well as improves the performance of recommendation to the users.
Recommender systems are software tools and techniques providing suggestions for items to be of interest to a user. Recommender systems have proved in recent years to be a valuable means of helping Web users by providing useful and effective recommendations or suggestions.
Matrix Factorization In Recommender SystemsYONG ZHENG
The document discusses matrix factorization techniques for recommender systems. It begins with an overview of recommender systems and their use of matrix factorization for dimensionality reduction. Principal component analysis and singular value decomposition are described as early linear algebra techniques used for this purpose. The document then focuses on how these techniques evolved into basic and extended matrix factorization methods in recommender systems, using the Netflix Prize competition as an example.
Deep Reinforcement Learning based Recommendation with Explicit User-ItemInter...Kishor Datta Gupta
—Recommendation is crucial in both academia andindustry, and various techniques are proposed such as content-based collaborative filtering, matrix factorization, logistic re-gression, factorization machines, neural networks and multi-armed bandits. However, most of the previous studies sufferfrom two limitations: (1) considering the recommendation asa static procedure and ignoring the dynamic interactive naturebetween users and the recommender systems; (2) focusing on theimmediate feedback of recommended items and neglecting thelong-term rewards. To address the two limitations, in this paperwe propose a novel recommendation framework based on deepreinforcement learning, called DRR. The DRR framework treatsrecommendation as a sequential decision making procedure andadopts an “Actor-Critic” reinforcement learning scheme to modelthe interactions between the users and recommender systems,which can consider both the dynamic adaptation and long-term rewards. Further more, a state representation module isincorporated into DRR, which can explicitly capture the interac-tions between items and users. Three instantiation structures aredeveloped. Extensive experiments on four real-world datasets areconducted under both the offline and online evaluation settings.The experimental results demonstrate the proposed DRR methodindeed outperforms the state-of-the-art competitors
The document summarizes recommender systems and different recommendation approaches. It introduces content-based recommendation which recommends items similar to those a user preferred. It also covers collaborative filtering which recommends items liked by similar users. Specific collaborative filtering techniques discussed include k-nearest neighbors, association rules, and matrix factorization using singular value decomposition as was done for the Netflix Prize contest.
This document outlines a proposed framework called TKmeans++ for identifying grey sheep users (GSU) and recommending items to them based on trust relations. The framework has two phases: a GSU identification phase that calculates user weights based on similarity, influence, and trust to assign users to clusters, and a recommendation phase that recommends top items to a GSU based on items positively rated by other GSU. An experimental study on the Epinions dataset shows TKmeans++ outperforms other clustering algorithms on MAE and coverage metrics for recommending to GSU. Future work could explore matrix factorization approaches or combining clustering and matrix factorization.
A Randomized Approach for Crowdsourcing in the Presence of Multiple Viewscollwe
The document presents a framework called M2VW for crowdsourcing in the presence of multiple views. M2VW models the multi-view learning problem using crowdsourced labels. It formulates an optimization problem to minimize prediction errors while encouraging worker consensus across views and feature sparsity. A randomized block coordinate descent algorithm is proposed to solve the optimization efficiently in linear time with respect to the number of workers. Experimental results on real and synthetic datasets demonstrate that M2VW achieves better performance than baseline methods and scales linearly with the number of workers.
This document provides an introduction to deep reinforcement learning. It begins with an overview of reinforcement learning and its key characteristics such as using reward signals rather than supervision and sequential decision making. The document then covers the formulation of reinforcement learning problems using Markov decision processes and the typical components of an RL agent including policies, value functions, and models. It discusses popular RL algorithms like Q-learning, deep Q-networks, and policy gradient methods. The document concludes by outlining some potential applications of deep reinforcement learning and recommending further educational resources.
Recommender systems help users deal with overwhelming choices by providing personalized recommendations. They are commonly used by websites like Amazon, Netflix, and YouTube. Research on recommender systems has grown significantly over the past 20 years. Common recommendation models include collaborative filtering, which predicts ratings based on similar users or items, and matrix factorization, which represents users and items as vectors in a latent space. Transfer learning techniques allow knowledge from related domains to improve recommendations for new users or items.
Recommender systems help users deal with large amounts of options by providing personalized recommendations. They are commonly used by websites like Amazon, Netflix, and YouTube. Twenty years of research on recommender systems has led to many different recommendation models, including collaborative filtering, content-based filtering, knowledge-based, and hybrid approaches. Collaborative filtering uses user ratings and preferences to find similar users or items and provide recommendations. It has been widely used by many companies and responsible for a large portion of their sales and views.
Low rank models for recommender systems with limited preference informationEvgeny Frolov
The document summarizes a PhD thesis focused on developing low rank approximation methods for recommender systems with limited preference information. The thesis consists of three parts: an overview of matrix and tensor approximation methods, proposed new methods (a higher order tensor model, SVD-based hybrid model, and combined hybrid tensor model), and software developed. The proposed methods aim to address problems with limited preference data by incorporating side information and more accurately representing user preferences. Evaluation shows the methods improve upon state-of-the-art results for recommender systems.
1. The document discusses various techniques for improving recommender systems, including incorporating trust and identifying expert recommenders.
2. It suggests identifying a subset of trusted users or "experts" who provide high quality recommendations to improve accuracy and coverage.
3. The techniques explored include modeling trust and reputation, identifying influential users in recommendation graphs, and adaptively selecting the best data sources depending on user characteristics.
Similar to Bando de Dados Avançados - Recommender Systems (20)
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
ESR spectroscopy in liquid food and beverages.pptxPRIYANKA PATEL
With increasing population, people need to rely on packaged food stuffs. Packaging of food materials requires the preservation of food. There are various methods for the treatment of food to preserve them and irradiation treatment of food is one of them. It is the most common and the most harmless method for the food preservation as it does not alter the necessary micronutrients of food materials. Although irradiated food doesn’t cause any harm to the human health but still the quality assessment of food is required to provide consumers with necessary information about the food. ESR spectroscopy is the most sophisticated way to investigate the quality of the food and the free radicals induced during the processing of the food. ESR spin trapping technique is useful for the detection of highly unstable radicals in the food. The antioxidant capability of liquid food and beverages in mainly performed by spin trapping technique.
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...University of Maribor
Slides from talk:
Aleš Zamuda: Remote Sensing and Computational, Evolutionary, Supercomputing, and Intelligent Systems.
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Inter-Society Networking Panel GRSS/MTT-S/CIS Panel Session: Promoting Connection and Cooperation
https://www.etran.rs/2024/en/home-english/
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...Travis Hills MN
Travis Hills of Minnesota developed a method to convert waste into high-value dry fertilizer, significantly enriching soil quality. By providing farmers with a valuable resource derived from waste, Travis Hills helps enhance farm profitability while promoting environmental stewardship. Travis Hills' sustainable practices lead to cost savings and increased revenue for farmers by improving resource efficiency and reducing waste.
The debris of the ‘last major merger’ is dynamically youngSérgio Sacani
The Milky Way’s (MW) inner stellar halo contains an [Fe/H]-rich component with highly eccentric orbits, often referred to as the
‘last major merger.’ Hypotheses for the origin of this component include Gaia-Sausage/Enceladus (GSE), where the progenitor
collided with the MW proto-disc 8–11 Gyr ago, and the Virgo Radial Merger (VRM), where the progenitor collided with the
MW disc within the last 3 Gyr. These two scenarios make different predictions about observable structure in local phase space,
because the morphology of debris depends on how long it has had to phase mix. The recently identified phase-space folds in Gaia
DR3 have positive caustic velocities, making them fundamentally different than the phase-mixed chevrons found in simulations
at late times. Roughly 20 per cent of the stars in the prograde local stellar halo are associated with the observed caustics. Based
on a simple phase-mixing model, the observed number of caustics are consistent with a merger that occurred 1–2 Gyr ago.
We also compare the observed phase-space distribution to FIRE-2 Latte simulations of GSE-like mergers, using a quantitative
measurement of phase mixing (2D causticality). The observed local phase-space distribution best matches the simulated data
1–2 Gyr after collision, and certainly not later than 3 Gyr. This is further evidence that the progenitor of the ‘last major merger’
did not collide with the MW proto-disc at early times, as is thought for the GSE, but instead collided with the MW disc within
the last few Gyr, consistent with the body of work surrounding the VRM.
Or: Beyond linear.
Abstract: Equivariant neural networks are neural networks that incorporate symmetries. The nonlinear activation functions in these networks result in interesting nonlinear equivariant maps between simple representations, and motivate the key player of this talk: piecewise linear representation theory.
Disclaimer: No one is perfect, so please mind that there might be mistakes and typos.
dtubbenhauer@gmail.com
Corrected slides: dtubbenhauer.com/talks.html
ESPP presentation to EU Waste Water Network, 4th June 2024 “EU policies driving nutrient removal and recycling
and the revised UWWTD (Urban Waste Water Treatment Directive)”
1. Recommender Systems
Collaborative Filtering &
Dimensionality Reduction
Mining of Massive Datasets
Jure Leskovec,Anand Rajaraman, Jeff Ullman
Stanford University
*Adapted by Gustavo Coutinho
Note to other teachers and users of these slides: We would be delighted if you found this our
material useful in giving your own lectures. Feel free to use these slides verbatim, or to modify
them to fit your own needs. If you make use of a significant portion of these slides in your own
lecture, please include this message, or a link to our web site: http://www.mmds.org
2. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Collaborative Filtering
Harnessing quality judgments of other
users
2
3. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Previously - Content-Based
3J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
4. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Previously - Content-Based
4J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
5. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Previously - Content-Based
5J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
6. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Utility Matrix
Users have preferences for certain items,
and these preferences must be teased out
of the data.
Lets represent it with an Utility Matrix!
Example:
6
7. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Collaborative Filtering
Consider user x
Find set N of other
users whose ratings
are “similar” to
x’s ratings
Estimate x’s ratings
based on ratings
of users in N
7
x
N
J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
8. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Collaborative Filtering
Different from Content-Based Filtering
We don’t need to understand the
content of an specific item!
Different user share their experiences
8J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
9. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Let rx and rx vectors of users x and y
ratings, respectively
Lets try to use the Jaccard Similarity as
a measure
9
Finding “Similar” Users
rx = [*, _, _, *, ***]
ry = [*, _, **, **, _]
10. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Now, rx and ry are considered as sets
Problem: Ignores the value of the rating!
10
Finding “Similar” Users
rx = { 1, 4, 5}
ry = { 1, 3, 4}
11. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
How to put the rating factor under a
formula?
Cosine Similarity measure
Now, rx and ry are considered as points
Problem: Treats missing ratings as
“negative”!
11
Finding “Similar” Users
similarity = cos(Θ) =
rx · ry
||rx|| · ||ry||
rx = { 1, 0, 0, 1, 3}
ry = { 1, 0, 2, 2, 0}
12. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
How do we balance the missing values?
Pearson correlation coefficient
Sxy= items rated by both users x and y
12
Finding “Similar” Users
sim(x, y) =
s∈Sxy
(rxs − rx)(rys − ry)
s∈Sxy
(rxs − rx)2
s∈Sxy
(rys − ry)2
rx and ry = average rating of “x” and “y”
13. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Similarity Metric
Lets consider de following Utility Matrix
of users and ratings
Intuitively we want: sim(A,B)>sim(A,C)
Using Jaccard: 1/5 < 2/4
Using Cosine: 0.386 > 0.322
13
14. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Similarity Metric
Now, we’re going to use Pearson
Correlation
Subtracting the (row) mean
Using Pearson: 0.092 > -0.559
Notice that Cosine Similarity is a
correlation when data is centered at 0
14
15. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Rating Predictions
How can we go from similarity metrics to
recommendations?
Let rx be the vector of user x’s ratings
Let N be the set of k users most similar to
x who have rated item i
Prediction for item s of user x:
Where sxy=sim(x,y)
15
rxi =
y∈N sxy · ryi
y∈N sxy
rxi =
1
k
·
y∈N
ryi
16. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Item-Item Collaborative Filtering
Until now we have used an User-User
approach.
What about an Item-Item?
▪ For item i, find other similar items
▪ Estimate rating for item i based on
ratings for similar items
▪ Can use the same similarity metrics and
predictions functions as in user-user
model
16
18. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Item-Item CF (|N|=2)
12 11 10 9 8 7 6 5 4 3 2 1
4 5 5 ? 3 1 1
3 1 2 4 4 5 2
5 3 4 3 2 1 4 2 3
2 4 5 4 2 4
5 2 2 4 3 4 5
4 2 3 3 1 6
Users
Movies
- estimate rating of movie 1 by user 5
18
19. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org 19
Item-Item CF (|N|=2)
1.00
-0.18
0.41
-0.10
-0.31
0.59
sim(1,m)12 11 10 9 8 7 6 5 4 3 2 1
4 5 5 ? 3 1 1
3 1 2 4 4 5 2
5 3 4 3 2 1 4 2 3
2 4 5 4 2 4
5 2 2 4 3 4 5
4 2 3 3 1 6
Users
Movies
- estimate rating of movie 1 by user 5
20. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Item-Item CF (|N|=2)
Neighbour selection: identify movies
similar to movie 1, rated by user 5
Here we use Pearson correlation as
similarity:
Subtract mean rating mi from each
movie i
m1=(1+3+5+5+4)/5 = 3.6
row1:[ -2.6, 0, -0.6, 0, 0, 1.4, 0, 0,
1.4, 0, 0.4, 0]
Compute cosine similarities between
rows 20
23. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Define similarity sij of items i and j
Select k nearest neighbors N(i; x)
▪ Items most similar to i, that were rated by x
Estimate rating rxi as the weighted
average:
CF: Common Practice
23
baseline estimate for rxi
µ = overall mean movie rating
bx = rating deviation of user x
= (avg. rating of user x) – µ
bi = rating deviation of movie i
∑
∑
∈
∈
−⋅
+=
);(
);(
)(
xiNj ij
xiNj xjxjij
xixi
s
brs
br
24. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Item-Item vs. User-User
In practice, it has been observed that item-item
often works better than user-user
Why? Items are simpler, users have multiple tastes
Avatar LOTR Matrix Pirates
Alice 1 0.8
Bob 0.5 0.3
Carol 0.9 1 0.8
David 1 0.4
24
25. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Works for any kind of item
No feature selection needed
Unexpected recommendations
A user may receive recommendations
different from active searches done by itself
Groups with similar ratings
Users may connect with each other and
create groups with similar interests
Pros/Cons of Collaborative Filtering
25
26. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Cold Start
Need enough users in the system to find a match
Sparsity
The user/ratings matrix is sparse
Hard to find users that have rated the same items
First rater
Cannot recommend an item that has not been
previously rated
New items, Esoteric items
Popularity bias
Cannot recommend items to someone with
unique taste
Tends to recommend popular items
Pros/Cons of Collaborative Filtering
26
27. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Hybrid Methods
Implement two or more different
recommenders and combine predictions
Perhaps using a linear model
Add content-based methods to
collaborative filtering
Item profiles for new item problem
Demographics to deal with new user
problem
27
28. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Remarks & Practical Tips
- Evaluation
- Error metrics
- Complexity / Speed
2828
29. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Evaluation
1 3 4
3 5 5
4 5 5
3
3
2 2 2
5
2 1 1
3 3
1
Users
Movies
29
30. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Evaluation
1 3 4
3 5 5
4 5 5
3
3
2 ? ?
?
2 1 ?
3 ?
1
Users
Movies
Test Data Set
30
31. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Collaborative Filtering: Complexity
Expensive step is finding k most similar
customers: O(|X|)
Too expensive to do at runtime
Could pre-compute
Naïve pre-computation takes time O(k·|X|)
We already know how to do this!
Near-neighbor search in high
dimensions (LSH)
Clustering
Dimensionality reduction
32
32. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Tip:Add Data
Leverage all the data
Don’t try to reduce data size in an
effort to make fancy algorithms work
Simple methods on large data do best
Add more data
e.g., add IMDB data on genres
More data beats better algorithms
http://anand.typepad.com/datawocky/2008/03/more-data-
usual.html
33
33. J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org
Questions
34