R exam (B) given in Paris-Dauphine, Licence Mido, Jan. 11, 2013Christian Robert
This is one of two exams given to our students this year. They had two hours to solve three problems and had to return R codes as well as handwritten explanations.
In this experiment, I tried to implement Minimum
error rate classifier using the posterior probabilities which
uses Normal distribution to calculate likelihood probabilities to
classify given sample points
R exam (B) given in Paris-Dauphine, Licence Mido, Jan. 11, 2013Christian Robert
This is one of two exams given to our students this year. They had two hours to solve three problems and had to return R codes as well as handwritten explanations.
In this experiment, I tried to implement Minimum
error rate classifier using the posterior probabilities which
uses Normal distribution to calculate likelihood probabilities to
classify given sample points
Linear Discriminant Analysis and Its Generalization일상 온
The brief introduction to the linear discriminant analysis and some extended methods. Much of the materials are taken from The Elements of Statistical Learning by Hastie et al. (2008).
Implicit schemes are needed in order to have fast runtime in wave models. Parallelization using the Message Passing Interface are needed in order to run on computers with thousands of processors. Implicit schemes rely on preconditioner in order for the iterative schemes to converge fast. Thus we need fast preconditioners and we present those here.
Introduction to machine learning terminology.
Applications within High Energy Physics and outside HEP.
* Basic problems: classification and regression.
* Nearest neighbours approach and spacial indices
* Overfitting (intro)
* Curse of dimensionality
* ROC curve, ROC AUC
* Bayes optimal classifier
* Density estimation: KDE and histograms
* Parametric density estimation
* Mixtures for density estimation and EM algorithm
* Generative approach vs discriminative approach
* Linear decision rule, intro to logistic regression
* Linear regression
Robot의 Gait optimization, Gesture Recognition, Optimal Control, Hyper parameter optimization, 신약 신소재 개발을 위한 optimal data sampling strategy등과 같은 ML분야에서 약방의 감초 같은 존재인 GP이지만 이해가 쉽지 않은 GP의 기본적인 이론 및 matlab code 소개
Error Estimates for Multi-Penalty Regularization under General Source Conditioncsandit
In learning theory, the convergence issues of the regression problem are investigated with
the least square Tikhonov regularization schemes in both the RKHS-norm and the L 2
-norm.
We consider the multi-penalized least square regularization scheme under the general source
condition with the polynomial decay of the eigenvalues of the integral operator. One of the
motivation for this work is to discuss the convergence issues for widely considered manifold
regularization scheme. The optimal convergence rates of multi-penalty regularizer is achieved
in the interpolation norm using the concept of effective dimension. Further we also propose
the penalty balancing principle based on augmented Tikhonov regularization for the choice of
regularization parameters. The superiority of multi-penalty regularization over single-penalty
regularization is shown using the academic example and moon data set.
A review of one of the most popular methods of clustering, a part of what is know as unsupervised learning, K-Means. Here, we go from the basic heuristic used to solve the NP-Hard problem to an approximation algorithm K-Centers. Additionally, we look at variations coming from the Fuzzy Set ideas. In the future, we will add more about On-Line algorithms in the line of Stochastic Gradient Ideas...
Linear Discriminant Analysis and Its Generalization일상 온
The brief introduction to the linear discriminant analysis and some extended methods. Much of the materials are taken from The Elements of Statistical Learning by Hastie et al. (2008).
Implicit schemes are needed in order to have fast runtime in wave models. Parallelization using the Message Passing Interface are needed in order to run on computers with thousands of processors. Implicit schemes rely on preconditioner in order for the iterative schemes to converge fast. Thus we need fast preconditioners and we present those here.
Introduction to machine learning terminology.
Applications within High Energy Physics and outside HEP.
* Basic problems: classification and regression.
* Nearest neighbours approach and spacial indices
* Overfitting (intro)
* Curse of dimensionality
* ROC curve, ROC AUC
* Bayes optimal classifier
* Density estimation: KDE and histograms
* Parametric density estimation
* Mixtures for density estimation and EM algorithm
* Generative approach vs discriminative approach
* Linear decision rule, intro to logistic regression
* Linear regression
Robot의 Gait optimization, Gesture Recognition, Optimal Control, Hyper parameter optimization, 신약 신소재 개발을 위한 optimal data sampling strategy등과 같은 ML분야에서 약방의 감초 같은 존재인 GP이지만 이해가 쉽지 않은 GP의 기본적인 이론 및 matlab code 소개
Error Estimates for Multi-Penalty Regularization under General Source Conditioncsandit
In learning theory, the convergence issues of the regression problem are investigated with
the least square Tikhonov regularization schemes in both the RKHS-norm and the L 2
-norm.
We consider the multi-penalized least square regularization scheme under the general source
condition with the polynomial decay of the eigenvalues of the integral operator. One of the
motivation for this work is to discuss the convergence issues for widely considered manifold
regularization scheme. The optimal convergence rates of multi-penalty regularizer is achieved
in the interpolation norm using the concept of effective dimension. Further we also propose
the penalty balancing principle based on augmented Tikhonov regularization for the choice of
regularization parameters. The superiority of multi-penalty regularization over single-penalty
regularization is shown using the academic example and moon data set.
A review of one of the most popular methods of clustering, a part of what is know as unsupervised learning, K-Means. Here, we go from the basic heuristic used to solve the NP-Hard problem to an approximation algorithm K-Centers. Additionally, we look at variations coming from the Fuzzy Set ideas. In the future, we will add more about On-Line algorithms in the line of Stochastic Gradient Ideas...
Data Steganography for Optical Color Image CryptosystemsCSCJournals
In this paper, an optical color image cryptosystem with a data hiding scheme is proposed. In the proposed optical cryptosystem, a confidential color image is embedded into the host image of the same size. Then the stego-image is encrypted by using the double random phase encoding algorithm. The seeds to generate random phase data are hidden in the encrypted stego-image by a content-dependent and low distortion data embedding technique. The confidential image and secret data delivery is accomplished by hiding the image into the host image and embedding the data into the encrypted stego-image. Experimental results show that the proposed data steganographic cryptosystem provides large data hiding capacity and high reconstructed image quality.
Fiduciary Wellness 06/11/13 - 5 things you should be, and 5 things you should...Michael Olah
ERISA fiduciaries often don't know what is expected of them - and some don't even know what it means to be a fiduciary, and even if they are one. This presentation walks through some basics of being an ERISA fiduciary, and then describes 5 things you should be doing as a fiduciary (but probably aren't) and 5 things you really shouldn't be doing (and hopefully aren't!).
Student and lecturers’ experiences of introducing a hybrid IBL approach to teaching Organisation Studies in a business school
Authors: M Page, H Gaggiotti, C Jarvis,, with E Attwell, M Lukaj, L McCann, S Hayward, L Hindson
Using Enquiry Based Learning to Create a Blended Academic Skills Development ...cilass.slideshare
For a number of years Academic Skills modules had been delivered to campus-based students in a blended mode. However the designs had not been able to fully engage students in a module that was seen as of little or no relevance to their academic or future careers. Inquiry based learning was used as the basis for a redesign of one such module allowing for the inclusion of authentic and group-based activities. The poster will outline the design, delivery and evaluation of a module and how undergraduate students have been brought to an awareness of the importance of independent learning skills and their value in HE and beyond.
Slides were formed by referring to the text Machine Learning by Tom M Mitchelle (Mc Graw Hill, Indian Edition) and by referring to Video tutorials on NPTEL
Slides for the presentation at ENBIS 2018 of "Deep k-Means: Jointly Clustering with k-Means and Learning Representations" by Thibaut Thonet. Joint work with Maziar Moradi Fard and Eric Gaussier.
Lecture 02: Machine Learning for Language Technology - Decision Trees and Nea...Marina Santini
In this lecture, we talk about two different discriminative machine learning methods: decision trees and k-nearest neighbors. Decision trees are hierarchical structures.k-nearest neighbors are based on two principles: recollection and resemblance.
Machine learning in science and industry — day 1arogozhnikov
A course of machine learning in science and industry.
- notions and applications
- nearest neighbours: search and machine learning algorithms
- roc curve
- optimal classification and regression
- density estimation
- Gaussian mixtures and EM algorithm
- clustering, an example of clustering in the opera
Mathematics (from Greek μάθημα máthēma, “knowledge, study, learning”) is the study of topics such as quantity (numbers), structure, space, and change. There is a range of views among mathematicians and philosophers as to the exact scope and definition of mathematics
* ML in HEP
* classification and regression
* knn classification and regression
* ROC curve
* optimal bayesian classifier
* Fisher's QDA
* intro to Logistic Regression
Information-theoretic clustering with applicationsFrank Nielsen
Information-theoretic clustering with applications
Abstract: Clustering is a fundamental and key primitive to discover structural groups of homogeneous data in data sets, called clusters. The most famous clustering technique is the celebrated k-means clustering that seeks to minimize the sum of intra-cluster variances. k-Means is NP-hard as soon as the dimension and the number of clusters are both greater than 1. In the first part of the talk, we first present a generic dynamic programming method to compute the optimal clustering of n scalar elements into k pairwise disjoint intervals. This case includes 1D Euclidean k-means but also other kinds of clustering algorithms like the k-medoids, the k-medians, the k-centers, etc.
We extend the method to incorporate cluster size constraints and show how to choose the appropriate number of clusters using model selection. We then illustrate and refine the method on two case studies: 1D Bregman clustering and univariate statistical mixture learning maximizing the complete likelihood. In the second part of the talk, we introduce a generalization of k-means to cluster sets of histograms that has become an important ingredient of modern information processing due to the success of the bag-of-word modelling paradigm.
Clustering histograms can be performed using the celebrated k-means centroid-based algorithm. We consider the Jeffreys divergence that symmetrizes the Kullback-Leibler divergence, and investigate the computation of Jeffreys centroids. We prove that the Jeffreys centroid can be expressed analytically using the Lambert W function for positive histograms. We then show how to obtain a fast guaranteed approximation when dealing with frequency histograms and conclude with some remarks on the k-means histogram clustering.
References: - Optimal interval clustering: Application to Bregman clustering and statistical mixture learning IEEE ISIT 2014 (recent result poster) http://arxiv.org/abs/1403.2485
- Jeffreys Centroids: A Closed-Form Expression for Positive Histograms and a Guaranteed Tight Approximation for Frequency Histograms.
IEEE Signal Process. Lett. 20(7): 657-660 (2013) http://arxiv.org/abs/1303.7286
http://www.i.kyoto-u.ac.jp/informatics-seminar/
RMD24 | Retail media: hoe zet je dit in als je geen AH of Unilever bent? Heid...BBPMedia1
Grote partijen zijn al een tijdje onderweg met retail media. Ondertussen worden in dit domein ook de kansen zichtbaar voor andere spelers in de markt. Maar met die kansen ontstaan ook vragen: Zelf retail media worden of erop adverteren? In welke fase van de funnel past het en hoe integreer je het in een mediaplan? Wat is nu precies het verschil met marketplaces en Programmatic ads? In dit half uur beslechten we de dilemma's en krijg je antwoorden op wanneer het voor jou tijd is om de volgende stap te zetten.
Attending a job Interview for B1 and B2 Englsih learnersErika906060
It is a sample of an interview for a business english class for pre-intermediate and intermediate english students with emphasis on the speking ability.
Skye Residences | Extended Stay Residences Near Toronto Airportmarketingjdass
Experience unparalleled EXTENDED STAY and comfort at Skye Residences located just minutes from Toronto Airport. Discover sophisticated accommodations tailored for discerning travelers.
Website Link :
https://skyeresidences.com/
https://skyeresidences.com/about-us/
https://skyeresidences.com/gallery/
https://skyeresidences.com/rooms/
https://skyeresidences.com/near-by-attractions/
https://skyeresidences.com/commute/
https://skyeresidences.com/contact/
https://skyeresidences.com/queen-suite-with-sofa-bed/
https://skyeresidences.com/queen-suite-with-sofa-bed-and-balcony/
https://skyeresidences.com/queen-suite-with-sofa-bed-accessible/
https://skyeresidences.com/2-bedroom-deluxe-queen-suite-with-sofa-bed/
https://skyeresidences.com/2-bedroom-deluxe-king-queen-suite-with-sofa-bed/
https://skyeresidences.com/2-bedroom-deluxe-queen-suite-with-sofa-bed-accessible/
#Skye Residences Etobicoke, #Skye Residences Near Toronto Airport, #Skye Residences Toronto, #Skye Hotel Toronto, #Skye Hotel Near Toronto Airport, #Hotel Near Toronto Airport, #Near Toronto Airport Accommodation, #Suites Near Toronto Airport, #Etobicoke Suites Near Airport, #Hotel Near Toronto Pearson International Airport, #Toronto Airport Suite Rentals, #Pearson Airport Hotel Suites
"𝑩𝑬𝑮𝑼𝑵 𝑾𝑰𝑻𝑯 𝑻𝑱 𝑰𝑺 𝑯𝑨𝑳𝑭 𝑫𝑶𝑵𝑬"
𝐓𝐉 𝐂𝐨𝐦𝐬 (𝐓𝐉 𝐂𝐨𝐦𝐦𝐮𝐧𝐢𝐜𝐚𝐭𝐢𝐨𝐧𝐬) is a professional event agency that includes experts in the event-organizing market in Vietnam, Korea, and ASEAN countries. We provide unlimited types of events from Music concerts, Fan meetings, and Culture festivals to Corporate events, Internal company events, Golf tournaments, MICE events, and Exhibitions.
𝐓𝐉 𝐂𝐨𝐦𝐬 provides unlimited package services including such as Event organizing, Event planning, Event production, Manpower, PR marketing, Design 2D/3D, VIP protocols, Interpreter agency, etc.
Sports events - Golf competitions/billiards competitions/company sports events: dynamic and challenging
⭐ 𝐅𝐞𝐚𝐭𝐮𝐫𝐞𝐝 𝐩𝐫𝐨𝐣𝐞𝐜𝐭𝐬:
➢ 2024 BAEKHYUN [Lonsdaleite] IN HO CHI MINH
➢ SUPER JUNIOR-L.S.S. THE SHOW : Th3ee Guys in HO CHI MINH
➢FreenBecky 1st Fan Meeting in Vietnam
➢CHILDREN ART EXHIBITION 2024: BEYOND BARRIERS
➢ WOW K-Music Festival 2023
➢ Winner [CROSS] Tour in HCM
➢ Super Show 9 in HCM with Super Junior
➢ HCMC - Gyeongsangbuk-do Culture and Tourism Festival
➢ Korean Vietnam Partnership - Fair with LG
➢ Korean President visits Samsung Electronics R&D Center
➢ Vietnam Food Expo with Lotte Wellfood
"𝐄𝐯𝐞𝐫𝐲 𝐞𝐯𝐞𝐧𝐭 𝐢𝐬 𝐚 𝐬𝐭𝐨𝐫𝐲, 𝐚 𝐬𝐩𝐞𝐜𝐢𝐚𝐥 𝐣𝐨𝐮𝐫𝐧𝐞𝐲. 𝐖𝐞 𝐚𝐥𝐰𝐚𝐲𝐬 𝐛𝐞𝐥𝐢𝐞𝐯𝐞 𝐭𝐡𝐚𝐭 𝐬𝐡𝐨𝐫𝐭𝐥𝐲 𝐲𝐨𝐮 𝐰𝐢𝐥𝐥 𝐛𝐞 𝐚 𝐩𝐚𝐫𝐭 𝐨𝐟 𝐨𝐮𝐫 𝐬𝐭𝐨𝐫𝐢𝐞𝐬."
Falcon stands out as a top-tier P2P Invoice Discounting platform in India, bridging esteemed blue-chip companies and eager investors. Our goal is to transform the investment landscape in India by establishing a comprehensive destination for borrowers and investors with diverse profiles and needs, all while minimizing risk. What sets Falcon apart is the elimination of intermediaries such as commercial banks and depository institutions, allowing investors to enjoy higher yields.
What is the TDS Return Filing Due Date for FY 2024-25.pdfseoforlegalpillers
It is crucial for the taxpayers to understand about the TDS Return Filing Due Date, so that they can fulfill your TDS obligations efficiently. Taxpayers can avoid penalties by sticking to the deadlines and by accurate filing of TDS. Timely filing of TDS will make sure about the availability of tax credits. You can also seek the professional guidance of experts like Legal Pillers for timely filing of the TDS Return.
As a business owner in Delaware, staying on top of your tax obligations is paramount, especially with the annual deadline for Delaware Franchise Tax looming on March 1. One such obligation is the annual Delaware Franchise Tax, which serves as a crucial requirement for maintaining your company’s legal standing within the state. While the prospect of handling tax matters may seem daunting, rest assured that the process can be straightforward with the right guidance. In this comprehensive guide, we’ll walk you through the steps of filing your Delaware Franchise Tax and provide insights to help you navigate the process effectively.
[Note: This is a partial preview. To download this presentation, visit:
https://www.oeconsulting.com.sg/training-presentations]
Sustainability has become an increasingly critical topic as the world recognizes the need to protect our planet and its resources for future generations. Sustainability means meeting our current needs without compromising the ability of future generations to meet theirs. It involves long-term planning and consideration of the consequences of our actions. The goal is to create strategies that ensure the long-term viability of People, Planet, and Profit.
Leading companies such as Nike, Toyota, and Siemens are prioritizing sustainable innovation in their business models, setting an example for others to follow. In this Sustainability training presentation, you will learn key concepts, principles, and practices of sustainability applicable across industries. This training aims to create awareness and educate employees, senior executives, consultants, and other key stakeholders, including investors, policymakers, and supply chain partners, on the importance and implementation of sustainability.
LEARNING OBJECTIVES
1. Develop a comprehensive understanding of the fundamental principles and concepts that form the foundation of sustainability within corporate environments.
2. Explore the sustainability implementation model, focusing on effective measures and reporting strategies to track and communicate sustainability efforts.
3. Identify and define best practices and critical success factors essential for achieving sustainability goals within organizations.
CONTENTS
1. Introduction and Key Concepts of Sustainability
2. Principles and Practices of Sustainability
3. Measures and Reporting in Sustainability
4. Sustainability Implementation & Best Practices
To download the complete presentation, visit: https://www.oeconsulting.com.sg/training-presentations
Premium MEAN Stack Development Solutions for Modern BusinessesSynapseIndia
Stay ahead of the curve with our premium MEAN Stack Development Solutions. Our expert developers utilize MongoDB, Express.js, AngularJS, and Node.js to create modern and responsive web applications. Trust us for cutting-edge solutions that drive your business growth and success.
Know more: https://www.synapseindia.com/technology/mean-stack-development-company.html
Cracking the Workplace Discipline Code Main.pptxWorkforce Group
Cultivating and maintaining discipline within teams is a critical differentiator for successful organisations.
Forward-thinking leaders and business managers understand the impact that discipline has on organisational success. A disciplined workforce operates with clarity, focus, and a shared understanding of expectations, ultimately driving better results, optimising productivity, and facilitating seamless collaboration.
Although discipline is not a one-size-fits-all approach, it can help create a work environment that encourages personal growth and accountability rather than solely relying on punitive measures.
In this deck, you will learn the significance of workplace discipline for organisational success. You’ll also learn
• Four (4) workplace discipline methods you should consider
• The best and most practical approach to implementing workplace discipline.
• Three (3) key tips to maintain a disciplined workplace.
3.0 Project 2_ Developing My Brand Identity Kit.pptxtanyjahb
A personal brand exploration presentation summarizes an individual's unique qualities and goals, covering strengths, values, passions, and target audience. It helps individuals understand what makes them stand out, their desired image, and how they aim to achieve it.
What are the main advantages of using HR recruiter services.pdfHumanResourceDimensi1
HR recruiter services offer top talents to companies according to their specific needs. They handle all recruitment tasks from job posting to onboarding and help companies concentrate on their business growth. With their expertise and years of experience, they streamline the hiring process and save time and resources for the company.
Putting the SPARK into Virtual Training.pptxCynthia Clay
This 60-minute webinar, sponsored by Adobe, was delivered for the Training Mag Network. It explored the five elements of SPARK: Storytelling, Purpose, Action, Relationships, and Kudos. Knowing how to tell a well-structured story is key to building long-term memory. Stating a clear purpose that doesn't take away from the discovery learning process is critical. Ensuring that people move from theory to practical application is imperative. Creating strong social learning is the key to commitment and engagement. Validating and affirming participants' comments is the way to create a positive learning environment.
Enterprise Excellence is Inclusive Excellence.pdfKaiNexus
Enterprise excellence and inclusive excellence are closely linked, and real-world challenges have shown that both are essential to the success of any organization. To achieve enterprise excellence, organizations must focus on improving their operations and processes while creating an inclusive environment that engages everyone. In this interactive session, the facilitator will highlight commonly established business practices and how they limit our ability to engage everyone every day. More importantly, though, participants will likely gain increased awareness of what we can do differently to maximize enterprise excellence through deliberate inclusion.
What is Enterprise Excellence?
Enterprise Excellence is a holistic approach that's aimed at achieving world-class performance across all aspects of the organization.
What might I learn?
A way to engage all in creating Inclusive Excellence. Lessons from the US military and their parallels to the story of Harry Potter. How belt systems and CI teams can destroy inclusive practices. How leadership language invites people to the party. There are three things leaders can do to engage everyone every day: maximizing psychological safety to create environments where folks learn, contribute, and challenge the status quo.
Who might benefit? Anyone and everyone leading folks from the shop floor to top floor.
Dr. William Harvey is a seasoned Operations Leader with extensive experience in chemical processing, manufacturing, and operations management. At Michelman, he currently oversees multiple sites, leading teams in strategic planning and coaching/practicing continuous improvement. William is set to start his eighth year of teaching at the University of Cincinnati where he teaches marketing, finance, and management. William holds various certifications in change management, quality, leadership, operational excellence, team building, and DiSC, among others.
Memorandum Of Association Constitution of Company.pptseri bangash
www.seribangash.com
A Memorandum of Association (MOA) is a legal document that outlines the fundamental principles and objectives upon which a company operates. It serves as the company's charter or constitution and defines the scope of its activities. Here's a detailed note on the MOA:
Contents of Memorandum of Association:
Name Clause: This clause states the name of the company, which should end with words like "Limited" or "Ltd." for a public limited company and "Private Limited" or "Pvt. Ltd." for a private limited company.
https://seribangash.com/article-of-association-is-legal-doc-of-company/
Registered Office Clause: It specifies the location where the company's registered office is situated. This office is where all official communications and notices are sent.
Objective Clause: This clause delineates the main objectives for which the company is formed. It's important to define these objectives clearly, as the company cannot undertake activities beyond those mentioned in this clause.
www.seribangash.com
Liability Clause: It outlines the extent of liability of the company's members. In the case of companies limited by shares, the liability of members is limited to the amount unpaid on their shares. For companies limited by guarantee, members' liability is limited to the amount they undertake to contribute if the company is wound up.
https://seribangash.com/promotors-is-person-conceived-formation-company/
Capital Clause: This clause specifies the authorized capital of the company, i.e., the maximum amount of share capital the company is authorized to issue. It also mentions the division of this capital into shares and their respective nominal value.
Association Clause: It simply states that the subscribers wish to form a company and agree to become members of it, in accordance with the terms of the MOA.
Importance of Memorandum of Association:
Legal Requirement: The MOA is a legal requirement for the formation of a company. It must be filed with the Registrar of Companies during the incorporation process.
Constitutional Document: It serves as the company's constitutional document, defining its scope, powers, and limitations.
Protection of Members: It protects the interests of the company's members by clearly defining the objectives and limiting their liability.
External Communication: It provides clarity to external parties, such as investors, creditors, and regulatory authorities, regarding the company's objectives and powers.
https://seribangash.com/difference-public-and-private-company-law/
Binding Authority: The company and its members are bound by the provisions of the MOA. Any action taken beyond its scope may be considered ultra vires (beyond the powers) of the company and therefore void.
Amendment of MOA:
While the MOA lays down the company's fundamental principles, it is not entirely immutable. It can be amended, but only under specific circumstances and in compliance with legal procedures. Amendments typically require shareholder
Accpac to QuickBooks Conversion Navigating the Transition with Online Account...PaulBryant58
This article provides a comprehensive guide on how to
effectively manage the convert Accpac to QuickBooks , with a particular focus on utilizing online accounting services to streamline the process.
8. Distancias
Distancia de Levenshtein, distancia de edición, o distancia entre palabras,
al número mínimo de operaciones requeridas para transformar una cadena de
caracteres en otra. Se entiende por operación: inserción, eliminación o la
sustitución de un carácter.
https://en.wikibooks.org/wiki/Algorithm_I
mplementation/Strings/Levenshtein_dista
nce#Python
13. ¿IBL?
La idea es simple:
La clase de una instancia debe ser similar a la clase asociada e ejemplos
parecidos.
Almacenar todo los ejemplos.
Cuando se recibe una instancia para clasificar se buscan los ejemplos “más
parecidos” y se analizan las clases asignadas.
Pero:
La clasificación puede ser costosa
¿Todos los atributos son igual de relevantes?
¿Cuántos son los ejemplos parecidos?
¿Si los ejemplos parecidos tienen clases disímiles?
¿Todos lso ejemplos parecidos “pesan” igual?
¿Qué tan parecidos deben ser los parecidos?
14. K-nearest neighbor
To define how similar two examples are we need a metric.
We assume all examples are points in an n-dimensional space Rn and use the
Euclidean distance:
Let Xi and Xj be two examples. Their distance d(Xi,Xj) is defined as:
d(Xi, Xj) = ( Σk [xik – xjk]2 ) ** 1/2
Where xik is the value of attribute k on example Xi.
16. Nearest Neighbor
Four things make a memory based learner:
1. A distance metric
Euclidian
2. How many nearby neighbors to look at?
One
3. A weighting function (optional)
Unused
4. How to fit with the local points?
Just predict the same output as the nearest neighbor.
17. Voronoi Diagram
Decision surface induced by a 1-nearest neighbor. The decision
surface is a combination of convex polyhedra surrounding each
training example.
24. k-Nearest Neighbour Classification Method
Key idea: keep all the training instances
Given query example, take vote amongst its k neighbours
Neighbours are determined by using a distance function
25. k-Nearest Neighbour Classification Method
(k=1)
(k=4)
Probability interpretation: estimate p(y|x) as
( ){ }, | , ( )
( | ) , ( ) is the neighborhood around
| ( ) |
i i i ix y y y x N x
p y x N x x
N x
= ∈
=
Sample adapted from Rong Jin’s slides
26. k-Nearest Neighbour Classification Method
Advantages:
Training is really fast
Can learn complex target functions
Disadvantages
Slow at query time: Efficient data structures are needed to speed
up the query
27. How to choose k?
Use validation with leave-one-out method
For k = 1, 2, …, K
Err(k) = 0;
1. Randomly select a training data point and
hide its class label
2. Using the remaining data and given k to
predict the class label for the left data point
3. Err(k) = Err(k) + 1 if the predicted label is
different from the true label
Repeat the procedure until all training examples
are tested
Choose the k whose Err(k) is minimal
28. How to choose k?
Use validation with leave-one-out method
For k = 1, 2, …, K
Err(k) = 0;
1. Randomly select a training data point and
hide its class label
2. Using the remaining data and given k to
predict the class label for the left data point
3. Err(k) = Err(k) + 1 if the predicted label is
different from the true label
Repeat the procedure until all training examples
are tested
Choose the k whose Err(k) is minimal
29. How to choose k?
Use validation with leave-one-out method
For k = 1, 2, …, K
Err(k) = 0;
1. Randomly select a training data point and
hide its class label
2. Using the remaining data and given k to
predict the class label for the left data point
3. Err(k) = Err(k) + 1 if the predicted label is
different from the true label
Repeat the procedure until all training examples
are tested
Choose the k whose Err(k) is minimal
(k=1)
30. How to choose k?
Use validation with leave-one-out method
For k = 1, 2, …, K
Err(k) = 0;
1. Randomly select a training data point and
hide its class label
2. Using the remaining data and given k to
predict the class label for the left data point
3. Err(k) = Err(k) + 1 if the predicted label is
different from the true label
Repeat the procedure until all training examples
are tested
Choose the k whose Err(k) is minimal
Err(1) = 1
31. How to choose k?
Use validation with leave-one-out method
For k = 1, 2, …, K
Err(k) = 0;
1. Randomly select a training data point and
hide its class label
2. Using the remaining data and given k to
predict the class label for the left data point
3. Err(k) = Err(k) + 1 if the predicted label is
different from the true label
Repeat the procedure until all training examples
are tested
Choose the k whose Err(k) is minimal
Err(1) = 1
32. How to choose k?
Use validation with leave-one-out method
For k = 1, 2, …, K
Err(k) = 0;
1. Randomly select a training data point and
hide its class label
2. Using the remaining data and given k to
predict the class label for the left data point
3. Err(k) = Err(k) + 1 if the predicted label is
different from the true label
Repeat the procedure until all training examples
are tested
Choose the k whose Err(k) is minimal
Err(1) = 3
Err(2) = 2
Err(3) = 6
k = 2
33. K-nearest neighbor for discrete classes
Algorithm (parameter k)
For each training example (X,C(X)) add the example to our training
list.
When a new example Xq arrives, assign class:
C(Xq) = majority voting on the k nearest neighbors of Xq
C(Xq) = argmax v Σi δ(v, C(Xi))
where δ(a,b) = 1 if a = b and 0 otherwise
34. K-nearest neighbor for real-valued functions
Algorithm (parameter k)
For each training example (X,C(X))
add the example to our training list.
When a new example Xq arrives, assign class:
C(Xq) = average value among k nearest neighbors of Xq
C(Xq) = Σ C(Xi) / k
35. Distance Weighted Nearest Neighbor
It makes sense to weight the contribution of each
example according to the distance to the new query
example.
C(Xq) = argmax v Σi wi δ(v, C(Xi))
For example, wi = 1 / d(Xq,Xi)
36. Nearest Neighbor
Four things make a memory based learner:
1. A distance metric
Euclidian
2. How many nearby neighbors to look at?
k
3. A weighting function (optional)
1 / d(Xq,Xi)
4. How to fit with the local points?
Just predict the same output as the nearest neighbor.
37. Distance Weighted Nearest Neighbor for Real-Valued
Functions
For real valued functions we average based on the weight
function and normalize using the sum of all weights.
C(Xq) = Σi wi C(Xi) / Σ wi
38. Problems with k-nearest Neighbor
The distance between examples is based on all attributes. What if some
attributes are irrelevant?
Consider the curse of dimensionality.
The larger the number of irrelevant attributes, the higher the effect on the
nearest-neighbor rule.
One solution is to use weights on the attributes. This is like stretching or
contracting the dimensions on the input space.
Ideally we would like to eliminate all irrelevant attributes.
39. Locally Weighted Regression
Let’s remember some terminology:
Regression: Is a problem similar to classification but the value to predict
is a real number.
Residual: The difference between the true target value f and our
approximation f’: f(X) – f’(X)
Kernel Function: The distance function that provides a weight to each
example. The kernel function K is a function of the distance between
examples: K = f(d(Xi,Xq))
40. Locally Weighted Regression
The method is called locally weighted regression for the following
reasons:
“Locally” because the predicted value for an example Xq is based only on
the vicinity or neighborhood around Xq.
“Weighted” because the contribution of each neighbor of Xq will depend
on the distance between the neighbor example and Xq.
“Regression” because the value to predict will be a real number.
41. Locally Weighted Regression
Consider the problem of approximating a target function using a
linear combination of attribute values:
f’(X) = w0 + w1x1 + w2x2 + … + wnxn
where X = (x1, x2, …, xn)
We want to find those coefficients that minimize the error: E = ½ Σk [f(X)
– f’(X)]2
42. Locally Weighted Regression
If we do this in the vicinity of an example Xq and we wish to use a
kernel function, we get a form of locally weighted regression:
E(Xq) = ½ Σk ( [f(X) – f’(X)]2 K(d(Xq,X) )
where the sum now goes over the neighbors of Xq.
43. Locally Weighted Regression
Using gradient descent search, the update rule is defined
as:
ΔΔ Wj = n Σk [f(X) – f’(X)] K(d(Xq,X) xj
where n is the learning rate and xj is the jth attribute of example
X.
45. Nearest Neighbor
1. A distance metric
Scaled Euclidian
2. How many nearby neighbors to look at?
All of them
3. A weighting function (optional)
w_k = exp(-D(x_k , x_query )^2 / Kw^2 )
4. How to fit with the local points?
First form a local linear model. Find the β that
minimizes the locally weighted sum of squared residuals:
46. Locally Weighted Regression
Remarks:
The literature contains other functions that are non linear.
There are many variations to locally weighted regression that use
different kernel functions.
Normally a linear model is sufficiently good to approximate the local
neighborhood of an example.