1) The document discusses conformal predictions, a machine learning technique that provides calibrated predictions along with confidence levels or regions.
2) Conformal predictions work by dividing data into training and calibration sets, fitting a model on training data, and using the calibration set to estimate prediction confidence without assumptions about the data distribution.
3) The document outlines how conformal predictions are applied in classification and regression problems, providing algorithms to compute prediction regions at a given confidence level for new observations.
4) As an example application, the document shows how conformal predictors are used for a car insurance classification problem to determine the confidence of predictions while minimizing false positives.
This document discusses the DEMATEL (Decision Making Trial and Evaluation Laboratory) method and its revised version. DEMATEL is used to analyze complex systems and determine the relationships between factors. The original DEMATEL assumes the total relationship matrix converges to zero, but this is not always true. The revised DEMATEL addresses this by introducing a small positive number to the normalization process, guaranteeing convergence. Steps for both methods and an example are provided to demonstrate the infeasibility of the original DEMATEL in some cases.
A Quick Introduction to the Chow Liu AlgorithmJee Vang, Ph.D.
The Chow-Liu algorithm is used to construct a maximum weight spanning tree from a dataset that approximates the dependencies between variables. It runs in O(n2 log(n)) time and finds the tree distribution closest to the true distribution according to Kullback-Leibler divergence. The algorithm takes a dataset and measures of association between variable pairs as input and outputs a dependency tree by sequentially connecting the most associated variable pairs.
Luteal phase support in assisted reproductive technology (ART)Dr. Sunita Chandra
This document provides information on luteal phase support in assisted reproductive technology (ART). It discusses that the luteal phase is characterized by progesterone secretion from the corpus luteum, which is important for endometrial receptivity and early pregnancy development. However, in stimulated ART cycles, high estrogen levels can cause luteal phase insufficiency. Various options for luteal phase support are discussed, including human chorionic gonadotropin, progestogens administered vaginally, intramuscularly or orally, and gonadotropin-releasing hormone agonists. The benefits and limitations of different agents are summarized.
This document discusses endometriosis and its relationship to infertility. It covers several key points:
1. Endometriosis has three main types - peritoneal, ovarian, and rectovaginal - which are different entities.
2. Endometriosis can result in infertility through mechanical effects, endocrine abnormalities, changes to peritoneal fluid, immune system issues, and defects in oocytes.
3. Diagnosis is confirmed through laparoscopy, and mild or minimal endometriosis associated with infertility can be treated through laparoscopic destruction, expectant management, or GnRH agonists. Surgery aims to decrease inflammation and toxicity.
4. For endometriomas,
This presentation provides an introduction to the Particle Swarm Optimization topic, it shows the PSO basic idea, PSO parameters, advantages, limitations and the related applications.
Dr. Sujoy Dasgupta is a reproductive medicine specialist who has extensive training and experience in managing male factor infertility. He discusses various cases involving different causes of male infertility such as mild abnormalities, varicocele, genetic issues, ejaculatory disorders, and physical abnormalities. He analyzes the diagnostic workup and treatment strategies for each case, including lifestyle modifications, medical therapy, surgical correction, assisted reproduction techniques, and counseling regarding prognosis.
This document discusses the case of a 33-year-old woman who has been married for 7 years with no living children and a history of 3 miscarriages between 16-20 weeks gestation. The next step is recommended to be a 3D pelvic ultrasound to assess the uterine anatomy and identify any congenital uterine abnormalities such as a septate or bicornuate uterus. If 3D ultrasound is not available, sonohysterography could be used instead. For a confirmed uterine septum, hysteroscopic resection may be considered to improve live birth and decrease miscarriage rates. For pregnancy, cervical cerclage would not be recommended in this case given her history does not meet criteria for history-indicated or ultrasound-
This document discusses the DEMATEL (Decision Making Trial and Evaluation Laboratory) method and its revised version. DEMATEL is used to analyze complex systems and determine the relationships between factors. The original DEMATEL assumes the total relationship matrix converges to zero, but this is not always true. The revised DEMATEL addresses this by introducing a small positive number to the normalization process, guaranteeing convergence. Steps for both methods and an example are provided to demonstrate the infeasibility of the original DEMATEL in some cases.
A Quick Introduction to the Chow Liu AlgorithmJee Vang, Ph.D.
The Chow-Liu algorithm is used to construct a maximum weight spanning tree from a dataset that approximates the dependencies between variables. It runs in O(n2 log(n)) time and finds the tree distribution closest to the true distribution according to Kullback-Leibler divergence. The algorithm takes a dataset and measures of association between variable pairs as input and outputs a dependency tree by sequentially connecting the most associated variable pairs.
Luteal phase support in assisted reproductive technology (ART)Dr. Sunita Chandra
This document provides information on luteal phase support in assisted reproductive technology (ART). It discusses that the luteal phase is characterized by progesterone secretion from the corpus luteum, which is important for endometrial receptivity and early pregnancy development. However, in stimulated ART cycles, high estrogen levels can cause luteal phase insufficiency. Various options for luteal phase support are discussed, including human chorionic gonadotropin, progestogens administered vaginally, intramuscularly or orally, and gonadotropin-releasing hormone agonists. The benefits and limitations of different agents are summarized.
This document discusses endometriosis and its relationship to infertility. It covers several key points:
1. Endometriosis has three main types - peritoneal, ovarian, and rectovaginal - which are different entities.
2. Endometriosis can result in infertility through mechanical effects, endocrine abnormalities, changes to peritoneal fluid, immune system issues, and defects in oocytes.
3. Diagnosis is confirmed through laparoscopy, and mild or minimal endometriosis associated with infertility can be treated through laparoscopic destruction, expectant management, or GnRH agonists. Surgery aims to decrease inflammation and toxicity.
4. For endometriomas,
This presentation provides an introduction to the Particle Swarm Optimization topic, it shows the PSO basic idea, PSO parameters, advantages, limitations and the related applications.
Dr. Sujoy Dasgupta is a reproductive medicine specialist who has extensive training and experience in managing male factor infertility. He discusses various cases involving different causes of male infertility such as mild abnormalities, varicocele, genetic issues, ejaculatory disorders, and physical abnormalities. He analyzes the diagnostic workup and treatment strategies for each case, including lifestyle modifications, medical therapy, surgical correction, assisted reproduction techniques, and counseling regarding prognosis.
This document discusses the case of a 33-year-old woman who has been married for 7 years with no living children and a history of 3 miscarriages between 16-20 weeks gestation. The next step is recommended to be a 3D pelvic ultrasound to assess the uterine anatomy and identify any congenital uterine abnormalities such as a septate or bicornuate uterus. If 3D ultrasound is not available, sonohysterography could be used instead. For a confirmed uterine septum, hysteroscopic resection may be considered to improve live birth and decrease miscarriage rates. For pregnancy, cervical cerclage would not be recommended in this case given her history does not meet criteria for history-indicated or ultrasound-
Science, practice and evidence are dynamic processes. This is typically vivid when it relates to Polycystic Ovarian Syndrome. PCOS is the commonest hyperandrogenic disorder in women and one of the most common causes of ovulatory infertility. Although polycystic ovaries were first described by the Italian scientist Vallisneri in 1721, it was largely forgotten until the 1930s, and then renamed after its rediscoverers as Stein-Leventhal syndrome. Even then, it still wasn’t until the invention of the ultrasound scanner in the 1980s and consensus of diagnosis in the early 1990s that PCOS was recognized on a wider scale in women of reproductive age. When attempting to diagnose with precision something that is complex, it is important that we first clearly define what it is we are trying to diagnose. PCOS is today seen as a heterogeneous syndrome where a range of symptoms may be present or absent, and may overlap with other conditions, it is perhaps best viewed as a spectrum of symptoms, pathologic findings and laboratory abnormalities. PCOS can be difficult to conceptualize, even for experts, as shown by the fact that there have been several different ways of diagnosing it over the years.
More recently, the fundamental role of hyperandrogenism has been pointed out.
However, PCOS compromises other pathological conditions that strongly modify the phenotype and play a dominant role in the pathophysiology of the disorder, including insulin resistance and hyperinsulinemia, obesity and metabolic disorders, all favoring together with androgen excess, an increased susceptibility to develop type 2 diabetes mellitus (T2DM) and, possibly, cardiovascular diseases. PCOS by itself may also have some genetic component as documented by familial aggregation and recent genetic studies. All the clinical features may however change throughout the lifespan, starting from adolescence to postmenopausal age. Therefore, PCOS should be considered as a lifetime disorder.
I sincerely hope that with the recommended readings attached and lecture, you will be able to strengthen your knowledge, thereby providing evidence-based medicine practice for the management of PCOS in a successful manner to improve and better women’s Health care. The best investment you can make is an investment in yourself. The more you learn, the more you’ll earn (Warren Buffett), so read as much as you can.
Thank You.
Regards: Rafi Rozan
Gradient boosting in practice: a deep dive into xgboostJaroslaw Szymczak
The document discusses tuning parameters for the XGBoost gradient boosting algorithm. It explores different parameters like max_depth, learning_rate, and n_estimators using a news article classification dataset. Experiments are performed to evaluate the effect of these parameters on model accuracy and training time. The learning curves are also plotted to analyze model performance over iterations.
This document discusses different types of matchings in graphs. A matching is a set of edges without common vertices. A maximum matching contains the largest possible number of edges. A maximal matching is one where no additional edges can be added without violating the matching property. A perfect matching is where every vertex is incident to exactly one edge, making it both maximum and maximal. The document provides definitions and properties of different matchings in graphs.
This document presents a new meta-heuristic optimization algorithm called Cuckoo Search (CS) that is inspired by the brood parasitism of some cuckoo species and the Lévy flight behavior of some birds and insects. The CS algorithm is formulated based on three idealized rules: each cuckoo lays one egg in a randomly selected nest; the best nests with high-quality eggs are carried over to subsequent generations; and a portion of the worst nests are abandoned. New solutions in CS are generated through Lévy flights. The performance of CS is validated on benchmark test functions and compared to genetic algorithms and particle swarm optimization. Results show that CS can find global optima efficiently.
OVARIAN REJUVENATION - ROLE OF PLATELET RICH PLASMA THERAPY BY DR SHASHWAT JANIDR SHASHWAT JANI
The document discusses ovarian rejuvenation using platelet-rich plasma (PRP) injections. PRP is extracted from a patient's own blood and contains growth factors that may reawaken dormant follicles in the ovaries. The procedure involves extracting blood, centrifuging it to separate PRP from other blood components, and injecting the PRP into the ovaries under ultrasound guidance, usually once a month for three months. The goal is to stimulate egg maturation and development, helping patients conceive. Side effects are minimal and may include pain, fever, or internal bleeding. Follow-up monitors hormone levels to check for signs of improved ovarian function.
Clique Relaxation Models in Networks: Theory, Algorithms, and ApplicationsSSA KPI
This document discusses clique relaxation models in networks. It provides an introduction to graph theory basics and defines common clique relaxation concepts like k-cliques, k-clubs, and k-plexes. The document outlines topics on the theory, algorithms, and applications of clique relaxations and discusses finding cohesive subgroups in social networks and other applications.
Hysteroscopy plays a crucial role in the management of infertility by allowing direct visualization of the uterine cavity. It can be used for both diagnostic and therapeutic purposes to assess and treat intrauterine abnormalities that may cause infertility such as submucosal fibroids, polyps, and adhesions. Hysteroscopy has advantages over other imaging modalities as it provides more accurate diagnosis and allows treatment to be performed in the same setting. Common procedures include polypectomy, myomectomy and adhesiolysis which have been shown to improve fertility outcomes.
This document provides guidelines for thromboprophylaxis during pregnancy, labor, and after vaginal delivery. It outlines various risk factors for venous thromboembolism (VTE) including pre-existing conditions like previous DVT or thrombophilia, as well as transient risks from procedures, immobilization, or medical complications. It recommends individual assessment and management based on risk factor profile, including consideration of antenatal low molecular weight heparin for high risk groups like those with previous VTE or inherited thrombophilia. Postpartum prophylaxis for at least 6 weeks is also advised for many groups based on their VTE risk.
Hysteroscopic surgery can effectively treat various intrauterine pathologies that cause infertility such as polyps, fibroids, adhesions, and septums. It allows for direct visualization and removal of abnormalities, improving chances for spontaneous or assisted conception. While hysteroscopy is considered the gold standard for diagnosing intrauterine issues, less invasive methods like ultrasound and HSG are usually sufficient. Routine hysteroscopy before first IVF is not recommended as it does not improve live birth rates, but may be beneficial after repeated failures. Operative hysteroscopy can significantly enhance fertility outcomes.
Bidirectional graph search techniques for finding shortest path in image base...Navin Kumar
The intriguing problem of solving a maze comes
under the territory of algorithms and artificial intelligence.
The maze solving using computers is quite of interest for many
researchers, hence, there had been many previous attempts to
come up with a solution which is optimum in terms of time and
space. Some of the best performing algorithms suitable for the
problem are breadth-first search, A* algorithm, best-first
search and many others which ultimately are the
enhancement of these basic algorithms. The images are
converted into graph data structures after which an algorithm
is applied eventually pointing the trace of the solution on the
maze image. This paper is an attempt to do the same by
implementing the bidirectional version of these well-known
algorithms and study their performance with the former. The
bidirectional approach is indeed capable of providing
improved results at an expense of space. The vital part of the
approach is to find the meeting point of the two bidirectional
searches which will be guaranteed to meet if there exists any
solution.
PANEL DISCUSSION ON PRACTICAL APPROACH TO ENDOMETRIOSISWith FOCUS ON DINOGESTLifecare Centre
PANEL DISCUSSION ON PRACTICAL APPROACH TO ENDOMETRIOSISWith FOCUS ON DINOGEST
UMA RAI
RAJ BOKARIA
JYOTI AGARWAL
JYOTI BHASKER
RENU CHAWLA
DIPTI NABH
VANDANA GUPTA
Particle swarm optimization is a population-based stochastic optimization technique inspired by bird flocking or fish schooling. It works by having a population of candidate solutions, called particles, and moving these particles around in the search space according to simple mathematical formulae over the particle's position and velocity. Each particle keeps track of its coordinates in the problem space which are associated with the best solution that particle has achieved so far. The main idea is that hope flies along with the flock.
PANEL DISCUSSION
MANAGEMENT OF PCOS - WOMB to TOMB
MODERATOR : Sharda Jain
PANELISTS : Dr.Chitra setia
Dr Puneet Arora
Dr. Ila Gupta
Dr. Rupam Arora
Dr. Archana Sharma
Dr. Sangeeta Gupta
Dermatologists
Dr. V.K. Upadhyay
Dr. S. Kandhari
The document discusses fuzzy logic and fuzzy sets. It begins by explaining fuzzy logic is used to model imprecise concepts and dependencies using natural language terms. It then defines fuzzy variables, universes of discourse, and fuzzy sets which have membership functions assigning a degree of membership between 0 and 1. Operations on fuzzy sets like intersection, union, and complement are also covered. The document also discusses fuzzy rules, relations, and approximate reasoning using max-min inference.
Selective progesterone receptor modulators (SPRMs)
Stimulates growth :
Up regulating epidermal growth factor (EGF)
Down regulating tumour necrosis factor-alpha expression
Inhibits growth :
Downregulating insulin-like growth factor-1 (IGF-1) expression
NO EFFECT ON ESTRADIOL LEVELS
Mifepristone : 5 or 10 mg per day for 1 year
Ulipristal acetate: 5-10mg/day for 13 weeks
Pro apoptotic and anti-proliferative effects on fibroid cells
The document discusses the Newton-Raphson method for solving nonlinear equations. It involves starting with an initial guess for the root and iteratively improving the estimate using the function value and derivative until reaching an acceptable approximation of the true root. The method relies on linearizing the function using its tangent line to generate improved estimates in each step.
WEKA:Credibility Evaluating Whats Been Learnedweka Content
- Training and test sets are used to measure classification success rates, with the test set being independent of the training set. The error rate on the training set is optimistic. Cross validation techniques like 10-fold stratified cross validation are used when data is limited.
- True success rates are predicted using properties of statistics and normal distributions. Confidence levels determine the range within which the true rate is expected to lie.
- Techniques like paired t-tests are used to statistically compare the performance of different algorithms or data mining methods. They determine if performance differences are statistically significant.
This document discusses various techniques for evaluating machine learning models and comparing their performance, including:
- Measuring error rates on separate test and training sets to avoid overfitting
- Using techniques like cross-validation, bootstrapping, and holdout validation when data is limited
- Comparing algorithms using statistical tests like paired t-tests
- Accounting for costs of different prediction outcomes in evaluation and model training
- Visualizing performance using lift charts and ROC curves to compare models
- The Minimum Description Length principle for selecting the model that best compresses the data
1) The paper introduces the influence function for interpreting black-box machine learning models. The influence function traces a model's predictions back to the training data by examining how the model's parameters would change if a particular training point was removed or perturbed.
2) The influence function approximates this change in parameters by assuming a quadratic approximation to the empirical risk function around the learned parameters and taking a single Newton step. It shows the parameter change due to removing a point is approximated by the influence function.
3) The paper demonstrates how the influence function can be used to understand model behavior, find adversarial examples, debug issues, and correct errors, among other applications. It also proposes practical methods to compute the influence function for
Understanding Blackbox Prediction via Influence FunctionsSEMINARGROOT
Pang Wei Koh and Percy Liang
"Understanding Black-Box prediction via influence functions" ICML 2017 Best paper
References:
https://youtu.be/0w9fLX_T6tY
https://arxiv.org/abs/1703.04730
A presentation about NGBoost (Natural Gradient Boosting) which I presented in the Information Theory and Probabilistic Programming course at the University of Oklahoma.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Science, practice and evidence are dynamic processes. This is typically vivid when it relates to Polycystic Ovarian Syndrome. PCOS is the commonest hyperandrogenic disorder in women and one of the most common causes of ovulatory infertility. Although polycystic ovaries were first described by the Italian scientist Vallisneri in 1721, it was largely forgotten until the 1930s, and then renamed after its rediscoverers as Stein-Leventhal syndrome. Even then, it still wasn’t until the invention of the ultrasound scanner in the 1980s and consensus of diagnosis in the early 1990s that PCOS was recognized on a wider scale in women of reproductive age. When attempting to diagnose with precision something that is complex, it is important that we first clearly define what it is we are trying to diagnose. PCOS is today seen as a heterogeneous syndrome where a range of symptoms may be present or absent, and may overlap with other conditions, it is perhaps best viewed as a spectrum of symptoms, pathologic findings and laboratory abnormalities. PCOS can be difficult to conceptualize, even for experts, as shown by the fact that there have been several different ways of diagnosing it over the years.
More recently, the fundamental role of hyperandrogenism has been pointed out.
However, PCOS compromises other pathological conditions that strongly modify the phenotype and play a dominant role in the pathophysiology of the disorder, including insulin resistance and hyperinsulinemia, obesity and metabolic disorders, all favoring together with androgen excess, an increased susceptibility to develop type 2 diabetes mellitus (T2DM) and, possibly, cardiovascular diseases. PCOS by itself may also have some genetic component as documented by familial aggregation and recent genetic studies. All the clinical features may however change throughout the lifespan, starting from adolescence to postmenopausal age. Therefore, PCOS should be considered as a lifetime disorder.
I sincerely hope that with the recommended readings attached and lecture, you will be able to strengthen your knowledge, thereby providing evidence-based medicine practice for the management of PCOS in a successful manner to improve and better women’s Health care. The best investment you can make is an investment in yourself. The more you learn, the more you’ll earn (Warren Buffett), so read as much as you can.
Thank You.
Regards: Rafi Rozan
Gradient boosting in practice: a deep dive into xgboostJaroslaw Szymczak
The document discusses tuning parameters for the XGBoost gradient boosting algorithm. It explores different parameters like max_depth, learning_rate, and n_estimators using a news article classification dataset. Experiments are performed to evaluate the effect of these parameters on model accuracy and training time. The learning curves are also plotted to analyze model performance over iterations.
This document discusses different types of matchings in graphs. A matching is a set of edges without common vertices. A maximum matching contains the largest possible number of edges. A maximal matching is one where no additional edges can be added without violating the matching property. A perfect matching is where every vertex is incident to exactly one edge, making it both maximum and maximal. The document provides definitions and properties of different matchings in graphs.
This document presents a new meta-heuristic optimization algorithm called Cuckoo Search (CS) that is inspired by the brood parasitism of some cuckoo species and the Lévy flight behavior of some birds and insects. The CS algorithm is formulated based on three idealized rules: each cuckoo lays one egg in a randomly selected nest; the best nests with high-quality eggs are carried over to subsequent generations; and a portion of the worst nests are abandoned. New solutions in CS are generated through Lévy flights. The performance of CS is validated on benchmark test functions and compared to genetic algorithms and particle swarm optimization. Results show that CS can find global optima efficiently.
OVARIAN REJUVENATION - ROLE OF PLATELET RICH PLASMA THERAPY BY DR SHASHWAT JANIDR SHASHWAT JANI
The document discusses ovarian rejuvenation using platelet-rich plasma (PRP) injections. PRP is extracted from a patient's own blood and contains growth factors that may reawaken dormant follicles in the ovaries. The procedure involves extracting blood, centrifuging it to separate PRP from other blood components, and injecting the PRP into the ovaries under ultrasound guidance, usually once a month for three months. The goal is to stimulate egg maturation and development, helping patients conceive. Side effects are minimal and may include pain, fever, or internal bleeding. Follow-up monitors hormone levels to check for signs of improved ovarian function.
Clique Relaxation Models in Networks: Theory, Algorithms, and ApplicationsSSA KPI
This document discusses clique relaxation models in networks. It provides an introduction to graph theory basics and defines common clique relaxation concepts like k-cliques, k-clubs, and k-plexes. The document outlines topics on the theory, algorithms, and applications of clique relaxations and discusses finding cohesive subgroups in social networks and other applications.
Hysteroscopy plays a crucial role in the management of infertility by allowing direct visualization of the uterine cavity. It can be used for both diagnostic and therapeutic purposes to assess and treat intrauterine abnormalities that may cause infertility such as submucosal fibroids, polyps, and adhesions. Hysteroscopy has advantages over other imaging modalities as it provides more accurate diagnosis and allows treatment to be performed in the same setting. Common procedures include polypectomy, myomectomy and adhesiolysis which have been shown to improve fertility outcomes.
This document provides guidelines for thromboprophylaxis during pregnancy, labor, and after vaginal delivery. It outlines various risk factors for venous thromboembolism (VTE) including pre-existing conditions like previous DVT or thrombophilia, as well as transient risks from procedures, immobilization, or medical complications. It recommends individual assessment and management based on risk factor profile, including consideration of antenatal low molecular weight heparin for high risk groups like those with previous VTE or inherited thrombophilia. Postpartum prophylaxis for at least 6 weeks is also advised for many groups based on their VTE risk.
Hysteroscopic surgery can effectively treat various intrauterine pathologies that cause infertility such as polyps, fibroids, adhesions, and septums. It allows for direct visualization and removal of abnormalities, improving chances for spontaneous or assisted conception. While hysteroscopy is considered the gold standard for diagnosing intrauterine issues, less invasive methods like ultrasound and HSG are usually sufficient. Routine hysteroscopy before first IVF is not recommended as it does not improve live birth rates, but may be beneficial after repeated failures. Operative hysteroscopy can significantly enhance fertility outcomes.
Bidirectional graph search techniques for finding shortest path in image base...Navin Kumar
The intriguing problem of solving a maze comes
under the territory of algorithms and artificial intelligence.
The maze solving using computers is quite of interest for many
researchers, hence, there had been many previous attempts to
come up with a solution which is optimum in terms of time and
space. Some of the best performing algorithms suitable for the
problem are breadth-first search, A* algorithm, best-first
search and many others which ultimately are the
enhancement of these basic algorithms. The images are
converted into graph data structures after which an algorithm
is applied eventually pointing the trace of the solution on the
maze image. This paper is an attempt to do the same by
implementing the bidirectional version of these well-known
algorithms and study their performance with the former. The
bidirectional approach is indeed capable of providing
improved results at an expense of space. The vital part of the
approach is to find the meeting point of the two bidirectional
searches which will be guaranteed to meet if there exists any
solution.
PANEL DISCUSSION ON PRACTICAL APPROACH TO ENDOMETRIOSISWith FOCUS ON DINOGESTLifecare Centre
PANEL DISCUSSION ON PRACTICAL APPROACH TO ENDOMETRIOSISWith FOCUS ON DINOGEST
UMA RAI
RAJ BOKARIA
JYOTI AGARWAL
JYOTI BHASKER
RENU CHAWLA
DIPTI NABH
VANDANA GUPTA
Particle swarm optimization is a population-based stochastic optimization technique inspired by bird flocking or fish schooling. It works by having a population of candidate solutions, called particles, and moving these particles around in the search space according to simple mathematical formulae over the particle's position and velocity. Each particle keeps track of its coordinates in the problem space which are associated with the best solution that particle has achieved so far. The main idea is that hope flies along with the flock.
PANEL DISCUSSION
MANAGEMENT OF PCOS - WOMB to TOMB
MODERATOR : Sharda Jain
PANELISTS : Dr.Chitra setia
Dr Puneet Arora
Dr. Ila Gupta
Dr. Rupam Arora
Dr. Archana Sharma
Dr. Sangeeta Gupta
Dermatologists
Dr. V.K. Upadhyay
Dr. S. Kandhari
The document discusses fuzzy logic and fuzzy sets. It begins by explaining fuzzy logic is used to model imprecise concepts and dependencies using natural language terms. It then defines fuzzy variables, universes of discourse, and fuzzy sets which have membership functions assigning a degree of membership between 0 and 1. Operations on fuzzy sets like intersection, union, and complement are also covered. The document also discusses fuzzy rules, relations, and approximate reasoning using max-min inference.
Selective progesterone receptor modulators (SPRMs)
Stimulates growth :
Up regulating epidermal growth factor (EGF)
Down regulating tumour necrosis factor-alpha expression
Inhibits growth :
Downregulating insulin-like growth factor-1 (IGF-1) expression
NO EFFECT ON ESTRADIOL LEVELS
Mifepristone : 5 or 10 mg per day for 1 year
Ulipristal acetate: 5-10mg/day for 13 weeks
Pro apoptotic and anti-proliferative effects on fibroid cells
The document discusses the Newton-Raphson method for solving nonlinear equations. It involves starting with an initial guess for the root and iteratively improving the estimate using the function value and derivative until reaching an acceptable approximation of the true root. The method relies on linearizing the function using its tangent line to generate improved estimates in each step.
WEKA:Credibility Evaluating Whats Been Learnedweka Content
- Training and test sets are used to measure classification success rates, with the test set being independent of the training set. The error rate on the training set is optimistic. Cross validation techniques like 10-fold stratified cross validation are used when data is limited.
- True success rates are predicted using properties of statistics and normal distributions. Confidence levels determine the range within which the true rate is expected to lie.
- Techniques like paired t-tests are used to statistically compare the performance of different algorithms or data mining methods. They determine if performance differences are statistically significant.
This document discusses various techniques for evaluating machine learning models and comparing their performance, including:
- Measuring error rates on separate test and training sets to avoid overfitting
- Using techniques like cross-validation, bootstrapping, and holdout validation when data is limited
- Comparing algorithms using statistical tests like paired t-tests
- Accounting for costs of different prediction outcomes in evaluation and model training
- Visualizing performance using lift charts and ROC curves to compare models
- The Minimum Description Length principle for selecting the model that best compresses the data
1) The paper introduces the influence function for interpreting black-box machine learning models. The influence function traces a model's predictions back to the training data by examining how the model's parameters would change if a particular training point was removed or perturbed.
2) The influence function approximates this change in parameters by assuming a quadratic approximation to the empirical risk function around the learned parameters and taking a single Newton step. It shows the parameter change due to removing a point is approximated by the influence function.
3) The paper demonstrates how the influence function can be used to understand model behavior, find adversarial examples, debug issues, and correct errors, among other applications. It also proposes practical methods to compute the influence function for
Understanding Blackbox Prediction via Influence FunctionsSEMINARGROOT
Pang Wei Koh and Percy Liang
"Understanding Black-Box prediction via influence functions" ICML 2017 Best paper
References:
https://youtu.be/0w9fLX_T6tY
https://arxiv.org/abs/1703.04730
A presentation about NGBoost (Natural Gradient Boosting) which I presented in the Information Theory and Probabilistic Programming course at the University of Oklahoma.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Logistic regression is a machine learning classification algorithm used to predict the probability of a categorical dependent variable given one or more independent variables. It uses a logit link function to transform the probability values into odds ratios between 0 and infinity. The model is trained by minimizing a cost function called logistic loss using gradient descent optimization. Model performance is evaluated using metrics like accuracy, precision, recall, and the confusion matrix, and can be optimized by adjusting the probability threshold for classifications.
The document provides an overview of machine learning, including definitions, types of machine learning (supervised, unsupervised, reinforcement learning), and evaluation metrics for machine learning models. It discusses classification metrics like accuracy, precision, recall, F1 score, and confusion matrices. For regression problems, it covers metrics like mean absolute error, mean squared error, R2 score. It also provides examples of calculating many of these common metrics in Python.
This document summarizes the NGBoost method for probabilistic regression. NGBoost uses gradient boosting to fit the parameters of an assumed probabilistic distribution for the target variable. It improves on existing probabilistic regression methods by using the natural gradient, which performs gradient descent in the space of distributions rather than the parameter space. This addresses issues with prior approaches and allows NGBoost to achieve state-of-the-art performance while remaining fast, flexible, and scalable. Future work may apply NGBoost to other problems like survival analysis or joint outcome regression.
This document summarizes several methods for estimating causal effects from observational data:
1. The back-door criterion provides a method for identifying when causal effects are identifiable based on observable variables. It requires adjusting for a set of variables S that block back-door paths between the treatment X and outcome Y.
2. Estimation methods described include calculating average treatment effects, avoiding estimating high-dimensional marginal distributions using sampling, matching on propensity scores, and using instrumental variables.
3. Propensity score matching involves estimating propensity scores via logistic regression and then matching treated and control units based on their propensity to receive treatment.
4. Instrumental variables estimation uses an instrument I that is associated with treatment X
1) Machine learning is a field of artificial intelligence that allows computers to learn without being explicitly programmed by finding patterns in data.
2) There are three main types of machine learning problems: supervised learning which uses labeled training data, unsupervised learning which finds hidden patterns in unlabeled data, and reinforcement learning where a system learns from feedback of rewards and punishments.
3) Key machine learning concepts include linear regression, which finds a linear relationship between variables, and gradient descent, an algorithm for minimizing cost functions to optimize model parameters like slope and intercept of a linear regression line.
This document summarizes domain adaptation from a theoretical machine learning perspective. It begins with an introduction to domain adaptation and an outline. It then provides background on machine learning concepts like empirical risk minimization and PAC learning. It formulates the domain adaptation problem and introduces a classifier-induced divergence measure to quantify differences between domains. A key theoretical guarantee is presented, bounding the target risk by the source risk plus a divergence term and constants. Finally, an example application to domain-adversarial neural networks is mentioned.
This document discusses supervised learning. Supervised learning uses labeled training data to train models to predict outputs for new data. Examples given include weather prediction apps, spam filters, and Netflix recommendations. Supervised learning algorithms are selected based on whether the target variable is categorical or continuous. Classification algorithms are used when the target is categorical while regression is used for continuous targets. Common regression algorithms discussed include linear regression, logistic regression, ridge regression, lasso regression, and elastic net. Metrics for evaluating supervised learning models include accuracy, R-squared, adjusted R-squared, mean squared error, and coefficients/p-values. The document also covers challenges like overfitting and regularization techniques to address it.
1) The document discusses using machine learning techniques like logistic regression, random forest, and K-means clustering to develop a credit scoring model based on financial ratios to predict a company's probability of default.
2) Random forest performed the best with an AUC of 0.87 and high precision and recall, while logistic regression had a AUC of 0.75 but issues with type II errors.
3) K-means clustering had a lower precision predicting defaults but an acceptable F1-score and AUC of 0.80.
The document discusses decision trees and ensemble methods. It begins with an agenda that covers the bias-variance tradeoff, generalizations of this concept, the ExtraTrees algorithm, its sklearn interface, and conclusions. It then reviews decision trees, plotting sample data and walking through how the tree would split the data. Next, it covers the general CART algorithm and different impurity measures. It discusses controlling overfitting via tree depth and other techniques. Finally, it delves into explaining the bias-variance decomposition and tradeoff in more detail.
Study on Evaluation of Venture Capital Based onInteractive Projection Algorithminventionjournals
International Journal of Business and Management Invention (IJBMI) is an international journal intended for professionals and researchers in all fields of Business and Management. IJBMI publishes research articles and reviews within the whole field Business and Management, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
This document provides an introduction to machine learning concepts including linear regression, linear classification, and the cross-entropy loss function. It discusses using gradient descent to fit machine learning models by minimizing a loss function on training data. Specifically, it describes how linear regression can be solved using mean squared error and gradient descent, and how linear classifiers can be trained with the cross-entropy loss and softmax activations. The goal is to choose model parameters that minimize the loss function for a given dataset.
1. Cross-validation is commonly used to evaluate machine learning algorithms and estimate their performance on new data. It involves partitioning the dataset into training and test sets and measuring the accuracy on the held-out test sets.
2. Tuning sets are often used to select hyperparameters like the number of hidden units. Performance on the tuning set is used to estimate future performance on new examples.
3. Statistical tests like paired t-tests are used to determine if differences in performance between algorithms on test sets are statistically significant.
1. Cross-validation is commonly used to evaluate machine learning algorithms and estimate their performance on new data. It involves partitioning the dataset into training and test sets and measuring the accuracy on the held-out test sets.
2. Tuning sets are often used to select hyperparameters like the number of hidden units. Performance on the tuning set is used to estimate future performance on new examples.
3. Statistical tests like paired t-tests are used to determine if differences in performance between algorithms on test sets are statistically significant.
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...sameer shah
"Join us for STATATHON, a dynamic 2-day event dedicated to exploring statistical knowledge and its real-world applications. From theory to practice, participants engage in intensive learning sessions, workshops, and challenges, fostering a deeper understanding of statistical methodologies and their significance in various fields."
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataKiwi Creative
Harness the power of AI-backed reports, benchmarking and data analysis to predict trends and detect anomalies in your marketing efforts.
Peter Caputa, CEO at Databox, reveals how you can discover the strategies and tools to increase your growth rate (and margins!).
From metrics to track to data habits to pick up, enhance your reporting for powerful insights to improve your B2B tech company's marketing.
- - -
This is the webinar recording from the June 2024 HubSpot User Group (HUG) for B2B Technology USA.
Watch the video recording at https://youtu.be/5vjwGfPN9lw
Sign up for future HUG events at https://events.hubspot.com/b2b-technology-usa/
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...Social Samosa
The Modern Marketing Reckoner (MMR) is a comprehensive resource packed with POVs from 60+ industry leaders on how AI is transforming the 4 key pillars of marketing – product, place, price and promotions.
1. How good is your prediction?
Quantifying uncertainty in Machine Learning predictions
PyData London 2019 (12th- 14th July)
Maria Navarro
2. Outline
Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
3. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
4. Motivating example
How good is your prediction?
Problem To find out whether a car is a total lo
To do it we have:
1. A set of historical observations 𝑥1; 𝑦1 , ⋯ , 𝑥 𝑁; 𝑦 𝑁 , where:
• 𝑥𝑖 describes the accident by age of the driver, model of the car, etc.
• 𝑦𝑖 is a label which identifies whether the car is reparable or not
2. A machine learning algorithm (h 𝑥 = 𝑦)
PROBLEM: To find out whether a car is a total loss or not
5. Motivating example
How good is your prediction, REALLY?
A new accident, 𝑥 𝑁+1, occurs. We run our model, and we obtain the following results:
1. The car is classified as total loss
2. The probability of total loss according to our model is 0.85
3. The model is roughly 91% accurate in training, test and validation sets, so we expect same
behaviour in production data
4. The model has an AUC of 0.88 in training, so again that is what we expect in production data
What do these measurements mean?
Do we have any guarantee about accident 𝑥 𝑁+1?
Are we confident about the prediction?
6. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
7. Introduction to conformal predictions
Why Conformal Predictions (CP) ?
1. There are several ad hoc ways to obtain some confidence around your predictions (resampling
methods, assume normality, etc.)
2. Conformal predictions assumes very little about the outcome you are trying to predict. It only
assume exchangeability.
3. It can be used with any machine learning algorithm.
4. It provides error bounds at a confidence level that we can select.
5. Probabilities are well-calibrated.
6. It is easy to implement.
7. The framework has been proven:
V. Vovk, A. Gammerman, G. Shafer
Algorithmic learning in a random walk, Springer 2005.
8. Introduction to conformal predictions
General idea
• Let 𝑍 be a probability distribution.
• f z → ℝ some function.
• We draw 5 samples from the distribution 𝑍 and apply 𝑓 𝑧 :
𝑓 𝑧𝑖 = 𝛼𝑖, with 𝑖 = 1, … , 5
For simplicity, we assume 𝛼1 ≤ 𝛼2 ≤ 𝛼3 ≤ 𝛼4 ≤ 𝛼5
• We estimate the cumulative distribution function (CDF) for the scores:
0 0.2 0.4 0.6 0.8 1
𝛼1 𝛼2 𝛼3 𝛼4 𝛼5
• We draw a new sample from z ∈ 𝑍. We assume exchangeability and compute 𝑓 𝑧 = 𝛼.
• We can estimate its probability: 𝑃 𝛼 ≤ 𝛼4 = 0.6 and 𝑃 𝛼 ≤ 𝛼2 = 0.2
9. Introduction to conformal predictions
Relation to our problem
• Let 𝑧𝑖 = 𝑥𝑖; 𝑦𝑖 with 𝑖 = 1, … , 𝑝 be a sample of the probability distribution, 𝑍 = 𝑋, 𝑌 , where:
𝑥𝑖 is our observables and 𝑦𝑖 the target we want to predict
• We define 𝑓 𝑧𝑖 = 𝑦𝑖 − ℎ 𝑥𝑖 , where:
ℎ 𝑥𝑖 is a regression model train on 𝑧𝑖 with 𝑖 = 5, … , 𝑝
• We apply 𝑓 𝑧 to the 5 remaining samples
𝑓 𝑧𝑖 = 𝛼𝑖, with 𝑖 = 1, … , 5
We can compute the exact values 0.10 ≤ 0.13 ≤ 0.28 ≤ 0.30 ≤ 0.38
• We estimate the cumulative distribution function (CDF) for the scores:
0 0.2 0.4 0.6 0.8 1
0.10 0.13 0.28 0.30 0.38
• We draw a new sample from z ∈ 𝑍. We assume exchangeability and compute 𝑓 𝑧 = 𝑦 − ℎ 𝑥 = 𝑦 − 2 .
• We can estimate its probability:
𝑃 𝑦 − 2 ≤ 0.30 = 0.6 and 𝑃 𝑦 − 2 ≤ 0.28 = 0.4
𝑃 𝑦 ∈ 2 ± 0.30 = 0.6 and 𝑃 𝑦 ∈ 2 ± 0.30 = 0.4
𝑦 𝜖 1.7, 2.3 with probability 0.6
10. Introduction to conformal predictions
Inputs for conformal predictions
• A set of training examples 𝑧𝑖 = 𝑥𝑖, 𝑦𝑖 with 𝑖 = 1, … , 𝑃
They must be drawn from an exchangeable distribution (the order of observations is
irrelevant).
• A non-conformity function 𝑓 𝑧 → ℝ
It measures the “weirdness” of an example 𝑥𝑖, 𝑦𝑖
It should give low scores to similar examples 𝑥𝑖, 𝑦𝑖 and high scores to different ones
𝑥𝑖, ¬𝑦𝑖
Common choice is take some function of the underlying model, but it can be anything: the
probability estimate for correct class, distance to neighbours with same class, probability from
the trees, absolute error of a regression model, etc.
• Set a significance level 𝛆 ∈ (0,1), so 1 − 𝜀 confidence level
11. Introduction to conformal predictions
How does conformal predictions work?
• Divide training set into two disjoint sets: 𝑍𝑡 with 𝑍𝑡 = 𝑚 and 𝑍 𝑐 with 𝑍 𝑐 = 𝑛, 𝑚 + 𝑛 = 𝑝
• Build the underlying model, ℎ, using 𝑍𝑡
• Apply 𝑓 𝑧𝑖 = 𝛼𝑖 to the elements of the set you did not use for training ℎ , and estimate its probability
distribution 𝛼1, … , 𝛼 𝑛 ~ 𝑄
• If a new example comes in 𝑥, ℎ 𝑥 = 𝑦 , then we will reject 𝑦
We will reject 𝑦 if 𝑓 (𝑥, 𝑦) = 𝛼 𝑦 does not belong to 𝑄
• We compute the non-conformity degree which is called p-value as follows:
𝑝 𝑦=
𝑧 𝑗 𝜖 𝑍 𝑐∶ 𝛼 𝑗 ≥ 𝛼 𝑦
𝑛+1
, 𝑝 𝑦 is the p-value
• Finally the prediction region:
Γ 𝜀
= 𝑦 𝜖 𝑌: 𝑝 𝑦 > 𝜀
Is 𝒚 a very non-conforming example?
12. Introduction to conformal predictions
Conformal prediction output
The prediction region Γ 𝜀
contains prediction 𝑦 with probability 1 − 𝜀
In classification :
𝛼 𝑦 is know, but we need to compute 𝑝 𝑦
The result is a set of labels:
Γ 𝜀
= 𝐶𝑙𝑎𝑠𝑠1, 𝐶𝑙𝑎𝑠𝑠3, 𝐶𝑙𝑎𝑠𝑠5 s. t. 𝑃 𝑦 ∈ Γ 𝜀
= 1 − 𝜀
o If Γ 𝜀
= ∅ , then always erroneous
o If Γ 𝜀
= 𝐶 (only one class), then always true (if it is the correct class)
o If Γ 𝜀
= 𝐶𝑙𝑎𝑠𝑠1, 𝐶𝑙𝑎𝑠𝑠3, … , 𝐶𝑙𝑎𝑠𝑠5 (several classes), then always correct
In regression is an interval:
𝑝 𝑦 is know, but we need to compute 𝛼 𝑦
The result is an interval:
Γ 𝜀
= 𝑎, 𝑏 where 𝑎, 𝑏 ∈ ℝ and s. t. 𝑃 𝑦 ∈ Γ 𝜀
= 1 − 𝜀
13. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
14. Conformal predictions in classification
Algorithm to compute conformal prediction regions in classification problems
Let 𝑍 = 𝑋, 𝑌 be the historical data set for our classification problem, where:
𝑍 = 𝑝, 𝑋 is the information about the problem and 𝑌 = 𝐶1 , … , 𝐶𝑠 set of labels.
𝑍 is exchangeable.
To obtain the prediction region:
1. Divide 𝑍 into two disjoint sets:
𝑍𝑡 proper training set with 𝑍𝑡 = 𝑚
𝑍 𝑐 calibration set with 𝑍 𝑐 = 𝑛
2. Fit a classifier, ℎ 𝑋 = 𝑌, using 𝑍𝑡
3. Define a non-conformity function 𝑓 𝑧 to measure the weirdness of your samples
4. Apply 𝑓 𝑧 to each element in 𝑍 𝑐 to obtain the calibration scores: 𝛼1, … , 𝛼 𝑛
5. Set a significance level 𝜀 𝜖 0, 1
15. Conformal predictions in classification
Algorithm to compute conformal predictions in classification problems
6. For a new sample 𝑥, 𝑦 compute the scoring value for each label in 𝑌:
∀ 𝐶𝑖 𝜖 𝑌 𝑓 𝑥, 𝑦 = 𝐶𝑖 = 𝛼 𝐶 𝑖
7. For each label in 𝑌 compute the p-value as follows:
∀ 𝐶𝑖 𝜖 𝑌 𝑝 𝐶 𝑖
=
𝑧 𝑗 𝜖 𝑍 𝑐∶ 𝛼 𝑗 ≥𝛼 𝐶 𝑖
𝑛+1
8. Finally build the prediction region as follows:
Γ 𝜀
= 𝐶𝑖 𝜖 𝑌: 𝑝 𝐶 𝑖
> 𝜀 , then
for the new prediction ℎ 𝑥 = 𝑦, 𝑃 𝑦 𝜖 Γ 𝜀
= 1 − ε
16. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
17. Conformal predictions in regression
Algorithm to compute conformal prediction regions in regression problems
Let 𝑍 = 𝑋, 𝑌 be the historical data set for our classification problem, where:
𝑍 = 𝑝, 𝑋 is the information about the problem and 𝑌 a continuous target.
𝑍 is exchangeable.
To obtain the prediction region:
1. Divide 𝑍 into two disjoint sets:
𝑍𝑡 proper training set with 𝑍𝑡 = 𝑚
𝑍 𝑐 calibration set with 𝑍 𝑐 = 𝑛
2. Fit a regression model, ℎ 𝑋 = 𝑌, using 𝑍𝑡
3. Define a non-conformity function 𝑓 𝑧 to measure the weirdness of your samples
4. Apply 𝑓 𝑧 to each element in 𝑍 𝑐 to obtain the calibration scores: 𝛼1, … , 𝛼 𝑛
5. Set a significance level 𝜀 𝜖 0, 1
18. Conformal predictions in regression
Algorithm to compute conformal predictions in regression problems
6. Sort calibrations scores 𝛼1, … , 𝛼 𝑛 in a descending order
7. Compute the index 𝑠 = 𝜀 𝑛 + 1
This is the index of the (1 − ε)-percentile of the non-conformity score 𝛼 𝑠
8. Finally the prediction region for a new sample:
Γ 𝜀
= ℎ 𝑥𝑖 ± 𝛼 𝑠, with 𝑃 ℎ(𝑥𝑖)𝜖 Γ 𝜀
= 1 − ε
19. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
20. Application
Classification with conformal predictors
• The dataset is imbalanced (Total Loss is the minority class)
• The model is XGBoost
• Model performance:
• A new accident happens the model says it is a Total Loss, but how confident we are?
• Due to business restrictions we have to minimize the number false positives in TL
PROBLEM: To find out whether a car is a total loss or not
21. Application
Classification with conformal predictors
• We take the test set, 𝑍𝑡𝑒𝑠𝑡 = (𝑥𝑖, 𝑦𝑖) with 𝑖 = 1, … , 𝑀
• We define a non-conformity function:
𝑓 𝑧 =
𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑐𝑙𝑎𝑠𝑠 𝑖 + 𝑐𝑎𝑙𝑖𝑏𝑟𝑎𝑡𝑒𝑑 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑐𝑙𝑎𝑠𝑠 𝑖
2
where:
𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦𝑐𝑙𝑎𝑠𝑠 𝑖 according to the model that 𝑦 = 𝑐𝑙𝑎𝑠𝑠 𝑖
𝑐𝑎𝑙𝑖𝑏𝑟𝑎𝑡𝑒𝑑 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦𝑐𝑙𝑎𝑠𝑠 𝑖 recalibrated probability that 𝑦 = 𝑐𝑙𝑎𝑠𝑠 𝑖
22. Application
Classification with conformal predictors
• Let us assume 𝑀 = 9 and apply 𝑓 𝑧 to each 𝑧𝑖 𝜖 𝑍𝑡𝑒𝑠𝑡
• We order the scores, and use them to compute the p-value per label for the new accident:
TL = 0.85 p-value TL = 8/(9+1) = 0.8 > 𝜀 = 0.05
Non-TL = 0.15 p-value non-TL = 2/(9+1) = 0.2 > 𝜀 = 0.05
Γ 𝜀
= 𝑇𝐿, 𝑛𝑜𝑛 − 𝑇𝐿 s. t. 𝑃 𝑦 ∈ Γ 𝜀
= 0.95
25. Application
Regression with conformal predictors
• The dataset is not correctly label there were some inconsistencies.
• The model is XGBoost.
• Model performance:
• The model output was the input to another model
PROBLEM: to compute/find out the price of a car
28. Application
Regression with conformal predictors
• We take the test set, 𝑍𝑡𝑒𝑠𝑡 = (𝑥𝑖, 𝑦𝑖) with 𝑖 = 1, … , 𝑀
• We define a non-conformity function:
𝑓 𝑧 = 𝑦 − ℎ(𝑥)
where:
𝑦 is the true value, and ℎ(𝑥) the model prediction
• Let us assume 𝑀 = 9 and apply 𝑓 𝑧 to each 𝑧𝑖 𝜖 𝑍𝑡𝑒𝑠𝑡
• We order in descending order
• We set 𝜀 = 0.2, then the index of the score 𝑠 = 0.2 ∙ 9 + 1 = 2 𝛼 𝑠=2
• The fixed width conformal interval would be: ℎ(𝑥) ± 189.52
30. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
31. Summary and conclusions
Take away
• Good model performance does not mean trustable predictions.
• Conformal predictions is a useful tool with different applications.
• It is easy to understand and to implement.
• Define a non-conformity function is not always easy.
• Confident areound predictions bring some
32. Motivating example
Introduction to conformal predictions
Conformal predictions in classification
Conformal predictions in regression
Application
Summary and conclusions
References
34. References
Some interesting readings
1. V. Vovk, A. Gammerman, G. Shafer, Algorithm learning in a random walk, Springer, 2005.
2. H. Linusson, An introduction to conformal predictions, 2017.
3. V. Vovk, Cross-conformal predictors, Annals of Mathematics and Artificial Intelligence, 1-20, 2013.
4. U. Johannsson, H. Bostrom, T. Lofstrom, H. Linusson, Regression conformal predictors with
random forest, Machine Learning, 95, 155-176, 2014.
5. V. Balasubramanian, S-S. Ho, V. Vovk, Conformal predictions for reliable machine learning, Science
Direct Journal and Book, 2014.
35. How is your prediction? Quantifying uncertainty in Machine Learning
predictions
Questions