The Medical Segmentation Decathlon provides a benchmark for evaluating the generalizability of semantic segmentation algorithms across a variety of anatomical structures and imaging modalities. The Decathlon includes 10 segmentation tasks with over 2,600 unique patient datasets. In Phase 1 of the challenge, participants developed algorithms to segment the structures and submitted results for evaluation. The top performing methods for each task are identified based on Dice scores and boundary accuracy metrics. Phase 2 will involve applying the previously developed algorithms to new datasets without modifications, to further evaluate generalizability.
Chap 8. Optimization for training deep modelsYoung-Geun Choi
연구실 내부 세미나 자료. Goodfellow et al. (2016), Deep Learning, MIT Press의 Chapter 8을 요약/발췌하였습니다. 깊은 신경망(deep neural network) 모형 훈련시 목적함수 최적화 방법으로 흔히 사용되는 방법들을 소개합니다.
This presentation material provides an introduction to graph grammar and its application to learning a graph generative model. Presented at IBIS 2019, Nagoya, Japan.
Unit 3 random number generation, random-variate generationraksharao
This document discusses random number generation and random variate generation. It covers:
1) Properties of random numbers such as uniformity, independence, maximum density, and maximum period.
2) Techniques for generating pseudo-random numbers such as the linear congruential method and combined linear congruential generators.
3) Tests for random numbers including Kolmogorov-Smirnov, chi-square, and autocorrelation tests.
4) Random variate generation techniques like the inverse transform method, acceptance-rejection technique, and special properties for distributions like normal, lognormal, and Erlang.
Adversarial machine learning for av softwarejunseok seo
Introduce practical guidances for developing adversarial machine model for anti-malware software. I didn't use reinforcement model yet, just proof-of-concept. If you have any questions about my work, email to me :)
nababora@naver.com
Pixel transforms,
Color transforms,
Histogram processing & equalization ,
Filtering,
Convolution,
Fourier transformation and its applications in sharpening,
Blurring and noise removal
Chap 8. Optimization for training deep modelsYoung-Geun Choi
연구실 내부 세미나 자료. Goodfellow et al. (2016), Deep Learning, MIT Press의 Chapter 8을 요약/발췌하였습니다. 깊은 신경망(deep neural network) 모형 훈련시 목적함수 최적화 방법으로 흔히 사용되는 방법들을 소개합니다.
This presentation material provides an introduction to graph grammar and its application to learning a graph generative model. Presented at IBIS 2019, Nagoya, Japan.
Unit 3 random number generation, random-variate generationraksharao
This document discusses random number generation and random variate generation. It covers:
1) Properties of random numbers such as uniformity, independence, maximum density, and maximum period.
2) Techniques for generating pseudo-random numbers such as the linear congruential method and combined linear congruential generators.
3) Tests for random numbers including Kolmogorov-Smirnov, chi-square, and autocorrelation tests.
4) Random variate generation techniques like the inverse transform method, acceptance-rejection technique, and special properties for distributions like normal, lognormal, and Erlang.
Adversarial machine learning for av softwarejunseok seo
Introduce practical guidances for developing adversarial machine model for anti-malware software. I didn't use reinforcement model yet, just proof-of-concept. If you have any questions about my work, email to me :)
nababora@naver.com
Pixel transforms,
Color transforms,
Histogram processing & equalization ,
Filtering,
Convolution,
Fourier transformation and its applications in sharpening,
Blurring and noise removal
This document provides an overview of Naive Bayes classification. It begins with background on classification methods, then covers Bayes' theorem and how it relates to Bayesian and maximum likelihood classification. The document introduces Naive Bayes classification, which makes a strong independence assumption to simplify probability calculations. It discusses algorithms for discrete and continuous features, and addresses common issues like dealing with zero probabilities. The document concludes by outlining some applications of Naive Bayes classification and its advantages of simplicity and effectiveness for many problems.
This presentation introduces anti-aliasing techniques. It begins with defining anti-aliasing as a technique to reduce jagged edges by softening lines and edges. It then shows before and after examples of anti-aliased versus non-anti-aliased polygons. Finally, it discusses the two main anti-aliasing approaches of area sampling and super-sampling and provides details on how each technique works.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
Image processing involves manipulating digital images through algorithms implemented on computers. A digital image is composed of picture elements called pixels arranged in a grid. Each pixel represents a color or intensity value. Common image processing tasks include computer vision, optical character recognition, medical imaging, and more. Key concepts in image processing include pixels, resolution, color depth, and filtering/manipulating pixel values.
Lec8: Medical Image Segmentation (II) (Region Growing/Merging)Ulaş Bağcı
. Region Growing algorithm
• Homogeneity Criteria
• Split/Merge algorithm
• Examples from CT, MRI, PET
• Limitations
Image Filtering, Enhancement, Noise Reduction, and Signal Processing • MedicalImageRegistration • MedicalImageSegmentation • MedicalImageVisualization • Machine Learning in Medical Imaging • Shape Modeling/Analysis of Medical Images Deep Learning in Radiology
Histograms show the distribution of pixel intensities in an image by counting the number of pixels for each intensity value. Normalized histograms provide an estimate of the probability of each intensity occurring. Histogram equalization transforms the pixel intensity distribution of an image to a uniform distribution in order to increase contrast. It does this by using the cumulative distribution function to map intensities to new output values. Local histogram equalization performs this on neighborhoods within an image to enhance local details. Arithmetic and logical operations can also be used for image enhancement, such as AND, OR, and subtraction between images on a pixel-by-pixel basis.
The document discusses key concepts in digital image fundamentals including:
1. The electromagnetic spectrum and how light attributes like intensity and luminance are measured.
2. How digital images are acquired through image sensing and sampling/quantization.
3. Methods for representing digital images through matrices and binary values, and how resolution affects gray-level detail.
4. Digital zooming techniques like nearest neighbor, bilinear, and bicubic interpolation that control blurring and edge effects.
5. Concepts like pixel adjacency, connectivity, and distance measures between pixels.
Lab manual of Digital image processing using python by khalid Shaikhkhalidsheikh24
This document is a practical workbook for a digital image processing course. It contains 8 lab sessions where a student learns how to install Python and PyCharm, read and display images, extract image pixel information, convert images between color spaces and formats, apply filters like blurring, and perform operations like edge detection and resizing. Each lab has the objective, task description, source code, and output for tasks related to foundational digital image processing techniques.
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
Slides for talk Abhishek Sharma and I gave at the Gennovation tech talks (https://gennovationtalks.com/) at Genesis. The talk was part of outreach for the Deep Learning Enthusiasts meetup group at San Francisco. My part of the talk is covered from slides 19-34.
Lec11: Active Contour and Level Set for Medical Image SegmentationUlaş Bağcı
ActiveContour(Snake) • LevelSet
• Applications
Enhancement, Noise Reduction, and Signal Processing • MedicalImageRegistration • MedicalImageSegmentation • MedicalImageVisualization • Machine Learning in Medical Imaging • Shape Modeling/Analysis of Medical Images Deep Learning in Radiology Fuzzy Connectivity (FC) – Affinity functions • Absolute FC • Relative FC (and Iterative Relative FC) • Successful example applications of FC in medical imaging • Segmentation of Airway and Airway Walls using RFC based method Energyfunctional – Data and Smoothness terms • GraphCut – Min cut – Max Flow • ApplicationsinRadiologyImages
Pattern Recognition is the branch of machine learning a computer science which deals with the regularities and patterns in the data that can further be used to classify and categorize the data with the help of Pattern Recognition System.
“The assignment of a physical object or event to one of several pre-specified categories”-- Duda & Hart
Pattern Recognition System is responsible for generating patterns and similarities among given problem/data space, that can further be used to generate solutions to complex problems effectively and efficiently.
Certain problems that can be solved by humans, can also be made to be solved by machine by using this process.
The document provides an overview of machine learning, including definitions of key concepts. It discusses what machine learning and artificial intelligence are, gives examples of machine learning applications, and describes different types of machine learning systems such as supervised, unsupervised, and reinforcement learning. It also outlines steps for getting started in machine learning, including recommended mathematics, programming, and tool knowledge.
1) The document discusses different types of attention mechanisms in CNNs including self-attention and simplified attention for recalibration.
2) It reviews the evolution of CNN architectures including AlexNet, VGG, ResNet and variants, DenseNet, ResNeXt, Xception, MobileNet and ShuffleNet.
3) These attention mechanisms and CNN architectures are applied to tasks like image recognition, machine translation and image captioning.
This document discusses decision theory and its applications in machine learning. It describes how decision theory uses probability to make optimal decisions given input and target data. It also discusses how to minimize expected error and loss when making predictions. Finally, it explains how inference and decision problems can be broken into two stages and different models like generative, discriminative, and discriminant functions can be used.
Image Segmentation
Types of Image Segmentation
Semantic Segmentation
Instance Segmentation
Types of Image Segmentation Techniques based on the image properties:
Threshold Method.
Edge Based Segmentation.
Region-Based Segmentation.
Clustering Based Segmentation.
Watershed Based Method.
Artificial Neural Network Based Segmentation.
This document discusses support vector machines (SVMs) for classification. It explains that SVMs find the optimal separating hyperplane that maximizes the margin between positive and negative examples. This is formulated as a convex optimization problem. Both primal and dual formulations are presented, with the dual having fewer variables that scale with the number of examples rather than dimensions. Methods for handling non-separable data using soft margins and kernels for nonlinear classification are also summarized. Popular kernel functions like polynomial and Gaussian kernels are mentioned.
The Hough transform is a feature extraction technique used in image analysis and computer vision to detect shapes within images. It works by detecting imperfect instances of objects of a certain class of shapes via a voting procedure. Specifically, the Hough transform can be used to detect lines, circles, and other shapes in an image if their parametric equations are known, and it provides robust detection even under noise and partial occlusion. It works by quantizing the parameter space that describes the shape and counting the number of votes each parametric description receives from edge points in the image.
This document discusses computer aided detection (CAD) of abnormalities in medical images. It begins by outlining CAD and some of the key machine learning challenges, including correlated training data, non-standard evaluation metrics, runtime constraints, lack of objective ground truths, and data shortages. It then describes solutions like multiple instance learning, batch classification, cascaded classifiers, crowdsourcing algorithms, and multi-task learning. The document concludes by reviewing the clinical impact of CAD systems through several independent studies, which demonstrated improved radiologist performance and sensitivity in detecting diseases.
Top 10 Data Science Practitioner PitfallsSri Ambati
Top 10 Data Science Practitioner Pitfalls Meetup with Erin LeDell and Mark Landry on 09.09.15
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
This document provides an overview of Naive Bayes classification. It begins with background on classification methods, then covers Bayes' theorem and how it relates to Bayesian and maximum likelihood classification. The document introduces Naive Bayes classification, which makes a strong independence assumption to simplify probability calculations. It discusses algorithms for discrete and continuous features, and addresses common issues like dealing with zero probabilities. The document concludes by outlining some applications of Naive Bayes classification and its advantages of simplicity and effectiveness for many problems.
This presentation introduces anti-aliasing techniques. It begins with defining anti-aliasing as a technique to reduce jagged edges by softening lines and edges. It then shows before and after examples of anti-aliased versus non-anti-aliased polygons. Finally, it discusses the two main anti-aliasing approaches of area sampling and super-sampling and provides details on how each technique works.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
Image processing involves manipulating digital images through algorithms implemented on computers. A digital image is composed of picture elements called pixels arranged in a grid. Each pixel represents a color or intensity value. Common image processing tasks include computer vision, optical character recognition, medical imaging, and more. Key concepts in image processing include pixels, resolution, color depth, and filtering/manipulating pixel values.
Lec8: Medical Image Segmentation (II) (Region Growing/Merging)Ulaş Bağcı
. Region Growing algorithm
• Homogeneity Criteria
• Split/Merge algorithm
• Examples from CT, MRI, PET
• Limitations
Image Filtering, Enhancement, Noise Reduction, and Signal Processing • MedicalImageRegistration • MedicalImageSegmentation • MedicalImageVisualization • Machine Learning in Medical Imaging • Shape Modeling/Analysis of Medical Images Deep Learning in Radiology
Histograms show the distribution of pixel intensities in an image by counting the number of pixels for each intensity value. Normalized histograms provide an estimate of the probability of each intensity occurring. Histogram equalization transforms the pixel intensity distribution of an image to a uniform distribution in order to increase contrast. It does this by using the cumulative distribution function to map intensities to new output values. Local histogram equalization performs this on neighborhoods within an image to enhance local details. Arithmetic and logical operations can also be used for image enhancement, such as AND, OR, and subtraction between images on a pixel-by-pixel basis.
The document discusses key concepts in digital image fundamentals including:
1. The electromagnetic spectrum and how light attributes like intensity and luminance are measured.
2. How digital images are acquired through image sensing and sampling/quantization.
3. Methods for representing digital images through matrices and binary values, and how resolution affects gray-level detail.
4. Digital zooming techniques like nearest neighbor, bilinear, and bicubic interpolation that control blurring and edge effects.
5. Concepts like pixel adjacency, connectivity, and distance measures between pixels.
Lab manual of Digital image processing using python by khalid Shaikhkhalidsheikh24
This document is a practical workbook for a digital image processing course. It contains 8 lab sessions where a student learns how to install Python and PyCharm, read and display images, extract image pixel information, convert images between color spaces and formats, apply filters like blurring, and perform operations like edge detection and resizing. Each lab has the objective, task description, source code, and output for tasks related to foundational digital image processing techniques.
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
Slides for talk Abhishek Sharma and I gave at the Gennovation tech talks (https://gennovationtalks.com/) at Genesis. The talk was part of outreach for the Deep Learning Enthusiasts meetup group at San Francisco. My part of the talk is covered from slides 19-34.
Lec11: Active Contour and Level Set for Medical Image SegmentationUlaş Bağcı
ActiveContour(Snake) • LevelSet
• Applications
Enhancement, Noise Reduction, and Signal Processing • MedicalImageRegistration • MedicalImageSegmentation • MedicalImageVisualization • Machine Learning in Medical Imaging • Shape Modeling/Analysis of Medical Images Deep Learning in Radiology Fuzzy Connectivity (FC) – Affinity functions • Absolute FC • Relative FC (and Iterative Relative FC) • Successful example applications of FC in medical imaging • Segmentation of Airway and Airway Walls using RFC based method Energyfunctional – Data and Smoothness terms • GraphCut – Min cut – Max Flow • ApplicationsinRadiologyImages
Pattern Recognition is the branch of machine learning a computer science which deals with the regularities and patterns in the data that can further be used to classify and categorize the data with the help of Pattern Recognition System.
“The assignment of a physical object or event to one of several pre-specified categories”-- Duda & Hart
Pattern Recognition System is responsible for generating patterns and similarities among given problem/data space, that can further be used to generate solutions to complex problems effectively and efficiently.
Certain problems that can be solved by humans, can also be made to be solved by machine by using this process.
The document provides an overview of machine learning, including definitions of key concepts. It discusses what machine learning and artificial intelligence are, gives examples of machine learning applications, and describes different types of machine learning systems such as supervised, unsupervised, and reinforcement learning. It also outlines steps for getting started in machine learning, including recommended mathematics, programming, and tool knowledge.
1) The document discusses different types of attention mechanisms in CNNs including self-attention and simplified attention for recalibration.
2) It reviews the evolution of CNN architectures including AlexNet, VGG, ResNet and variants, DenseNet, ResNeXt, Xception, MobileNet and ShuffleNet.
3) These attention mechanisms and CNN architectures are applied to tasks like image recognition, machine translation and image captioning.
This document discusses decision theory and its applications in machine learning. It describes how decision theory uses probability to make optimal decisions given input and target data. It also discusses how to minimize expected error and loss when making predictions. Finally, it explains how inference and decision problems can be broken into two stages and different models like generative, discriminative, and discriminant functions can be used.
Image Segmentation
Types of Image Segmentation
Semantic Segmentation
Instance Segmentation
Types of Image Segmentation Techniques based on the image properties:
Threshold Method.
Edge Based Segmentation.
Region-Based Segmentation.
Clustering Based Segmentation.
Watershed Based Method.
Artificial Neural Network Based Segmentation.
This document discusses support vector machines (SVMs) for classification. It explains that SVMs find the optimal separating hyperplane that maximizes the margin between positive and negative examples. This is formulated as a convex optimization problem. Both primal and dual formulations are presented, with the dual having fewer variables that scale with the number of examples rather than dimensions. Methods for handling non-separable data using soft margins and kernels for nonlinear classification are also summarized. Popular kernel functions like polynomial and Gaussian kernels are mentioned.
The Hough transform is a feature extraction technique used in image analysis and computer vision to detect shapes within images. It works by detecting imperfect instances of objects of a certain class of shapes via a voting procedure. Specifically, the Hough transform can be used to detect lines, circles, and other shapes in an image if their parametric equations are known, and it provides robust detection even under noise and partial occlusion. It works by quantizing the parameter space that describes the shape and counting the number of votes each parametric description receives from edge points in the image.
This document discusses computer aided detection (CAD) of abnormalities in medical images. It begins by outlining CAD and some of the key machine learning challenges, including correlated training data, non-standard evaluation metrics, runtime constraints, lack of objective ground truths, and data shortages. It then describes solutions like multiple instance learning, batch classification, cascaded classifiers, crowdsourcing algorithms, and multi-task learning. The document concludes by reviewing the clinical impact of CAD systems through several independent studies, which demonstrated improved radiologist performance and sensitivity in detecting diseases.
Top 10 Data Science Practitioner PitfallsSri Ambati
Top 10 Data Science Practitioner Pitfalls Meetup with Erin LeDell and Mark Landry on 09.09.15
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
The document discusses using theory-based research to improve health informatics (HI). It provides examples of testing theories from fields like communication, decision-making, and behavior change to optimize eHealth interventions before randomized controlled trials. Specific theories and studies testing things like how alert formatting impacts prescribing are summarized. The document argues this approach can help establish HI as a professional discipline by building a scientific evidence base for more reliable eHealth tools.
Data Science for Business Managers - An intro to ROI for predictive analyticsAkin Osman Kazakci
This module addresses critical business aspects related to launching a predictive analytics project. How to establish the relationship with business KPIs is discussed. A notion of data hunt, for planning & acquiring external data for better predictions is introduced. Model quality and it's role for ROI of data and prediction tasks are explained. The module is concluded with a glimpse on how collaborative data challenges can improve predictive model quality in no time.
Developing and validating statistical models for clinical prediction and prog...Evangelos Kritsotakis
Talk on clinical prediction models presented at the Joint Seminar Series in Translational and Clinical Medicine organised by the University of Crete Medical School, the Institute of Molecular Biology and Biotechnology of the Foundation for Research and Technology Hellas (IMBB-FORTH), and the University of Crete Research Center (UCRC), Heraklion [online], Greece, April 7, 2021.
Machine Learning in Modern Medicine with Erin LeDell at Stanford MedSri Ambati
Machine learning and AI company H2O.ai presented on machine learning applications in modern medicine. They discussed how electronic health records, genomics, wearables, and other data sources can be used with machine learning for personalized healthcare, disease prediction and prevention. H2O's software platform allows building models at scale from large datasets using algorithms like random forests, deep learning and ensembles. Demonstrations showed predicting HIV treatment failure and classifying breast cancer malignancy from medical images, achieving high accuracy. H2O aims to make machine learning accessible and scalable for improving medical research and care.
AI in Healthcare: Real-World Machine Learning Use CasesHealth Catalyst
Levi Thatcher, PhD, VP of Data Science at Health Catalyst will share practical AI use cases and distill the lessons into a framework you can use when evaluating AI healthcare projects. Specifically, Levi will answer these questions:
What are great healthcare business cases for AI/ML?
What kind of data do you need?
What tools / talent do you need?
How do you integrate AI/ML into the daily workflow?
Batch -13.pptx lung cancer detection using transfer learninghananth1513
Embedded systems
Embedded systems are special-purpose computing systems embedded in application environments or in other computing systems and provide specialized support. The decreasing cost of processing power, combined with the decreasing cost of memory and the ability to design low-cost systems on chip, has led to the development and deployment of embedded computing systems in a wide range of application environments. Examples include network adapters for computing systems and mobile phones, control systems for air conditioning, industrial systems, and cars,
Hybrid filtering methods for feature selection in high-dimensional cancer dataIJECEIAES
Statisticians in both academia and industry have encountered problems with high-dimensional data. The rapid feature increase has caused the feature count to outstrip the instance count. There are several established methods when selecting features from massive amounts of breast cancer data. Even so, overfitting continues to be a problem. The challenge of choosing important features with minimum loss in a different sample size is another area with room for development. As a result, the feature selection technique is crucial for dealing with high-dimensional data classification issues. This paper proposed a new architecture for high-dimensional breast cancer data using filtering techniques and a logistic regression model. Essential features are filtered out using a combination of hybrid chi–square and hybrid information gain (hybrid IG) with logistic regression as classifier. The results showed that hybrid IG performed the best for high-dimensional breast and prostate cancer data. The top 50 and 22 features outperformed the other configurations, with the highest classification accuracies of 86.96% and 82.61%, respectively, after integrating the hybrid information gain and logistic function (hybrid IG+LR) with a sample size of 75. In the future, multiclass classification of multidimensional medical data to be evaluated using data from a different domain.
IRJET - Breast Cancer Prediction using Supervised Machine Learning Algorithms...IRJET Journal
This document describes a study that uses supervised machine learning algorithms to predict breast cancer. Three algorithms - decision tree, logistic regression, and random forest - are applied to preprocessed breast cancer data. The random forest model achieved the best accuracy at 98.6% for predicting whether a tumor was benign or malignant. The study aims to develop an early prediction system for breast cancer using machine learning techniques.
Statistics in the age of data science, issues you can not ignoreTuri, Inc.
This document discusses issues in statistics that data scientists can and cannot ignore when working with large datasets. It begins by outlining the talk and defining key terms in data science. It then explains that model assessment, such as estimating model performance on new data, becomes easier with more data as statistical adjustments are not needed. However, more data and variables are not always better, as noise, collinearity, and overfitting can still occur. Several examples are given where common machine learning algorithms can be fooled into achieving high accuracy on training data even when the target variable is random. The conclusion emphasizes that data science, statistics, and domain expertise each provide unique perspectives, and effective teams need to understand all views.
Design of an Intelligent System for Improving Classification of Cancer DiseasesMohamed Loey
The methodologies that depend on gene expression profile have been able to detect cancer since its inception. The previous works have spent great efforts to reach the best results. Some researchers have achieved excellent results in the classification process of cancer based on the gene expression profile using different gene selection approaches and different classifiers
Early detection of cancer increases the probability of recovery. This thesis presents an intelligent decision support system (IDSS) for early diagnosis of cancer-based on the microarray of gene expression profiles. The problem of this dataset is the little number of examples (not exceed hundreds) comparing to a large number of genes (in thousands). So, it became necessary to find out a method for reducing the features (genes) that are not relevant to the investigated disease to avoid overfitting. The proposed methodology used information gain (IG) for selecting the most important features from the input patterns. Then, the selected features (genes) are reduced by applying the Gray Wolf Optimization algorithm (GWO). Finally, the methodology exercises support vector machine (SVM) for cancer type classification. The proposed methodology was applied to three data sets (breast, colon, and CNS) and was evaluated by the classification accuracy performance measurement, which is most important in the diagnosis of diseases. The best results were gotten when integrating IG with GWO and SVM rating accuracy improved to 96.67% and the number of features was reduced to 32 feature of the CNS dataset.
This thesis investigates several classification algorithms and their suitability to the biological domain. For applications that suffer from high dimensionality, different feature selection methods are considered for illustration and analysis. Moreover, an effective system is proposed. In addition, Experiments were conducted on three benchmark gene expression datasets. The proposed system is assessed and compared with related work performance.
1) The document discusses different metrics for optimizing predictive models, noting that squared error can emphasize outliers while lift charts are better. It recommends not optimizing AUC alone.
2) Global search algorithms may be needed if the model and error metric are not simple. The goal of the project and what to optimize should be considered.
3) Case studies are presented showing how optimizing for the problem goal, like flagging account outliers for fraud detection, led to better outcomes than a general classification approach.
1) The document discusses how Cancer Research UK and its Drug Discovery Unit are using AI/ML techniques like deep learning for drug discovery applications such as identifying potential small molecules that can selectively induce senescence without cell death.
2) As a case study, they trained a neural network classification model on a dataset of compounds labeled as senescence inducers or not to predict new senescence inducing compounds.
3) The best model was an ensemble of 10 neural networks that was used to screen a virtual library of 2 million compounds, filtering the results to propose 147 compounds for further validation in cell-based screens.
ICU Patient Deterioration Prediction : A Data-Mining Approachcsandit
A huge amount of medical data is generated every da
y, which presents a challenge in analysing
these data. The obvious solution to this challenge
is to reduce the amount of data without
information loss. Dimension reduction is considered
the most popular approach for reducing
data size and also to reduce noise and redundancies
in data. In this paper, we investigate the
effect of feature selection in improving the predic
tion of patient deterioration in ICUs. We
consider lab tests as features. Thus, choosing a su
bset of features would mean choosing the
most important lab tests to perform. If the number
of tests can be reduced by identifying the
most important tests, then we could also identify t
he redundant tests. By omitting the redundant
tests, observation time could be reduced and early
treatment could be provided to avoid the risk.
Additionally, unnecessary monetary cost would be av
oided. Our approach uses state-of-the-art
feature selection for predicting ICU patient deteri
oration using the medical lab results. We
apply our technique on the publicly available MIMIC
-II database and show the effectiveness of
the feature selection. We also provide a detailed a
nalysis of the best features identified by our
approach.
ICU PATIENT DETERIORATION PREDICTION: A DATA-MINING APPROACHcscpconf
A huge amount of medical data is generated every day, which presents a challenge in analysing
these data. The obvious solution to this challenge is to reduce the amount of data without
information loss. Dimension reduction is considered the most popular approach for reducing
data size and also to reduce noise and redundancies in data. In this paper, we investigate the
effect of feature selection in improving the prediction of patient deterioration in ICUs. We
consider lab tests as features. Thus, choosing a subset of features would mean choosing the
most important lab tests to perform. If the number of tests can be reduced by identifying the
most important tests, then we could also identify the redundant tests. By omitting the redundant
tests, observation time could be reduced and early treatment could be provided to avoid the risk.
Additionally, unnecessary monetary cost would be avoided. Our approach uses state-of-the-art
feature selection for predicting ICU patient deterioration using the medical lab results. We
apply our technique on the publicly available MIMIC-II database and show the effectiveness of
the feature selection. We also provide a detailed analysis of the best features identified by our
approach.
AI is the science and engineering of creating intelligent machines and software. It draws from fields like computer science, biology, psychology and linguistics. The goal is to develop systems that can perform tasks normally requiring human intelligence, like visual perception, decision making and language translation. Some key applications of AI include machine learning, expert systems, natural language processing and computer vision. As AI systems continue advancing, they are becoming better than humans at certain tasks like playing strategic games.
This document discusses several common problems with data handling and quality including building and testing models with the same data, confusion between biological and technical replicates, and identification and handling of outliers. It provides examples and explanations of key concepts such as experimental and sampling units, pseudo-replication, outliers versus high influence points, and leverage plots. The importance of proper data handling techniques like dividing data into training, test, and confirmation sets and using cross-validation is emphasized to avoid overfitting models and generating spurious findings.
Big Data and machine learning are increasingly important in biomedical science and clinical practice. Big Data refers to large and complex datasets that are too large for traditional tools to handle. Machine learning involves algorithms that can recognize patterns in data without being explicitly programmed. Some challenges of working with big data and machine learning include issues with data volume, variety, and veracity. However, techniques like distributed analysis, standards, and validation can help address these challenges.
Making Radiology AI Models more robust: Federated Learning and other Approachesimgcommcall
Daniel Rubin discusses approaches for making AI models more robust by accessing larger amounts of medical image data. Centralized data pooling is challenging due to data sharing barriers. Federated learning, which trains models across sites without sharing patient data, is presented as an alternative. However, federated learning requires common data standards for image annotations. The talk explores existing annotation standards and tools that could enable federated learning to leverage multi-institutional medical image data for developing more generalizable AI models.
This document provides an agenda for the NCI Imaging Informatics Webinar held on July 6, 2020. The agenda included presentations on distributed learning of deep learning in medical imaging, an update on the MedICI website, an update on The Cancer Imaging Archive, and announcements about future community calls, the community call wiki page, and where recordings could be accessed. The next scheduled community calls were listed as August 3, 2020 and September 14, 2020.
The document discusses the American College of Radiology's (ACR) efforts to advance the appropriate use of data science and artificial intelligence in radiology. It provides details on ACR's Data Science Institute (DSI) programs and initiatives to help define, validate, and monitor AI algorithms. These include ACR Define-AI, ACR Assess-AI, and ACR Certify-AI. The DSI aims to establish ACR as a leader in the radiology AI ecosystem and ensure the safe, effective integration of AI into clinical practice and medical education. The rest of the document discusses ACR's AI-LAB program to pilot AI algorithms and collect clinical feedback at partner sites.
The document summarizes the agenda for an NCI Imaging Informatics Webinar on April 6, 2020. The webinar included presentations on PathPresenter, a web-based digital pathology and image viewer, and an update on The Cancer Imaging Archive. It was announced that the webinar recordings and slides would be made available online on the NCI Imaging Community Call Wiki page and SlideShare account. The next webinars were scheduled for May 4 and June 1, 2020.
The January 6, 2020 Imaging Community Call featured presentations on the Medical Segmentation Decathlon, Imaging Data Commons updates, NBIA updates, and TCIA updates. Announcements were made about recording the presentations, the community call wiki page with recordings and slides, and the next scheduled community calls in February and March 2020.
NCI Cancer Research Data Commons - Overviewimgcommcall
The NCI Cancer Research Data Commons aims to enable sharing of diverse cancer research data across institutions by providing easy access to data stored in domain-specific repositories through a common authentication and authorization mechanism. It utilizes a framework of reusable components including data nodes, a cancer data aggregator, and cloud resources to integrate genomic, imaging, proteomic, and other data types while controlling access. The goals are to facilitate discovery and analysis tools as well as sustainably sharing data publicly to advance cancer research.
Imaging Data Commons (IDC) - Introduction and intital approachimgcommcall
The document introduces the Imaging Data Commons (IDC) which will connect researchers to cancer image collections, metadata, and tools for searching, viewing, and analyzing imaging data and related data types. The IDC will build on existing technologies and collaborations, with an initial focus on radiology and pathology images stored in DICOM format. It will utilize public cancer image collections from the Cancer Imaging Archive and integrate with other nodes in the Cancer Research Data Commons. The team has experience with open-source imaging tools, cloud infrastructure, and standards development. The initial implementation phases will focus on defining the data model and use cases, evaluating existing tools, and developing a minimal viable product hosted on the Google Cloud platform.
The document outlines the agenda for an Imaging Community Call on July 1, 2019. The agenda included welcome remarks, overviews of the Clinical Proteomic Tumor Analysis Consortium project and how its image, proteomic, and genomic data can be accessed through various data portals. It concluded with announcements about joining a community call group, an imaging community call wiki, and details on the next call in August 2019 focusing on tissue cytometry presentations.
The document discusses the Office of Cancer Clinical Proteomics Research (OCCPR) and its tumor characterization programs, including the Clinical Proteomic Tumor Analysis Consortium (CPTAC). CPTAC applies proteogenomics to characterize tumors and generate public resources of proteomic and genomic data. It builds on data from The Cancer Genome Atlas (TCGA) by characterizing proteins and genes to understand cancer. CPTAC data, along with clinical and genomic data, can be found on the Genomic Data Commons (GDC) portal. The document provides information on accessing, exploring, and analyzing CPTAC and other proteomic data deposited on the GDC.
CPTAC Data Portal and Proteomics Data Commonsimgcommcall
The CPTAC Data Coordinating Center houses proteomic datasets from CPTAC studies in its public data portal and assay portal. It analyzes data through a common pipeline and enables high-speed access. The Proteomic Data Commons is being developed to provide unified access to mass spectrometry data from multiple sources and allow analysis tools to access data in the cloud. It currently hosts data from CPTAC studies and is working to integrate with other cancer research data clouds. The goal is to improve data sharing, reuse and reproducibility across proteomic studies.
The National Cancer Institute’s Clinical Proteomic Tumor Analysis Consortium (CPTAC) is a national effort to accelerate the understanding of the molecular basis of cancer through the application of large-scale proteome and genome analysis, or proteogenomics.
The PRISM semantic integration approach aims to integrate diverse datasets from The Cancer Imaging Archive by representing data using shared ontologies. This removes obstacles to combining data about the same individuals from different sources. The Arkansas Image Enterprise System (ARIES) is an instance of PRISM that integrates imaging, clinical, and cognitive data from three Parkinson's disease cohorts. Semantic representation allows linking image volumes to cognitive assessments across cohorts. Ongoing work expands data integration and develops semantic query tools.
New manual contours for evaluating lung segmentation algorithms, additional MRI sequences for prostate data, and three new pathology datasets were added to TCIA. Diffusion and dynamic contrast MRI were added to the QIN-Prostate-Repeatability collection. 2018 Crowds Cure Cancer data was also released. The CPTAC lung adenocarcinoma proteomics/clinical discovery cohort data is now available. An upcoming Community Call will discuss the CPTAC program and data access policies. Multiple TCIA datasets were used in an AI cancer detection paper covered by major media outlets. A paper also discussed vulnerabilities in radiomic signature development using the NSCLC dataset. The NBIA data portal was updated with species selection and automatic inclusion of annotation files.
The document summarizes the agenda for an Imaging Community Call on June 3, 2019. The agenda included welcome remarks, updates on the PRISM project and TCIA, and announcements. The announcements section provided information on joining the community call mailing list, the Imaging Community Call Wiki page, and the next scheduled call on July 1, 2019.
This document summarizes the new features and improvements in the NBIA 7.0 GA Community Version release, including a new search interface, support for fielded text search, an improved data retriever application, upgrades for better Java support, and testing of load balancing capabilities. It also provides information on how to obtain the community version release and where to find additional documentation and support resources.
The document summarizes the agenda for an Imaging Community Call on May 6, 2019. The agenda included welcome remarks, updates on ITCR, TCIA, CPTAC SIG, and NBIA 7.0GA, and announcements. Under announcements, it invites the community to future calls, topics for discussion, and links to the community call wiki and SlideShare page for presentation materials. The next scheduled calls were for June 3rd with a PRISM update and July 1st.
The Cancer Imaging Archive provides several updates, including a text search feature to search DICOM metadata fields, filtering data collections by available data types, and new web pages on imaging proteogenomics and clinical trials. A prostate MRI repeatability collection was added and publications were highlighted using TCIA MRI and genetic data to develop models predicting glioma prognosis and treatment response. The NCI Imaging Data Commons RFP was promoted and BraTS test data is now available upon request.
Muktapishti is a traditional Ayurvedic preparation made from Shoditha Mukta (Purified Pearl), is believed to help regulate thyroid function and reduce symptoms of hyperthyroidism due to its cooling and balancing properties. Clinical evidence on its efficacy remains limited, necessitating further research to validate its therapeutic benefits.
TEST BANK For Community Health Nursing A Canadian Perspective, 5th Edition by...Donc Test
TEST BANK For Community Health Nursing A Canadian Perspective, 5th Edition by Stamler, Verified Chapters 1 - 33, Complete Newest Version Community Health Nursing A Canadian Perspective, 5th Edition by Stamler, Verified Chapters 1 - 33, Complete Newest Version Community Health Nursing A Canadian Perspective, 5th Edition by Stamler Community Health Nursing A Canadian Perspective, 5th Edition TEST BANK by Stamler Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Pdf Chapters Download Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Pdf Download Stuvia Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Study Guide Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Ebook Download Stuvia Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Questions and Answers Quizlet Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Studocu Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Quizlet Test Bank For Community Health Nursing A Canadian Perspective, 5th Edition Stuvia Community Health Nursing A Canadian Perspective, 5th Edition Pdf Chapters Download Community Health Nursing A Canadian Perspective, 5th Edition Pdf Download Course Hero Community Health Nursing A Canadian Perspective, 5th Edition Answers Quizlet Community Health Nursing A Canadian Perspective, 5th Edition Ebook Download Course hero Community Health Nursing A Canadian Perspective, 5th Edition Questions and Answers Community Health Nursing A Canadian Perspective, 5th Edition Studocu Community Health Nursing A Canadian Perspective, 5th Edition Quizlet Community Health Nursing A Canadian Perspective, 5th Edition Stuvia Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Pdf Chapters Download Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Pdf Download Stuvia Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Study Guide Questions and Answers Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Ebook Download Stuvia Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Questions Quizlet Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Studocu Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Quizlet Community Health Nursing A Canadian Perspective, 5th Edition Test Bank Stuvia
Local Advanced Lung Cancer: Artificial Intelligence, Synergetics, Complex Sys...Oleg Kshivets
Overall life span (LS) was 1671.7±1721.6 days and cumulative 5YS reached 62.4%, 10 years – 50.4%, 20 years – 44.6%. 94 LCP lived more than 5 years without cancer (LS=2958.6±1723.6 days), 22 – more than 10 years (LS=5571±1841.8 days). 67 LCP died because of LC (LS=471.9±344 days). AT significantly improved 5YS (68% vs. 53.7%) (P=0.028 by log-rank test). Cox modeling displayed that 5YS of LCP significantly depended on: N0-N12, T3-4, blood cell circuit, cell ratio factors (ratio between cancer cells-CC and blood cells subpopulations), LC cell dynamics, recalcification time, heparin tolerance, prothrombin index, protein, AT, procedure type (P=0.000-0.031). Neural networks, genetic algorithm selection and bootstrap simulation revealed relationships between 5YS and N0-12 (rank=1), thrombocytes/CC (rank=2), segmented neutrophils/CC (3), eosinophils/CC (4), erythrocytes/CC (5), healthy cells/CC (6), lymphocytes/CC (7), stick neutrophils/CC (8), leucocytes/CC (9), monocytes/CC (10). Correct prediction of 5YS was 100% by neural networks computing (error=0.000; area under ROC curve=1.0).
8 Surprising Reasons To Meditate 40 Minutes A Day That Can Change Your Life.pptxHolistified Wellness
We’re talking about Vedic Meditation, a form of meditation that has been around for at least 5,000 years. Back then, the people who lived in the Indus Valley, now known as India and Pakistan, practised meditation as a fundamental part of daily life. This knowledge that has given us yoga and Ayurveda, was known as Veda, hence the name Vedic. And though there are some written records, the practice has been passed down verbally from generation to generation.
Osteoporosis - Definition , Evaluation and Management .pdfJim Jacob Roy
Osteoporosis is an increasing cause of morbidity among the elderly.
In this document , a brief outline of osteoporosis is given , including the risk factors of osteoporosis fractures , the indications for testing bone mineral density and the management of osteoporosis
Integrating Ayurveda into Parkinson’s Management: A Holistic ApproachAyurveda ForAll
Explore the benefits of combining Ayurveda with conventional Parkinson's treatments. Learn how a holistic approach can manage symptoms, enhance well-being, and balance body energies. Discover the steps to safely integrate Ayurvedic practices into your Parkinson’s care plan, including expert guidance on diet, herbal remedies, and lifestyle modifications.
Promoting Wellbeing - Applied Social Psychology - Psychology SuperNotesPsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Rasamanikya is a excellent preparation in the field of Rasashastra, it is used in various Kushtha Roga, Shwasa, Vicharchika, Bhagandara, Vatarakta, and Phiranga Roga. In this article Preparation& Comparative analytical profile for both Formulationon i.e Rasamanikya prepared by Kushmanda swarasa & Churnodhaka Shodita Haratala. The study aims to provide insights into the comparative efficacy and analytical aspects of these formulations for enhanced therapeutic outcomes.
One health condition that is becoming more common day by day is diabetes.
According to research conducted by the National Family Health Survey of India, diabetic cases show a projection which might increase to 10.4% by 2030.
Our backs are like superheroes, holding us up and helping us move around. But sometimes, even superheroes can get hurt. That’s where slip discs come in.
8. Origins
I have loads data!
Would be delighted to share it if that
helps finding such a solution.
9. • Semantic segmentation algorithms are increasingly general purpose
o Models are translatable to unseen tasks.
• Algorithmic advances are commonly validated on one or two tasks
o This limits our understanding of the generalisability of the proposed contributions.
• Imaging-based care has many “tasks”
o A model that ”just works” would have a tremendous impact on healthcare.
Challenges – Algorithm Generalisability
10. • The field is missing a benchmark for general purpose algorithmic validation
• The approach should be fully open source and comprehensive
• Benchmark should be testing for a large span of challenges
o Big vs Small data
o Balanced vs Unbalanced labels
o Small vs Large Objects
o Single vs Multi-class labels
o Mono- vs Multi-modal Imaging
Challenges – Open Benchmark
11. • Low Cardinality
o Finding data is hard
o Getting ethics/governance approval is harder
• Constrained License
o Limits publishing rights
o Data reuse
o Cant be used for commercial applications
• Low coverage of anatomical appearances
Challenges – Open Data
12. • Metrics/Statistics
o Maier-Hein et al. “Is the winner really the best?”
o Tried and tested metrics
o Pre-submission Published Ranking
• Open Implementation
o COMIC
o Open Metrics code
• Continuous submission system
Challenges – Best Practices
14. • Registration free data download
• Permissive copyright-license (CC-BY-SA 4.0),
o This allows for data to be shared, distributed and improved upon.
o Only constraints:
- Attribution
- Share-Alike
• All data has been labelled and verified by an expert human rater
o Best effort to mimic the accuracy required for clinical use.
Open Data
15. The aim is to develop an algorithm or learning system that can solve each task,
separately, without human interaction. This can be achieved through the use of a
single learner, an ensemble of multiple learners, architecture search, curriculum
learning, or any other technique, as long as task-specific model parameters are
not human-defined.
Problem Statement
16. Phase 1
1. Data for 7 tasks was released on the 11th of May.
2. Develop Algorithm
3. Train/Test without human interaction
4. Submit the segmentation results by the 5th of August.
Phase 2
1. Release 3 more tasks on the 6th of August.
2. Train their previously developed algorithm, without any software modifications
3. Submit results of the last 3 tasks by the 31st August.
Process - Phase 1
18. • Phase 1:
o Best Method
• Phase 2:
o Best Method
o Runner Up x2
• NVIDIATitanV ($2999)
o 110 teraflops of compute power
• Sponsored by NVIDIA
Award
19. 15:00 - 15:10: Introduction
15:10 - 15:20: Data Description
15:20 - 15:35: Metrics and Statistics Methodology
15:35 - 15:50: NVIDIA
15:50 - 16:00: Phase 1 Results
16:00 - 16:25: Presentation: Top 5 methods (Phase 1)
16:25 - 16:35: Phase 2 Results
16:35 - 16:55: Panel discussion
16:55 - 17:00: Conclusions and closing
Plan for today
22. Target: Liver and tumour
Modality: Portal venous phase CT
Size: 201 3D volumes (131 Training + 70 Testing)
Source: IRCAD Hôpitaux Universitaires
Challenge: Label unbalance with a large (liver) and small
(tumour) target
Data
Liver Tumors
23. Target: Gliomas segmentation necrotic/active tumour and
oedema
Modality: Multimodal multisite MRI data (FLAIR, T1w,
T1gd,T2w)
Size: 750 4D volumes (484 Training + 266 Testing)
Source: BRATS 2016 and 2017 datasets
Challenge: Complex and heterogeneously-located targets
Data
Brain Tumors
24. Target: Hippocampus head and body
Modality: Mono-modal MRI
Size: 394 3D volumes (263 Training + 131 Testing)
Source: Vanderbilt University Medical Center
Challenge: Segmenting two neighbouring small structures
with high precision
Data
Hippocampus
25. Target: Lung and tumours
Modality: CT
Size: 96 3D volumes (64 Training + 32 Testing)
Source: The Cancer Imaging Archive
Challenge: Segmentation of a small target (cancer) in a large
image
Data
Lung Tumors
26. Target: Prostate central gland and peripheral zone
Modality: Multimodal MR (T2, ADC)
Size: 48 4D volumes (32 Training + 16 Testing)
Source: Radboud University, Nijmegen Medical Centre
Challenge: Segmenting two adjoint regions with large inter-
subject variations
Data
Prostate
27. Target: Left Atrium
Modality: Mono-modal MRI
Size: 30 3D volumes (20 Training + 10 Testing)
Source: King’s College London
Challenge: Small training dataset with large variability
Data
Cardiac
28. Target: Pancreas and tumour/cystic lesion
Modality: Portal venous phase CT
Size: 420 3D volumes (282 Training +139 Testing)
Source: Memorial Sloan Kettering Cancer Center
Challenge: Label unbalance with large (background), medium
(pancreas) and small (tumour) structures
Data
Pancreas Tumor/Cystic Lesions
30. Target: Colon tumor
Modality: CT
Size: 190 3D volumes (126 Training + 64 Testing)
Source: Memorial Sloan Kettering Cancer Center
Challenge: Heterogeneous appearance, very hard to annotate
Data
Colon Tumor
31. Target: Hepatic vessels and tumour
Modality: CT
Size: 443 3D volumes (303 Training + 140 Testing)
Source: Memorial Sloan Kettering Cancer Center
Challenge: Tubular small structures next to heterogeneous
tumour
Data
Hepatic Vessels
33. Target: Spleen
Modality: CT
Size: 61 3D volumes (41 Training + 20 Testing)
Source: Memorial Sloan Kettering Cancer Center
Challenge: Large ranging foreground size
Data
Spleen
36. • Do not touch the challenge data! Come up with ranking scheme before the challenge.
• Experiments were performed with 56 segmentation competitions for which per case
data was available
• You can read a lot of the details in [1]
Strategy for generation of ranking scheme
[1] Maier-Hein, et al.:Is the winner really the best? A critical analysis of common research practice in biomedical image analysis competitions, ArXiv:180602051 (2018)
44. • Bootstrapping with case resampling
oRepeat 1000 times:
- Resample all metric values with replacement. Size of the resample must be
equal to the number of test cases.
- Compute a new ranking based on the modified set of results
- Determine the challenge winner
oCompute % of cases where the winner stayed the winner
oCompute % of algorithms that were ranked first in at least 1% of the simulations.
How to measure ranking stability?
45. Metric-based aggregation with mean most robust
Maier-Hein, et al.:Is the winner really the best? A critical analysis of common research practice in biomedical image analysis competitions, ArXiv:180602051 (2018)
Conclusion:
Metric-based aggregration (aggregate then rank) with mean is most
robust ranking scheme from commonly applied ones
46. Done?
• Conclusion?Take metric-based aggregation with mean as standard method
• Some problems:
oMissing value handling statistically suboptimal when using the mean
(because it depends on the “punishing value“)
oArbitrarily small difference in aggregated metric values results in different ranks
oStatistical tests not straightforward to apply because pairwise comparisons may
result in „inconsistent“ rankings
47. Significance Ranking
ci
a1
ti1
mNm
(a1, ti1)
...
...
tiNi
m1(a1, tiNi
)
mNm
(a1, tiNi
)
...aNA
ti1
mNm
(aNA
, ti1)
...
...
tiNi
m1(aNA
, tiNi
)
mNm
(aNA
, tiNi
)
...
...
2. Pairwise comparisons of
algorithms
Wilcoxon signed rank test:
mj(al, tik) - mj(al‘, tik)
3. Significance scoring
sij(al): Number of algorithms
performing significantly
worse than al according to mj
4. Significance ranking
Rank algorithms according
to sij(ax), x = 1,...,NA
=> Highest score: Rank 1
1. Performance
assessment per
case tik
m1(a1, ti1)
m1(aNA
, ti1)
_
m1(aNA
, ti1)
m1(aNA
, ti1)
Proposed
48. • Robustness similar to that of metric-based aggregation
• Significance level of statistical test has minor influence (default 0.05)
• More shared places compared to metric-based aggregation
o1 winner: 73% / 100%
o2 winners: 14% / 0%
o3 winners: 11% / 0%
o4 winners: 2% / 0%
Experiments with significance ranking
49. Thank you!
• News on the challenge initiative will be tweeted:
• @cami_dkfz #BiomedicalChallenges
Further reading: Maier-Hein, et al.:Is the winner really the best? A critical analysis of common research practice in
biomedical image analysis competitions,ArXiv:180602051 (2018)
51. Other metrics considered
• Robust Dice (ignoring border region)
• Volume Difference (RMSE)
• Voxel classification statistics
o True positives
o False positives
o True negatives
o False negatives
Volume-based Performance Metrics
M1
M2
Dice Sørensen Coefficient
52. Surface-based Performance Metrics
Gold Standard
Model
prediction
Standard surface metrics
• Maximum surface distance
( a.k.a Hausdorff)
• 95 percentile of surface distances
(Hausdorff95)
• Mean surface distance
• Median surface distance
53. Surface DSC*
acceptable
deviation τ
Gold Standard
Model
prediction
• Acceptable deviation τ defined by
data set provider
• Green: acceptable surface parts
(distance between surfaces ≤ τ)
• Pink: unacceptable parts of the
surfaces (distance between surfaces >
τ ).
• Surface DSC measures the fraction of
acceptable surface parts
*called Normalised Surface Distance (NSD) in the challenge
54. • Blue: ground truth
• Green: predicted
• Yellow: distances ≤ τ
• Red: distances > τ
• τ = 0.2
• For illustration only
showing distances
from predicted
contour to ground
truth countour
Surface DSC Illustration
Hausdorff: 0.14
Hausdorff95: 0.13
Surface DSC: 100%
Hausdorff: 1.38
Hausdorff95: 0.65
Surface DSC: 88%
Hausdorff: 0.61
Hausdorff95: 0.58
Surface DSC: 60%
55. Surface DSC Computation
total surface
area
0.0 = no overlap
1.0 = perfect overlap
"overlapping"
surface area
Raw mask Surface
Border region
at tolerance τ
Forward and backward
“overlap”
84. Panel Discussion
• Even more varied data?
• Complex vs simple ranking/metrics?
• Freedom vs Dockerization?
• Fairness in Compute Power?
85. Closing Remarks
• Article publication
o Data
o Full Paper
• Re-open submission system
o Monthly ranking release
o Effort to avoid overfitting
o Standard Benchmark for Segmentation
• The future
o We aim to achieve the dodecathlon 2019 (20 tasks)
o RSNA/ACR data validation