Extracting Functional-Connectome Biomarkers with Machine Learning: a talk in the symposium on how do current predictive connectivity models meet clinician’s needs?
This talk is a bit provocative and first sets visions, before bringing a few technical suggestions
Towards psychoinformatics with machine learning and brain imagingGael Varoquaux
Informatics in the psychological sciences brings fascinating challenges as mental processes or pathologies have fuzzy definition and are hard to quantify. Brain imaging brings rich data on the neural substrate of these concepts, yet it is a non trivial link.
The goal of this presentation is to put forward basic ideas of "psychoinformatics", using advanced processing on brain images to quantify better the elements of psychology.
It discusses how machine learning can bridge brain images to behavior: to describe better mental processes involved in brain activity, or to extract biomarkers of pathologies, individual traits, or cognition.
Machine learning for functional connectomesGael Varoquaux
A tutorial on using machine-learning for functional-connectomes, for instance on resting-state fMRI. This is typically useful for population imaging: comparing traits or conditions across subjects.
Estimating Functional Connectomes: Sparsity’s Strength and LimitationsGael Varoquaux
Talk given at the OHBM 2017 education course.
I present the challenges and techniques to estimating meaningful brain functional connectomes from fMRI: why sparsity in inverse covariance leads to models that can interpreted as interactions between regions.
Then I discuss the limitations of sparse estimators and introduce shrinkage as an alternative. Finally, I discuss how to compare multiple functional connectomes.
A presentation on deciding when the scores from two tests, which are from the same CHC domain (e.g., Gwm), and which may have the same narrow CHC classifications, are different enough to warrant clinical interpretation.
How to evaulate the unusualness (base rate) of WJ IV cluster or test score di...Kevin McGrew
The WJ IV provides two primary methods for comparing tests or cluster scores. One is based on a predictive model (the variation and comparison procedures) and the other allows comparisons of SEM confidence bands, which takes into account each measures reliability. A third method for comparing scores, one that takes into account the correlation between compared measures (ability cohesion model) is not provided, but is frequently used by assessment professionals. The three types of score comparison methods are described and new information, via a "rule of thumb" summary slide and nomograph, are provided to allow WJ IV users to evaluate scores via all three methods.
Towards psychoinformatics with machine learning and brain imagingGael Varoquaux
Informatics in the psychological sciences brings fascinating challenges as mental processes or pathologies have fuzzy definition and are hard to quantify. Brain imaging brings rich data on the neural substrate of these concepts, yet it is a non trivial link.
The goal of this presentation is to put forward basic ideas of "psychoinformatics", using advanced processing on brain images to quantify better the elements of psychology.
It discusses how machine learning can bridge brain images to behavior: to describe better mental processes involved in brain activity, or to extract biomarkers of pathologies, individual traits, or cognition.
Machine learning for functional connectomesGael Varoquaux
A tutorial on using machine-learning for functional-connectomes, for instance on resting-state fMRI. This is typically useful for population imaging: comparing traits or conditions across subjects.
Estimating Functional Connectomes: Sparsity’s Strength and LimitationsGael Varoquaux
Talk given at the OHBM 2017 education course.
I present the challenges and techniques to estimating meaningful brain functional connectomes from fMRI: why sparsity in inverse covariance leads to models that can interpreted as interactions between regions.
Then I discuss the limitations of sparse estimators and introduce shrinkage as an alternative. Finally, I discuss how to compare multiple functional connectomes.
A presentation on deciding when the scores from two tests, which are from the same CHC domain (e.g., Gwm), and which may have the same narrow CHC classifications, are different enough to warrant clinical interpretation.
How to evaulate the unusualness (base rate) of WJ IV cluster or test score di...Kevin McGrew
The WJ IV provides two primary methods for comparing tests or cluster scores. One is based on a predictive model (the variation and comparison procedures) and the other allows comparisons of SEM confidence bands, which takes into account each measures reliability. A third method for comparing scores, one that takes into account the correlation between compared measures (ability cohesion model) is not provided, but is frequently used by assessment professionals. The three types of score comparison methods are described and new information, via a "rule of thumb" summary slide and nomograph, are provided to allow WJ IV users to evaluate scores via all three methods.
Measuring mental health with machine learning and brain imagingGael Varoquaux
The study of mental health relies vastly on behavior testing and questionnaires. I discuss how
machine learning on large brain-imaging cohorts can open new alleys for markers of mental health. My
claims are that challenges are the amount of diagnosed conditions rather than heterogeneity of the
conditions and that we should turn to proxy labels. I discuss another fundamental challenge to this
agenda: the external and construct validity of brain-imaging based markers.
Machine learning and cognitive neuroimaging: new tools can answer new questionsGael Varoquaux
Machine learning is geared towards prediction. However, aside diagnosis or prognosis in the clinics, cognitive neuroimaging strives for uncovering insights from the data, rather than minimizing prediction error. I review various inferences on brain function that have been drawn using pattern recognition techniques, focusing on decoding. In particular, I discuss using generalization as a test for information, multivariate analysis to interpret overlapping activation patterns, and decoding for principled reverse inference. I give each time a statistical view and a cognitive imaging view.
This paper studies an identification problem that arises when clinicians seek to personalize patient care by predicting health outcomes conditional on observed patient covariates. Let y be an outcome of interest and let (x = k, w = j) be observed patient covariates. Suppose a clinician wants to choose a care option that maximizes a patient's expected utility conditional on the observed covariates. To accomplish this, the clinician needs to know the conditional probability distribution P(y|x = k, w = j). It is common to have a trustworthy evidence-based risk assessment that predicts y conditional on a subset of the observed covariates, say x, but not conditional on (x, w). Then the clinician knows P(y|x = k) but not P(y|x = k, w = j). Research on the ecological inference problem studies partial identification of P(y∣x, w) given knowledge of P(y|x) and P(w|x). Combining this knowledge with structural assumptions yields tighter conclusions. A psychological literature comparing actuarial predictions and clinical judgments has concluded that clinicians should not attempt to subjectively predict patient outcomes conditional on covariates that are not utilized in evidence-based risk assessments. I argue that formalizing clinical judgment through analysis of the identification problem can improve risk assessments and care decisions.
Running head PROJECT PHASE 4-INFECTIOUS DISEASES1PROJECT PHASE.docxtoltonkendal
Running head: PROJECT PHASE 4-INFECTIOUS DISEASES 1
PROJECT PHASE 4 – INFECTIOUS DISEASES 13
Project Phase 4 – Infectious Diseases
Author Note
This paper is being submitted on
Project Phase 4 – Infectious Diseases
Introduction:
As a healthcare professional, you will work to improve and maintain the health of individuals, families, and communities in various settings. Basic statistical analysis can be used to gain an understanding of current problems. Understanding the current situation is the first step in discovering where an opportunity for improvement exists. This course project will assist you in applying basic statistical principles to a fictional scenario in order to impact the health and wellbeing of the clients being served.
This assignment will be completed in phases throughout the quarter. As you gain additional knowledge through the didactic portion of this course, you will be able to apply your new knowledge to this project. You will receive formative feedback from your instructor on each submission. The final project will be due on week 5.
Scenario:
You are currently working at NCLEX Memorial Hospital in the Infectious Diseases Unit. Over the past few days, you have noticed an increase in patients admitted with a particular infectious disease. You believe that the ages of these patients play a critical role in the method used to treat the patients. You decide to speak to your manager and together you work to use statistical analysis to look more closely at the ages of these patients. You do some research and put together a spreadsheet of the data that contains the following information:
· Client number
· Infection Disease Status
· Age of the patient
You need the preliminary findings immediately so that you can start treating these patients. So, let’s get to work!!!!
Background information on the Data:
The data set consists of 60 patients that have the infectious disease with ages ranging from 35 years of age to 76 years of age for NCLEX Memorial Hospital. Remember this assignment will be completed over the duration of the course.
To begin let’s learn what infectious disease is. Infectious diseases are caused by pathogenic microorganisms, which are bacteria, viruses, parasites or fungi; the diseases can be spread directly or indirectly, through one person to another (WHO, 2017).
This scenario will aim to improve the quality of healthcare services that are provided to individuals, families, and communities at different levels of age. Therefore, the project utilized at NCLEX Memorial Hospital, over the past few days has seen a larger level of infectious disease occurrences. The data set composed was for sixty patients ranging in age from thirty-five to seventy-six.
1)
a) Qualitative infectious: Disease
b) Quantitative: Age
2) Age is a constant variable as it may take any value.
3) A variable is any quantity that can be measured and whose value differs through the
Population and here we se ...
Bayesain Hypothesis of Selective Attention - Raw 2011 posterGiacomo Veneri
The aim of the study is to understand the process of target averaging during the selection process. We analyzed the probability to select the target after a fixation outside ROIs from the duration of fixations and the distance to the target. We aimed to respond to the question “is it possible to predict the selected area?” . In this study we tested the presence of information in non-ROI fixation data about the occurrence of a target at the next saccade. A classification algorithm was trained to predict the target vs. non-target outcome (dependent variable) of a saccade from summary statistics of fixation data (covariates). We claim that significantly accurate predictions are substantial evidence to support the hypothesis of "presence of information".
Success stories of the Big Data paradigm and Predictive Analytics in many application areas led to the wide recognition of their high potential impact application areas like healthcare, marketing, finance etc. However, there is still a large gap between actual and potential data usage, because of numerous challenges: high dimensionality, sparsity, data heterogeneity, privacy concerns, the need for collaboration between domain experts and data scientists, demand for highly accurate and interpretable models etc. On the other side, extensive efforts of scientific research offer many partial or complete solutions for the aforementioned challenges. Coordination of research and industry efforts (fusion of cutting edge predictive analytics methodologies with commercial or non-commercial products) should lead to increased exploitation of Big Data promise, better satisfaction of industry needs and new methodological breakthroughs.
Evaluating machine learning models and their diagnostic valueGael Varoquaux
Model evaluation is, in my opinion, the most overlooked step of the machine-learning pipeline. Reliably estimating a model's performance for a given purpose is crucial and difficult. In this talk, I first discuss choosing metric informative for the application, stressing the importance of the class prevalence in classification settings. I will then discussing procedures to estimate the generalization performance, drawing a distinction between evaluating a learning procedure or a prediction rule, and discussing how to give confidence intervals to the performance estimates.
A tutorial on machine learning to build prediction models with missing values.
The slides cover both theoretical results (statistical learning) and practical advice, with a focus on implementation in Python with scikit-learn
More Related Content
Similar to Functional-connectome biomarkers to meet clinical needs?
Measuring mental health with machine learning and brain imagingGael Varoquaux
The study of mental health relies vastly on behavior testing and questionnaires. I discuss how
machine learning on large brain-imaging cohorts can open new alleys for markers of mental health. My
claims are that challenges are the amount of diagnosed conditions rather than heterogeneity of the
conditions and that we should turn to proxy labels. I discuss another fundamental challenge to this
agenda: the external and construct validity of brain-imaging based markers.
Machine learning and cognitive neuroimaging: new tools can answer new questionsGael Varoquaux
Machine learning is geared towards prediction. However, aside diagnosis or prognosis in the clinics, cognitive neuroimaging strives for uncovering insights from the data, rather than minimizing prediction error. I review various inferences on brain function that have been drawn using pattern recognition techniques, focusing on decoding. In particular, I discuss using generalization as a test for information, multivariate analysis to interpret overlapping activation patterns, and decoding for principled reverse inference. I give each time a statistical view and a cognitive imaging view.
This paper studies an identification problem that arises when clinicians seek to personalize patient care by predicting health outcomes conditional on observed patient covariates. Let y be an outcome of interest and let (x = k, w = j) be observed patient covariates. Suppose a clinician wants to choose a care option that maximizes a patient's expected utility conditional on the observed covariates. To accomplish this, the clinician needs to know the conditional probability distribution P(y|x = k, w = j). It is common to have a trustworthy evidence-based risk assessment that predicts y conditional on a subset of the observed covariates, say x, but not conditional on (x, w). Then the clinician knows P(y|x = k) but not P(y|x = k, w = j). Research on the ecological inference problem studies partial identification of P(y∣x, w) given knowledge of P(y|x) and P(w|x). Combining this knowledge with structural assumptions yields tighter conclusions. A psychological literature comparing actuarial predictions and clinical judgments has concluded that clinicians should not attempt to subjectively predict patient outcomes conditional on covariates that are not utilized in evidence-based risk assessments. I argue that formalizing clinical judgment through analysis of the identification problem can improve risk assessments and care decisions.
Running head PROJECT PHASE 4-INFECTIOUS DISEASES1PROJECT PHASE.docxtoltonkendal
Running head: PROJECT PHASE 4-INFECTIOUS DISEASES 1
PROJECT PHASE 4 – INFECTIOUS DISEASES 13
Project Phase 4 – Infectious Diseases
Author Note
This paper is being submitted on
Project Phase 4 – Infectious Diseases
Introduction:
As a healthcare professional, you will work to improve and maintain the health of individuals, families, and communities in various settings. Basic statistical analysis can be used to gain an understanding of current problems. Understanding the current situation is the first step in discovering where an opportunity for improvement exists. This course project will assist you in applying basic statistical principles to a fictional scenario in order to impact the health and wellbeing of the clients being served.
This assignment will be completed in phases throughout the quarter. As you gain additional knowledge through the didactic portion of this course, you will be able to apply your new knowledge to this project. You will receive formative feedback from your instructor on each submission. The final project will be due on week 5.
Scenario:
You are currently working at NCLEX Memorial Hospital in the Infectious Diseases Unit. Over the past few days, you have noticed an increase in patients admitted with a particular infectious disease. You believe that the ages of these patients play a critical role in the method used to treat the patients. You decide to speak to your manager and together you work to use statistical analysis to look more closely at the ages of these patients. You do some research and put together a spreadsheet of the data that contains the following information:
· Client number
· Infection Disease Status
· Age of the patient
You need the preliminary findings immediately so that you can start treating these patients. So, let’s get to work!!!!
Background information on the Data:
The data set consists of 60 patients that have the infectious disease with ages ranging from 35 years of age to 76 years of age for NCLEX Memorial Hospital. Remember this assignment will be completed over the duration of the course.
To begin let’s learn what infectious disease is. Infectious diseases are caused by pathogenic microorganisms, which are bacteria, viruses, parasites or fungi; the diseases can be spread directly or indirectly, through one person to another (WHO, 2017).
This scenario will aim to improve the quality of healthcare services that are provided to individuals, families, and communities at different levels of age. Therefore, the project utilized at NCLEX Memorial Hospital, over the past few days has seen a larger level of infectious disease occurrences. The data set composed was for sixty patients ranging in age from thirty-five to seventy-six.
1)
a) Qualitative infectious: Disease
b) Quantitative: Age
2) Age is a constant variable as it may take any value.
3) A variable is any quantity that can be measured and whose value differs through the
Population and here we se ...
Bayesain Hypothesis of Selective Attention - Raw 2011 posterGiacomo Veneri
The aim of the study is to understand the process of target averaging during the selection process. We analyzed the probability to select the target after a fixation outside ROIs from the duration of fixations and the distance to the target. We aimed to respond to the question “is it possible to predict the selected area?” . In this study we tested the presence of information in non-ROI fixation data about the occurrence of a target at the next saccade. A classification algorithm was trained to predict the target vs. non-target outcome (dependent variable) of a saccade from summary statistics of fixation data (covariates). We claim that significantly accurate predictions are substantial evidence to support the hypothesis of "presence of information".
Success stories of the Big Data paradigm and Predictive Analytics in many application areas led to the wide recognition of their high potential impact application areas like healthcare, marketing, finance etc. However, there is still a large gap between actual and potential data usage, because of numerous challenges: high dimensionality, sparsity, data heterogeneity, privacy concerns, the need for collaboration between domain experts and data scientists, demand for highly accurate and interpretable models etc. On the other side, extensive efforts of scientific research offer many partial or complete solutions for the aforementioned challenges. Coordination of research and industry efforts (fusion of cutting edge predictive analytics methodologies with commercial or non-commercial products) should lead to increased exploitation of Big Data promise, better satisfaction of industry needs and new methodological breakthroughs.
Similar to Functional-connectome biomarkers to meet clinical needs? (20)
Evaluating machine learning models and their diagnostic valueGael Varoquaux
Model evaluation is, in my opinion, the most overlooked step of the machine-learning pipeline. Reliably estimating a model's performance for a given purpose is crucial and difficult. In this talk, I first discuss choosing metric informative for the application, stressing the importance of the class prevalence in classification settings. I will then discussing procedures to estimate the generalization performance, drawing a distinction between evaluating a learning procedure or a prediction rule, and discussing how to give confidence intervals to the performance estimates.
A tutorial on machine learning to build prediction models with missing values.
The slides cover both theoretical results (statistical learning) and practical advice, with a focus on implementation in Python with scikit-learn
Dirty data science machine learning on non-curated dataGael Varoquaux
These slides are a one-hour course on machine learning with non-curated data.
According to industry surveys, the number one hassle of data scientists is cleaning the data to analyze it. Here, I survey what "dirtyness" forces time-consuming cleaning. We will then cover two specific aspects of dirty data: non-normalized entries and missing values. I show how, for these two problems, machine-learning practice can be adapted to work directly on a data table without curation. The normalization problem can be tackled by adapting methods from natural language processing. The missing-values problem will lead us to revisit classic statistical results in the setting of supervised learning.
Representation learning in limited-data settingsGael Varoquaux
A 4-hour long didactic course on simple notions of representations and how to use them in limited-data settings:
- A supervised learning point of view, giving intuitions and math on what are representations are why they matter
- Building simple unsupervised learning models to extract representation: from matrix decomposition for signals to embeddings of entities
- Evaluating models in limited-data settings, often a bottleneck
This slide-deck was given as a course at the 2021 DeepLearn summer school.
Better neuroimaging data processing: driven by evidence, open communities, an...Gael Varoquaux
My current thoughts about methods validity and design in brain imaging.
Data processing is a significant part of a neuroimaging study. The choice of corresponding methods and tools is crucial. I will give an opinionated view how on a path to building better data processing for neuroimaging. I will take examples on endeavors that I contributed to: defining standards for functional-connectivity analysis, the nilearn neuroimaging tool, the scikit-learn machine-learning toolbox -an industry standard with a million regular users. I will cover not only the technical process -statistics, signal processing, software engineering- but also the epistemology of methods development. Methods govern our results, they are more than a technical detail.
Atlases of cognition with large-scale human brain mappingGael Varoquaux
Cognitive neuroscience uses neuroimaging to identify brain systems engaged in specific cognitive tasks. However, linking unequivocally brain systems with cognitive functions is difficult: each task probes only a small number of facets of cognition, while brain systems are often engaged in many tasks. We develop a new approach to generate a functional atlas of cognition, demonstrating brain systems selectively associated with specific cognitive functions. This approach relies upon an ontology that defines specific cognitive functions and the relations between them, along with an analysis scheme tailored to this ontology. Using a database of thirty neuroimaging studies, we show that this approach provides a highly-specific atlas of mental functions, and that it can decode the mental processes engaged in new tasks.
Similarity encoding for learning on dirty categorical variablesGael Varoquaux
For statistical learning, categorical variables in a table are usually considered as discrete entities and encoded separately to feature vectors, e.g., with one-hot encoding. "Dirty" non-curated data gives rise to categorical variables with a very high cardinality but redundancy: several categories reflect the same entity. In databases, this issue is typically solved with a deduplication step. We show that a simple approach that exposes the redundancy to the learning algorithm brings significant gains. We study a generalization of one-hot encoding, similarity encoding, that builds feature vectors from similarities across categories. We perform a thorough empirical validation on non-curated tables, a problem seldom studied in machine learning. Results on seven real-world datasets show that similarity encoding brings significant gains in prediction in comparison with known encoding methods for categories or strings, notably one-hot encoding and bag of character n-grams. We draw practical recommendations for encoding dirty categories: 3-gram similarity appears to be a good choice to capture morphological resemblance. For very high-cardinality, dimensionality reduction significantly reduces the computational cost with little loss in performance: random projections or choosing a subset of prototype categories still outperforms classic encoding approaches.
Simple representations for learning: factorizations and similarities Gael Varoquaux
Real-life data seldom comes in the ideal form for statistical learning.
This talk focuses on high-dimensional problems for signals and
discrete entities: when dealing with many, correlated, signals or
entities, it is useful to extract representations that capture these
correlations.
Matrix factorization models provide simple but powerful representations. They are used for recommender systems across discrete entities such as users and products, or to learn good dictionaries to represent images. However they entail large computing costs on very high-dimensional data, databases with many products or high-resolution images. I will present an
algorithm to factorize huge matrices based on stochastic subsampling that gives up to 10-fold speed-ups [1].
With discrete entities, the explosion of dimensionality may be due to variations in how a smaller number of categories are represented. Such a problem of "dirty categories" is typical of uncurated data sources. I will discuss how encoding this data based on similarities recovers a useful category structure with no preprocessing. I will show how it interpolates between one-hot encoding and techniques used in character-level natural language processing.
[1] Stochastic subsampling for factorizing huge matrices, A Mensch, J Mairal, B Thirion, G Varoquaux, IEEE Transactions on Signal Processing 66 (1), 113-128
[2] Similarity encoding for learning with dirty categorical variables. P Cerda, G Varoquaux, B Kégl Machine Learning (2018): 1-18
A tutorial on Machine Learning, with illustrations for MR imagingGael Varoquaux
Machine learning builds predictive models from the data. It is massive used on medical images these days, for a variety of applications ranging from segmentation to diagnosis.
This is an introductory tutorial to machine learning from giving intuitions on the statistical point of view. It introduce the methodology, the concepts behind the central models, the validation framework and some caveats to look for.
It also discusses some applications to drawing conclusions from brain imaging, and use these applications to highlight various technical aspects to running machine learning models on high-dimensional data such as medical imaging.
Scikit-learn and nilearn: Democratisation of machine learning for brain imagingGael Varoquaux
This talk describe our efforts to bring easily usable machine learning to brain mapping. It covers both questions that machine learning can answer as well as two softwares developed to facilitate machine learning and it's application to neuroimaging.
Computational practices for reproducible scienceGael Varoquaux
Reconciling bleeding-edge scientific results and reproducible research may seem a conundrum in our fast-paced high-pressure academic world. I discuss the practices that I found useful in computational work. At a high level, it is important to navigate the space between rapid experimentation and industrial-grade software development. I advocate adopting more and more software-engineering best practices as a project matures. I will also discuss how to turn the computational work into libraries, and to ensure the quality of the resulting libraries. And I conclude on how those libraries need to fit in the larger picture of the exercise of research to give better science.
Slides for my keynote at Scipy 2017
https://youtu.be/eVDDL6tgsv8
Computing has been driving forward a revolution in how science and technology can solve new problems. Python has grown to be a central player in this game, from computational physics to data science. I would like to explore some lessons learned doing science with Python as well as doing Python libraries for science. What are the ingredients that the scientists need? What technical and project-management choices drove the success of projects I've been involved with? How do these demands and offers shape our ecosystem?
In this talk, I'd like to share a few thoughts on how we code for science and innovation, with the modest goal of changing the world.
Data science calls for rapid experimentation and building intuitions from the data. Yet, data science also underpins crucial decisions and operational logic. Writing production-ready and robust statistical analysis without cognitive overhead may seem a conundrum. I will explore simple, and less simple, practices for fast turn around and consolidation of data-science code. I will discuss how these considerations led to the design of scikit-learn, that enables easy machine learning yet is used in production. Finally, I will mention some scikit-learn gems, new or forgotten.
Scientist meets web dev: how Python became the language of dataGael Varoquaux
Python started as a scripting language, but now it is the new trend everywhere and in particular for data science, the latest rage of computing. It didn’t get there by chance: tools and concepts built by nerdy scientists and geek sysadmins provide foundations for what is said to be the sexiest job: data scientist.
In this talk I give a personal perspective on the progress of the scientific Python ecosystem, from numerical physics to data mining. What made Python suitable for science; Why the cultural gap between scientific Python and the broader Python community turned out to be a gold mine; And where this richness might lead us.
The talk will discuss low-level and high-level technical aspects, such as how the Python world makes it easy to move large chunks of number across code. It will touch upon current technical details that make scikit-learn and joblib stand.
Talk giving at PRNI 2016 for the paper https://arxiv.org/pdf/1606.06439v1.pdf
Abstract — Spatially-sparse predictors are good models for
brain decoding: they give accurate predictions and their weight
maps are interpretable as they focus on a small number of
regions. However, the state of the art, based on total variation or
graph-net, is computationally costly. Here we introduce sparsity
in the local neighborhood of each voxel with social-sparsity, a
structured shrinkage operator. We find that, on brain imaging
classification problems, social-sparsity performs almost as well as
total-variation models and better than graph-net, for a fraction
of the computational cost. It also very clearly outlines predictive
regions. We give details of the model and the algorithm
Personal point of view on scikit-learn: past, present, and future.
This talks gives a bit of history, mentions exciting development, and a personal vision on the future.
Inter-site autism biomarkers from resting state fMRIGael Varoquaux
We present an automated pipeline to learn predictive biomarkers from resting-state fMRI. We apply it to classifying autism on unseen sites, demonstrating the feasibility of biomarkers on weakly standardized functional imaging data.
We study the steps of the pipeline that are important to predict and can show that 1) the choice of atlas is the most important choice. Ideally the atlas should be made of functional regions learned from the data. 2) "tangent space" parametrization of the connectivity is the best performer.
We conclude on general recommendations for predictive biomarkers from resting-state fMRI
Brain maps from machine learning? Spatial regularizationsGael Varoquaux
Pattern Recognition for NeuroImaging (PR4NI)
We will show empirically how the pattern recognition techniques-commonly used, such as SVMs, provide low-quality brain maps, eventhough they give very good prediction accuracy. We will give an overview of recently developed techniques to impose priors on patterns particularly well suited to neuroimaging: selecting a small number of spatially-structured predictive brain regions. These tools reconcile machine learning with
brain mapping by giving maps more useful to draw neuroscientific conclusions. In addition, they are more robust to cross-individuals spatial variability and thus generalize well across subjects.
Scikit-learn for easy machine learning: the vision, the tool, and the projectGael Varoquaux
Scikit-learn is a popular machine learning tool. What can it do for you?Why you you want to use it? What can you do with it? Where is it going?In this talk, I will discuss why and how scikit-learn became popular. Iwill argue that it is successful because of its vision: it fills an important slot in the rich ecosystem of data science. I will demonstrate how scikit-learn makes predictive analysis easy and yet versatile.I will shed some light on our development process: how do we, as a community, ensure the quality and the growth of scikit-learn?
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
zkStudyClub - Reef: Fast Succinct Non-Interactive Zero-Knowledge Regex ProofsAlex Pruden
This paper presents Reef, a system for generating publicly verifiable succinct non-interactive zero-knowledge proofs that a committed document matches or does not match a regular expression. We describe applications such as proving the strength of passwords, the provenance of email despite redactions, the validity of oblivious DNS queries, and the existence of mutations in DNA. Reef supports the Perl Compatible Regular Expression syntax, including wildcards, alternation, ranges, capture groups, Kleene star, negations, and lookarounds. Reef introduces a new type of automata, Skipping Alternating Finite Automata (SAFA), that skips irrelevant parts of a document when producing proofs without undermining soundness, and instantiates SAFA with a lookup argument. Our experimental evaluation confirms that Reef can generate proofs for documents with 32M characters; the proofs are small and cheap to verify (under a second).
Paper: https://eprint.iacr.org/2023/1886
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
2. Extracting Functional-Connectome Biomarkers
with Machine Learning
Ga¨el Varoquaux
How Do Current Predictive Connectivity Models
Meet Clinician’s Needs?
This house believes that predictive biomarkers are, today,
the most useful endeavor for clinical application of func-
tional connectivity
3. Extracting Functional-Connectome Biomarkers
with Machine Learning
Ga¨el Varoquaux
How Do Current Predictive Connectivity Models
Meet Clinician’s Needs?
This house believes that predictive biomarkers are, today,
the most useful endeavor for clinical application of func-
tional connectivity
They are just not reliable enough
7. What if I told youWhat if I told you
Brain imaging predicts the risk that a 2 year-oldBrain imaging predicts the risk that a 2 year-old
develops on the autism spectrumdevelops on the autism spectrum
Brain imaging predicts long-term cognitive deficitBrain imaging predicts long-term cognitive deficit
after strokeafter stroke
G Varoquaux 5
8. 1 Heterogeneity is a roadblock?
[Abraham... 2017]
Autism:
ill-defined diagnostic criteria
sensitive to parents’ social-economic status
ABIDE:
post-hoc aggregation of data
across many cities and countries
Can autism biomarkers carry over to new sites?
Training set Testing set
G Varoquaux 6
9. 1 Heterogeneity is a roadblock?
[Abraham... 2017]
Autism:
ill-defined diagnostic criteria
sensitive to parents’ social-economic status
ABIDE:
post-hoc aggregation of data
across many cities and countries
Can autism biomarkers carry over to new sites?
Training set Testing set
Accuracy
Fraction of subjects used
Prediction to new sites works as well
G Varoquaux 6
10. Yes we can
extract biomarkersextract biomarkers
despite heterogeneitydespite heterogeneity
Multi-variate predictive models, unlike
classical statistics, can learn to reject
confounds, given examples of confound-
ing heterogeneity
G Varoquaux 7
11. 1 Proxy clinical outcomes
[Liem... 2016]
Predicting brain aging = chronological age
Predicts age with a mean absolute error of 4.3 years
G Varoquaux 8
12. 1 Proxy clinical outcomes
[Liem... 2016]
Predicting brain aging = chronological age
Predicts age with a mean absolute error of 4.3 years
Discrepency with chronological age
correlates with cognitive impairment
0 2 4
Brain aging discrepancy
(years)
-0.38
0.74
1.72
Objective Cognitive
Impairment group
Normal
Mild
Major
Biomarker
surrogate,
but useful
G Varoquaux 8
13. 1 Better descriptions of subjects?
[Rahim... 2017]
An individual should not be reduced to
a single diagnostic or behavioral quantity
G Varoquaux 9
14. 1 Better descriptions of subjects?
[Rahim... 2017]
Multi-output prediction
Predict jointly multiple individual phenotypes
behavioral scores diagnostic status
They improve eachother’s prediction
Adding MMSE as a target improves AD prediction
Functional
connectivity (fMRI)
Protein
biomarkers (CSF)
Hippocampus
volumetry (MRI)
50% 60% 70% 80% 90%
Cross-validation accuracy
Stacked predictions
of fMRI, MRI, CSF
mono-
modal
multi-
modal
Classification: AD vs. MCI
Single-output
Multi-output
G Varoquaux 9
16. 1 Trustworthy biomarkers
[Woo... 2017]
Good biomarkers generalize to new subjects
to new sites
Bad biomarkers overly adapt
to a few subjects
to site observation noise
Predictive modeling: machine learning
Prediction rather than association
out-of-sample statistics
G Varoquaux 10
17. One does not simplyOne does not simply
claim predictionclaim prediction
G Varoquaux 11
18. 1 Prediction requires more than association
[R. Poldrack, G. Huckins, G. Varoquaux, submitted]
2 1 0 1 2
5.0
7.5
10.0
12.5
15.0
17.5
20.0
22.5
25.0
order = 1
0
0
20
40
60
80
100
Meansquarederror
G Varoquaux 12
19. 1 Prediction requires more than association
[R. Poldrack, G. Huckins, G. Varoquaux, submitted]
2 1 0 1 2
5.0
7.5
10.0
12.5
15.0
17.5
20.0
22.5
25.0
order = 1
order = 2
0
0
20
40
60
80
100
Meansquarederror
G Varoquaux 12
20. 1 Prediction requires more than association
[R. Poldrack, G. Huckins, G. Varoquaux, submitted]
2 1 0 1 2
5.0
7.5
10.0
12.5
15.0
17.5
20.0
22.5
25.0
order = 1
order = 2
order = 15
0
0
20
40
60
80
100
Meansquarederror
Quality of fit on data used to fit is not meaningful
Only new (test) data, can measure prediction
G Varoquaux 12
21. 1 Evidence for prediction
[Varoquaux... 2017]
Established by cross-validation
Test setTrain set
Full data
G Varoquaux 13
22. [R. Poldrack, G. Huckins, G. Varoquaux, submitted]
One does not simplyOne does not simply
claim predictionclaim prediction
100 last publications on
“fMRI prediction”
0 20 40
None
K-fold
Leave-one-out
Leave-X-out
Other
G Varoquaux 14
24. 1 Cross-validation is solid evidence?
[Varoquaux 2017]
Trivial analytic variations on a permuted data:
smoothing, SVM vs log-reg, feature selection
30% 40% 50% 60% 70%
Crossvalidation scores for different decoders
4 first
4 last
6 first
6 last
all 12
Sessions used
25% 39%
40% 71%
38% 57%
47% 57%
44% 52%
n~72
n~72
n~108
n~108
n~216
G Varoquaux 15
25. 1 Cross-validation is solid evidence?
[Varoquaux 2017]
Trivial analytic variations on a permuted data:
smoothing, SVM vs log-reg, feature selection
30% 40% 50% 60% 70%
Crossvalidation scores for different decoders
4 first
4 last
6 first
6 last
all 12
Sessions used
25% 39%
40% 71%
38% 57%
47% 57%
44% 52%
n~72
n~72
n~108
n~108
n~216
With small n, by chance, some analytic
choices give seemingly good predictions
G Varoquaux 15
26. 1 Cross-validation is solid evidence?
[Varoquaux 2017]
30
100
200
300
umber of available samples 19% +15%
20% +18%
10% +8%
10% +10%
7% +5%
7% +7%
5% +4%
6% +6%
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
Sampling distribution of test error for n = 30
G Varoquaux 15
27. 1 Cross-validation is solid evidence?
[Varoquaux 2017]
30
100
200
300
1000
Number of available samples
19% +15%
20% +18%
10% +8%
10% +10%
7% +5%
7% +7%
5% +4%
6% +6%
3% +2%
3% +3%
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
LOO
50 splits, 20% test
G Varoquaux 15
28. 1 Cross-validation is solid evidence?
[Varoquaux 2017]
45% 30% 15% 0% +15% +30%
Difference between public and private scores
15% +14%
Kaggle competition on r-fMRI for Schizophrenia
2 different test sets: size 30 and 28
G Varoquaux 15
29. One does not simplyOne does not simply
claim predictionclaim prediction
We need
Clean cross-validation
strong-generalization
= testing on data never seen
Several 100s subjects
G Varoquaux 16
30. Yes we can
Reliable prediction of clinical end-
points would be game changing
But we need larger sizes, reduced
analytical variability, and clean
validation
G Varoquaux 17
31. 2 Extracting biomarkers from
rest fMRI
Addressing the perils of
analytical variabality
Systematic study:
6 different cohorts
More than 2000 individuals
[Dadi... 2019]
G Varoquaux 18
32. From rest-fMRI to biomarkers
No salient features in rest fMRI
G Varoquaux 19
33. From rest-fMRI to biomarkers
Define functional regions
G Varoquaux 19
34. From rest-fMRI to biomarkers
Define functional regions
Learn interactions
G Varoquaux 19
35. From rest-fMRI to biomarkers
Define functional regions
Learn interactions
Detect differences
G Varoquaux 19
36. From rest-fMRI to biomarkers
Functional
connectivity
matrix
Time series
extraction
Region
definition
Supervised learning
RS-fMRI
G Varoquaux 20
37. 2 Defining regions
Anatomical atlases
Clustering
k-means
ward
[Thirion... 2014]
... ... ...
... ...
G Varoquaux 21
38. 2 Defining regions
Anatomical atlases
Clustering
k-means
ward
[Thirion... 2014]
Decomposition models
time
voxels
time
voxels
time voxels
Y +E · S=
25
N
G Varoquaux 21
39. 2 Defining regions
Anatomical atlases
Clustering
k-means
ward
[Thirion... 2014]
Decomposition models
ICA:
seek independence of maps
Sparse dictionary learning:
seek sparse maps
G Varoquaux 21
40. 2 In connectome prediction settings
RS-fMRI
Functional
connectivity
Time series
2
4
3
1
Diagnosis
ROIs
Choice of regions for best prediction?
G Varoquaux 22
41. 2 In connectome prediction settings
RS-fMRI
Functional
connectivity
Time series
2
4
3
1
Diagnosis
ROIs
Choice of regions for best prediction?
[Dadi... 2019]
G Varoquaux 22
42. 2 Connectome: building a connectivity matrix
How to capture and represent interactions?
G Varoquaux 23
46. 2 Connectivity matrix for predictive models
Time series
2
RS-fMRI
41
Diagnosis
ROIs Functional
connectivity
3
[Dadi... 2019]
G Varoquaux 26
47. 2 Machine learning for connectome prediction
Functional
connectivity
Time series
3
4
Diagnosis
2
RS-fMRI
1 ROIs
Supervised learning step
Linear models
Random forests
Sparse or non sparse?
G Varoquaux 27
48. 2 Machine learning for connectome prediction
Functional
connectivity
Time series
3
4
Diagnosis
2
RS-fMRI
1 ROIs
Supervised learning step
Linear models
Random forests
Sparse or non sparse? [Dadi... 2019]
G Varoquaux 27
50. @GaelVaroquaux
Functionnal-connectome biomarkers
Biomarkers game-changing if trustworthy
Rest-fMRI biomarkers extraction
Functional regions (extracted by dictionary learning)
Tangent space to compare connectomes
Linear model for supervised learning
RS-fMRI
Diagnosis
Connectivity
Parameterization
Supervised
Learning
Defining Brain
ROIs
1 2 3
Software: nilearn ni
51. References I
A. Abraham, M. Milham, A. Di Martino, R. C. Craddock,
D. Samaras, B. Thirion, and G. Varoquaux. Deriving
reproducible biomarkers from multi-site resting-state data:
An autism-based example. NeuroImage, 2017.
K. Dadi, M. Rahim, A. Abraham, D. Chyzhyk, M. Milham,
B. Thirion, G. Varoquaux, and A. D. N. Initiative.
Benchmarking functional connectome-based predictive
models for resting-state fmri. NeuroImage, 2019.
F. Liem, G. Varoquaux, J. Kynast, F. Beyer, S. K. Masouleh,
J. M. Huntenburg, L. Lampe, M. Rahim, A. Abraham, R. C.
Craddock, ... Predicting brain-age from multimodal imaging
data captures cognitive impairment. NeuroImage, 2016.
M. Rahim, B. Thirion, D. Bzdok, I. Buvat, and G. Varoquaux.
Joint prediction of multiple scores captures better individual
traits from brain images. Neuroimage, in rev, 2017.
52. References II
B. Thirion, G. Varoquaux, E. Dohmatob, and J. Poline. Which
fMRI clustering gives good brain parcellations? Name:
Frontiers in Neuroscience, 8:167, 2014.
G. Varoquaux. Cross-validation failure: small sample sizes lead
to large error bars. NeuroImage, 2017.
G. Varoquaux, F. Baronnet, A. Kleinschmidt, P. Fillard, and
B. Thirion. Detection of brain functional-connectivity
difference in post-stroke patients using group-level
covariance modeling. In MICCAI. 2010.
G. Varoquaux, P. R. Raamana, D. A. Engemann,
A. Hoyos-Idrobo, Y. Schwartz, and B. Thirion. Assessing
and tuning brain decoders: cross-validation, caveats, and
guidelines. NeuroImage, 145:166–179, 2017.
C.-W. Woo, L. J. Chang, M. A. Lindquist, and T. D. Wager.
Building better biomarkers: brain models in translational
neuroimaging. Nature neuroscience, 20(3):365, 2017.