The document discusses computational models in systems biology and improving their reproducibility and reuse. It describes how models evolve over time through various versions as corrections and improvements are made. This can include changes to the structure and equations of the models. The author presents their work developing methods to detect, characterize, and communicate the differences between versions of computational models. This includes an algorithm to identify changes, formats to report changes, and an ontology called COMODI to semantically describe change types. The goal is to improve understanding of a model's evolution and history to increase trust and reuse of models.
The document discusses the evolution of computational models over time. It provides the example of models of the cell cycle control network in the bacterium Clostridium acetobutylicum that have been developed and expanded on since 1984. The models have become more detailed over time, incorporating new experimental findings and insights. The document also discusses how models of the core cell cycle oscillator have evolved since 1991 with the addition of new molecular components and regulatory interactions based on studies in different organisms.
M2CAT: Extracting reproducible simulation studies from model repositories usi...Martin Scharm
Martin Scharm presents M2CAT, a workflow to extract reproducible simulation studies from model repositories. It searches the graph database Masymos for relevant models, simulations, and related data. M2CAT then uses the CombineArchive Toolkit to export the selected studies as COMBINE archive files, which provide a standard format for sharing complete simulation experiments. These archives can be explored and modified using the CombineArchive Web interface or other tools. The goal is to make the large amounts of data in repositories more usable and enable researchers to more easily reproduce and build upon existing computational studies.
This document provides an overview of standards and best practices for making computational models reusable through the use of model repositories and standard formats. It discusses the COMBINE initiative for standardizing the encoding of models and simulations. The document encourages authors to make their models and data FAIR (Findable, Accessible, Interoperable, Reusable) by using community standards for publishing, exchanging, and archiving models. Examples of open model repositories and standards-compliant tools and libraries are provided to demonstrate how authors can improve sharing and reuse of their models.
Applications of Computer Science in Environmental ModelsIJLT EMAS
Computation is now regarded as an equal and
indispensable partner, along with theory and experiment, in the
advance of scientific knowledge and engineering practice.
Numerical simulation enables the study of complex systems and
natural phenomena that would be too expensive or dangerous, or
even impossible, to study by direct experimentation. The quest
for ever higher levels of detail and realism in such simulations
requires enormous computational capacity, and has provided the
impetus for dramatic breakthroughs in computer algorithms and
architectures. Due to these advances, computational scientists
and engineers can now solve large-scale problems that were once
thought intractable. Computational science and engineering
(CSE) is a rapidly growing multidisciplinary area with
connections to the sciences, engineering, and mathematics and
computer science. CSE focuses on the development of problemsolving
methodologies and robust tools for the solution of
scientific and engineering problems. We believe that CSE will
play an important if not dominating role for the future of the
scientific discovery process and engineering design. The
computation science is now being used widely for environmental
engineering calculations. The behavior of environmental
engineering systems and processes can be studied with the help
of computation science and understanding as well as better
solutions to environmental engineering problems can be
obtained.
M2CAT: Extracting reproducible simulation studies from model repositories usi...Martin Scharm
The document discusses M2CAT, a workflow that extracts reproducible simulation studies from model repositories. It searches the model repository Masymos for relevant studies, retrieves the necessary data, and exports it as a COMBINE archive using the CombineArchive Toolkit. This packages all the files into a single container that can be shared, modified and explored using various CAT tools. The workflow aims to make simulation studies more reproducible and accessible by bundling related models, data and descriptions into standardized packages.
Friday, October 15th, 2021, Sapporo, Hokkaido, Japan.
Hokkaido University ICReDD - Faculty of Medicine Joint Symposium
https://www.icredd.hokudai.ac.jp/event/5993
ICReDD (Institute for Chemical Reaction Design and Discovery)
https://www.icredd.hokudai.ac.jp
On August 29, 2005, Hurricane Katrina devastated New Orleans and t.docxhopeaustin33688
On August 29, 2005, Hurricane Katrina devastated New Orleans and the Gulf coast. Proctor & Gamble coffee manufacturing, with brands such as Folgers that get over half of their supply from sites in New Orleans, was severely impacted by the hurricane. Six months later, there were, as a P&G executive told the New York Times “still holes on the shelves” where P&G’s brands should be. Given your new insights from supply chain management, how would you solve/avoid a situation like this?
Module 2 - Case
Supply Chain Design
Case Assignment
Welcome to the second case study for this course.
Assignment: Please read the article below (available through ProQuest), then in a 3-4 page paper discuss the article and integrated supply chains.
Assignment Expectations
The authors of the article do a pretty good job of explaining the concept but I would like you to tell me what they mean in your own words. What is an integrated supply chain? What are the key elements/challenges in an integrated supply chain? What are the specific benefits to firms that implement superior supply chain management?Write a 3-4 page paper, using the same format as module one.
Integrated supply chains to be exploredAlan Johnson. Manufacturers' Monthly. Sydney: May 2007. It is attached
Abstract:
"The key challenge is to integrate supply chain capabilities to provide a seamless solution from potential design through to end delivery. End users are looking for a complete supply chain where there is single accountability and responsibility for delivery," said O'[Brien].
Module 2 - Background
Supply Chain Design
The following information will give you a good background on the importance of having a properly designed supply chain. Please review the information presented below to assist you with the assignments. I encourage you to surf the internet for more information on supply chain design.
Required Materials
You are not required to read all of these articles, but you may if you wish to choose to read several to further your knowledge. You are encouraged to surf the internet to gather additional resources in order to research your topic more thoroughly.
Start off this module by reviewing the article below on Supply Chain Design.
Beamon, B. M. (1998). Supply Chain Design and Analysis: Models and Methods. International Journal of Production Economics, 55(3), 281-294. Accessed August 10, 2009, at http://www.sclgme.org
The ProQuestdata base provides the articles below concerning changing supply chains and supply chain security.
Johnson, A., (2007). Integrated supply chains to be explored. Manufacturers' Monthly. Sydney. Available in the Trident Online Library.
Rogers, D., Lockman, D., Schwerdt, G., O'Donnell, B., & Huff, R., (2004). Supply Chain Security. Material Handling Management, 59(2), 15-18. Available in the Trident Online Library.
Supply Chain Design and Analysis:
Models and Methods
Benita M. Beamon
University of Washington
Industrial Engineering
Box 352650
Seattle, WA 98195.
This document introduces BioPreDyn-bench, a suite of benchmark problems for dynamic modelling in systems biology. The suite contains 6 benchmark problems ranging from medium to large-scale kinetic models of organisms such as E. coli, S. cerevisiae, D. melanogaster, and human cells. For each benchmark, the document provides a description, implementations in various formats, computational results from specific solvers, and analysis. The suite aims to serve as reference test cases to evaluate and compare parameter estimation methods for dynamic models in systems biology.
The document discusses the evolution of computational models over time. It provides the example of models of the cell cycle control network in the bacterium Clostridium acetobutylicum that have been developed and expanded on since 1984. The models have become more detailed over time, incorporating new experimental findings and insights. The document also discusses how models of the core cell cycle oscillator have evolved since 1991 with the addition of new molecular components and regulatory interactions based on studies in different organisms.
M2CAT: Extracting reproducible simulation studies from model repositories usi...Martin Scharm
Martin Scharm presents M2CAT, a workflow to extract reproducible simulation studies from model repositories. It searches the graph database Masymos for relevant models, simulations, and related data. M2CAT then uses the CombineArchive Toolkit to export the selected studies as COMBINE archive files, which provide a standard format for sharing complete simulation experiments. These archives can be explored and modified using the CombineArchive Web interface or other tools. The goal is to make the large amounts of data in repositories more usable and enable researchers to more easily reproduce and build upon existing computational studies.
This document provides an overview of standards and best practices for making computational models reusable through the use of model repositories and standard formats. It discusses the COMBINE initiative for standardizing the encoding of models and simulations. The document encourages authors to make their models and data FAIR (Findable, Accessible, Interoperable, Reusable) by using community standards for publishing, exchanging, and archiving models. Examples of open model repositories and standards-compliant tools and libraries are provided to demonstrate how authors can improve sharing and reuse of their models.
Applications of Computer Science in Environmental ModelsIJLT EMAS
Computation is now regarded as an equal and
indispensable partner, along with theory and experiment, in the
advance of scientific knowledge and engineering practice.
Numerical simulation enables the study of complex systems and
natural phenomena that would be too expensive or dangerous, or
even impossible, to study by direct experimentation. The quest
for ever higher levels of detail and realism in such simulations
requires enormous computational capacity, and has provided the
impetus for dramatic breakthroughs in computer algorithms and
architectures. Due to these advances, computational scientists
and engineers can now solve large-scale problems that were once
thought intractable. Computational science and engineering
(CSE) is a rapidly growing multidisciplinary area with
connections to the sciences, engineering, and mathematics and
computer science. CSE focuses on the development of problemsolving
methodologies and robust tools for the solution of
scientific and engineering problems. We believe that CSE will
play an important if not dominating role for the future of the
scientific discovery process and engineering design. The
computation science is now being used widely for environmental
engineering calculations. The behavior of environmental
engineering systems and processes can be studied with the help
of computation science and understanding as well as better
solutions to environmental engineering problems can be
obtained.
M2CAT: Extracting reproducible simulation studies from model repositories usi...Martin Scharm
The document discusses M2CAT, a workflow that extracts reproducible simulation studies from model repositories. It searches the model repository Masymos for relevant studies, retrieves the necessary data, and exports it as a COMBINE archive using the CombineArchive Toolkit. This packages all the files into a single container that can be shared, modified and explored using various CAT tools. The workflow aims to make simulation studies more reproducible and accessible by bundling related models, data and descriptions into standardized packages.
Friday, October 15th, 2021, Sapporo, Hokkaido, Japan.
Hokkaido University ICReDD - Faculty of Medicine Joint Symposium
https://www.icredd.hokudai.ac.jp/event/5993
ICReDD (Institute for Chemical Reaction Design and Discovery)
https://www.icredd.hokudai.ac.jp
On August 29, 2005, Hurricane Katrina devastated New Orleans and t.docxhopeaustin33688
On August 29, 2005, Hurricane Katrina devastated New Orleans and the Gulf coast. Proctor & Gamble coffee manufacturing, with brands such as Folgers that get over half of their supply from sites in New Orleans, was severely impacted by the hurricane. Six months later, there were, as a P&G executive told the New York Times “still holes on the shelves” where P&G’s brands should be. Given your new insights from supply chain management, how would you solve/avoid a situation like this?
Module 2 - Case
Supply Chain Design
Case Assignment
Welcome to the second case study for this course.
Assignment: Please read the article below (available through ProQuest), then in a 3-4 page paper discuss the article and integrated supply chains.
Assignment Expectations
The authors of the article do a pretty good job of explaining the concept but I would like you to tell me what they mean in your own words. What is an integrated supply chain? What are the key elements/challenges in an integrated supply chain? What are the specific benefits to firms that implement superior supply chain management?Write a 3-4 page paper, using the same format as module one.
Integrated supply chains to be exploredAlan Johnson. Manufacturers' Monthly. Sydney: May 2007. It is attached
Abstract:
"The key challenge is to integrate supply chain capabilities to provide a seamless solution from potential design through to end delivery. End users are looking for a complete supply chain where there is single accountability and responsibility for delivery," said O'[Brien].
Module 2 - Background
Supply Chain Design
The following information will give you a good background on the importance of having a properly designed supply chain. Please review the information presented below to assist you with the assignments. I encourage you to surf the internet for more information on supply chain design.
Required Materials
You are not required to read all of these articles, but you may if you wish to choose to read several to further your knowledge. You are encouraged to surf the internet to gather additional resources in order to research your topic more thoroughly.
Start off this module by reviewing the article below on Supply Chain Design.
Beamon, B. M. (1998). Supply Chain Design and Analysis: Models and Methods. International Journal of Production Economics, 55(3), 281-294. Accessed August 10, 2009, at http://www.sclgme.org
The ProQuestdata base provides the articles below concerning changing supply chains and supply chain security.
Johnson, A., (2007). Integrated supply chains to be explored. Manufacturers' Monthly. Sydney. Available in the Trident Online Library.
Rogers, D., Lockman, D., Schwerdt, G., O'Donnell, B., & Huff, R., (2004). Supply Chain Security. Material Handling Management, 59(2), 15-18. Available in the Trident Online Library.
Supply Chain Design and Analysis:
Models and Methods
Benita M. Beamon
University of Washington
Industrial Engineering
Box 352650
Seattle, WA 98195.
This document introduces BioPreDyn-bench, a suite of benchmark problems for dynamic modelling in systems biology. The suite contains 6 benchmark problems ranging from medium to large-scale kinetic models of organisms such as E. coli, S. cerevisiae, D. melanogaster, and human cells. For each benchmark, the document provides a description, implementations in various formats, computational results from specific solvers, and analysis. The suite aims to serve as reference test cases to evaluate and compare parameter estimation methods for dynamic models in systems biology.
The document presents a theoretical approach to engineering bioinspired systems through the design and characterization of non-coded amino acids. It discusses the characterization of arginine surrogates like c5Arg and (βPro)Arg through computational methods. It also characterizes indoline carboxylic acid and α-methyl indoline carboxylic acid as proline surrogates, analyzing their conformational preferences and energies using different theoretical methods and solvent models.
This document provides information about Bio-Modeling Systems SAS, including contact details and an overview of their services. It discusses the need to change the drug discovery paradigm to address issues like high failure rates and unreliable research. Bio-Modeling Systems provides a platform called CADITM Discovery that uses a mechanisms-based approach and principles like negative selection to model biological complexity and address uncertainties in scientific data. The goal is to improve medical research and validation processes to enhance patient health outcomes.
Model Management in Systems Biology: Challenges – Approaches – SolutionsMartin Scharm
I gave this talk as a webinar in the FAIRDOM webinar series 2016. The recordings of the webinar are available from http://fair-dom.org/knowledgehub/webinars-2/martin-scharm/
Introduction to chemical kinetics - WT/EBI course systems biology 2018Nicolas Le Novère
The document discusses modelling chemical kinetics and reactions. It describes how processes can be combined in ordinary differential equations (ODEs) for deterministic simulations or transformed into propensities for stochastic simulations. It also discusses the law of mass action, enzyme kinetics including Michaelis-Menten and Hill equations, and homeostasis - maintaining stable levels through negative feedback loops.
2. leiviskä k (1996) simulation in pulp and paper industry. february 1996Huy Nguyen
This document discusses simulation in the pulp and paper industry. It provides an overview of how mathematical models and simulation are used at different levels, including research and development, product and process design, control engineering, and process operation. Simulation is employed for tasks such as disturbance analysis, start-up and shut-down planning, real-time control and optimization, and operator training. The document also covers the basics of simulation, including model classification, selection, scope, and solution principles.
This talk was part of the 2020 Disease Map Modeling Community meeting, covering the steps towards publishing reproducible simulation studies (based on a reused model). Links to different COMBINE guidelines, tutorials and efforts. Grants: European Commission: EOSCsecretariat.eu - EOSCsecretariat.eu (831644)
Paper Annotated: SinGAN-Seg: Synthetic Training Data Generation for Medical I...Devansh16
YouTube video: https://www.youtube.com/watch?v=Ao-19L0sLOI
SinGAN-Seg: Synthetic Training Data Generation for Medical Image Segmentation
Vajira Thambawita, Pegah Salehi, Sajad Amouei Sheshkal, Steven A. Hicks, Hugo L.Hammer, Sravanthi Parasa, Thomas de Lange, Pål Halvorsen, Michael A. Riegler
Processing medical data to find abnormalities is a time-consuming and costly task, requiring tremendous efforts from medical experts. Therefore, Ai has become a popular tool for the automatic processing of medical data, acting as a supportive tool for doctors. AI tools highly depend on data for training the models. However, there are several constraints to access to large amounts of medical data to train machine learning algorithms in the medical domain, e.g., due to privacy concerns and the costly, time-consuming medical data annotation process. To address this, in this paper we present a novel synthetic data generation pipeline called SinGAN-Seg to produce synthetic medical data with the corresponding annotated ground truth masks. We show that these synthetic data generation pipelines can be used as an alternative to bypass privacy concerns and as an alternative way to produce artificial segmentation datasets with corresponding ground truth masks to avoid the tedious medical data annotation process. As a proof of concept, we used an open polyp segmentation dataset. By training UNet++ using both the real polyp segmentation dataset and the corresponding synthetic dataset generated from the SinGAN-Seg pipeline, we show that the synthetic data can achieve a very close performance to the real data when the real segmentation datasets are large enough. In addition, we show that synthetic data generated from the SinGAN-Seg pipeline improving the performance of segmentation algorithms when the training dataset is very small. Since our SinGAN-Seg pipeline is applicable for any medical dataset, this pipeline can be used with any other segmentation datasets.
Subjects: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
Cite as: arXiv:2107.00471 [eess.IV]
(or arXiv:2107.00471v1 [eess.IV] for this version)
Reach out to me:
Check out my other articles on Medium. : https://machine-learning-made-simple....
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn: https://www.linkedin.com/in/devansh-d...
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
Statistical modeling in pharmaceutical research and developmentPV. Viji
Statistical modeling in pharmaceutical research and development , Statistical Modeling , Descriptive Versus Mechanistic Modeling , Statistical Parameters Estimation , Confidence Regions , Non Linearity at the Optimum , Sensitivity Analysis , Optimal Design , Population Modeling
LIFE EXPECTANCY PREDICTION FOR POST THORACIC SURGERYIRJET Journal
This document discusses using machine learning algorithms to predict life expectancy after thoracic surgery. Researchers used attribute ranking and selection methods to identify the most important attributes from a dataset of patient health records. They evaluated algorithms like logistic regression and random forest on the reduced dataset. Logistic regression achieved the highest prediction accuracy of 85%. The goal was to more accurately predict mortality risk based on a patient's underlying health issues and attributes related to lung cancer.
IRJET-Automatic RBC And WBC Counting using Watershed Segmentation AlgorithmIRJET Journal
This document presents a method for automatically counting red blood cells (RBCs) and white blood cells (WBCs) using image processing techniques. It discusses the limitations of conventional manual counting methods and proposes a software-based watershed segmentation algorithm to segment and count blood cells from microscope images. The algorithm involves preprocessing the image, applying filters, segmenting cells using markers and boundaries, and counting the segmented cells. Experimental results found the automatic method took 14.43 seconds on average and achieved 94.58% accuracy, faster and more accurate than manual counting. This software-based solution provides a low-cost alternative for blood cell analysis in medical laboratories.
The Cytell Cell Imaging System is a compact, automated cell imaging system that combines digital microscopy, image cytometry and cell counting capabilities. It provides high quality cellular and subcellular image data and analysis through the use of preconfigured biological applications. These applications simplify routine assays and tasks while generating detailed experimental reports. The system allows non-experts to perform assays but also provides flexibility for experienced researchers. It captures multi-channel fluorescence and brightfield images to analyze cells in a variety of sample formats.
IRJET - IoT based Steroid Measurement in Milk ProductsIRJET Journal
The document discusses an IOT-based system to measure steroid levels in milk products. It aims to address the issue of dairy farmers illegally injecting large amounts of steroids into cattle to increase milk production, which can harm human health. The proposed system uses sensors to analyze steroid levels in cattle and milk at various stages, and indicates which organs may be affected based on the chemicals detected. It analyzes the data using machine learning algorithms and displays the results through a graphical interface. The system aims to more accurately detect illegal steroid usage and protect consumer health.
An introductory workshop about machine learning in chemistry. This workshop is a set of slides and jupyter notebooks intended to give an overview of machine learning in chemistry to graduate students in chemical sciences, which was originally presented during a research trip to Ben Gurion University and the Hebrew University in Jerusalem in February 2019. Part 1 of 2.
The workshop lives at https://github.com/jpjanet/ML-chem-workshop where it is maintained in an up-to-date fashion. Notebook examples can be obtained from the GitHub page.
Economy and forecast for 2020 3 key trends in the futureeSAT Journals
Abstract The article deals with 3 key trends in the future and their general implications including 3D, RFID, Business Intelligence and new managerial positions. 3D by 2020 could replace conventional mass-production. The basic trends in the RFID aplications will be: RFID Wearables,RFID On Merchandise, Host Card Emulation (HCE) Payment Solutions,Printed RFID Technology, RFID chip tracking everyone everywhere in the near future. Business intelligence will be transformed to the general intelligence.The contribution covers the the following topics: selected Aspects of economy and social Aspects of Information Systems, complex technological and human Issues in today’s globalized and interconnected World and presents new results in the diffused way. Key words: 3 Key trends, 3D, RFID, Business Intelligence, Computer Feudal Monarchy, New Managerial Positions JEL Classification: A10, A11, A19, E27, E69
Data analytics and software sensors for single use bioprocessingExputec
This document discusses data analytics and software sensors for single-use bioprocessing. It provides an overview of Exputec, which offers solutions to improve biopharmaceutical development and production processes using data analytics. Software sensors are described as using models to monitor process variables like substrates and metabolites, complementing hardware sensors. Challenges of data analytics include data integration from collaborators and dealing with outliers. The document advocates for using robust multivariate statistics rather than removing outliers to avoid incorrect results from data analytics.
Towards Non-invasive Labour Detection: A Free- Living EvaluationMarco Altini
Slides of my presentation at EMBC 2018, more info on this research is available here: https://www.researchgate.net/project/Bloomlife-improving-prenatal-health-through-longitudinal-physiological-monitoring-at-large-scale?_sg=pbraocCDNc2lJd9v5GESvRhkmffW99OTeeNMkalglCirK5r-ZECp2XRy_5Otk-_B-_dlCalxvKUVtex9MkAHUPFKhHT56GfrO6h3
Statistical quality control applied industrial and manufacturing operations. Case study regarding the use of these tools. Description of statistical tools used in quality control and inspection.
This document provides an introduction to mathematical modeling. It discusses key concepts such as dimensional analysis, the Buckingham Pi theorem, types of mathematical models including static vs dynamic, discrete vs continuous, deterministic vs probabilistic, and linear vs nonlinear models. Examples of mathematical modeling applications in various fields like physics, engineering, biology and combat modeling are provided. The modeling process from defining the real-world problem to model formulation, solution, evaluation and refinement is outlined. Dimensional analysis using Rayleigh, Buckingham and Bridgman methods is explained through a heat transfer example.
ReComp and P4@NU: Reproducible Data Science for HealthPaolo Missier
brief overview of the ReComp project (http://recomp.org.uk) on Selective recurring re-computation of complex analytics, and a brief outlook for the P4@NU project on seeking digital biomarkers for age-0related metabolic diseases
Invited to introduce Docker to the Dept. for Information and Communication Services (Informations- und Kommunikationsdienste - IuK) at the University of Rostock.
More Related Content
Similar to Improving Reproducibility and Reuse of Modelling Results in the Life Sciences
The document presents a theoretical approach to engineering bioinspired systems through the design and characterization of non-coded amino acids. It discusses the characterization of arginine surrogates like c5Arg and (βPro)Arg through computational methods. It also characterizes indoline carboxylic acid and α-methyl indoline carboxylic acid as proline surrogates, analyzing their conformational preferences and energies using different theoretical methods and solvent models.
This document provides information about Bio-Modeling Systems SAS, including contact details and an overview of their services. It discusses the need to change the drug discovery paradigm to address issues like high failure rates and unreliable research. Bio-Modeling Systems provides a platform called CADITM Discovery that uses a mechanisms-based approach and principles like negative selection to model biological complexity and address uncertainties in scientific data. The goal is to improve medical research and validation processes to enhance patient health outcomes.
Model Management in Systems Biology: Challenges – Approaches – SolutionsMartin Scharm
I gave this talk as a webinar in the FAIRDOM webinar series 2016. The recordings of the webinar are available from http://fair-dom.org/knowledgehub/webinars-2/martin-scharm/
Introduction to chemical kinetics - WT/EBI course systems biology 2018Nicolas Le Novère
The document discusses modelling chemical kinetics and reactions. It describes how processes can be combined in ordinary differential equations (ODEs) for deterministic simulations or transformed into propensities for stochastic simulations. It also discusses the law of mass action, enzyme kinetics including Michaelis-Menten and Hill equations, and homeostasis - maintaining stable levels through negative feedback loops.
2. leiviskä k (1996) simulation in pulp and paper industry. february 1996Huy Nguyen
This document discusses simulation in the pulp and paper industry. It provides an overview of how mathematical models and simulation are used at different levels, including research and development, product and process design, control engineering, and process operation. Simulation is employed for tasks such as disturbance analysis, start-up and shut-down planning, real-time control and optimization, and operator training. The document also covers the basics of simulation, including model classification, selection, scope, and solution principles.
This talk was part of the 2020 Disease Map Modeling Community meeting, covering the steps towards publishing reproducible simulation studies (based on a reused model). Links to different COMBINE guidelines, tutorials and efforts. Grants: European Commission: EOSCsecretariat.eu - EOSCsecretariat.eu (831644)
Paper Annotated: SinGAN-Seg: Synthetic Training Data Generation for Medical I...Devansh16
YouTube video: https://www.youtube.com/watch?v=Ao-19L0sLOI
SinGAN-Seg: Synthetic Training Data Generation for Medical Image Segmentation
Vajira Thambawita, Pegah Salehi, Sajad Amouei Sheshkal, Steven A. Hicks, Hugo L.Hammer, Sravanthi Parasa, Thomas de Lange, Pål Halvorsen, Michael A. Riegler
Processing medical data to find abnormalities is a time-consuming and costly task, requiring tremendous efforts from medical experts. Therefore, Ai has become a popular tool for the automatic processing of medical data, acting as a supportive tool for doctors. AI tools highly depend on data for training the models. However, there are several constraints to access to large amounts of medical data to train machine learning algorithms in the medical domain, e.g., due to privacy concerns and the costly, time-consuming medical data annotation process. To address this, in this paper we present a novel synthetic data generation pipeline called SinGAN-Seg to produce synthetic medical data with the corresponding annotated ground truth masks. We show that these synthetic data generation pipelines can be used as an alternative to bypass privacy concerns and as an alternative way to produce artificial segmentation datasets with corresponding ground truth masks to avoid the tedious medical data annotation process. As a proof of concept, we used an open polyp segmentation dataset. By training UNet++ using both the real polyp segmentation dataset and the corresponding synthetic dataset generated from the SinGAN-Seg pipeline, we show that the synthetic data can achieve a very close performance to the real data when the real segmentation datasets are large enough. In addition, we show that synthetic data generated from the SinGAN-Seg pipeline improving the performance of segmentation algorithms when the training dataset is very small. Since our SinGAN-Seg pipeline is applicable for any medical dataset, this pipeline can be used with any other segmentation datasets.
Subjects: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
Cite as: arXiv:2107.00471 [eess.IV]
(or arXiv:2107.00471v1 [eess.IV] for this version)
Reach out to me:
Check out my other articles on Medium. : https://machine-learning-made-simple....
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn: https://www.linkedin.com/in/devansh-d...
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
Statistical modeling in pharmaceutical research and developmentPV. Viji
Statistical modeling in pharmaceutical research and development , Statistical Modeling , Descriptive Versus Mechanistic Modeling , Statistical Parameters Estimation , Confidence Regions , Non Linearity at the Optimum , Sensitivity Analysis , Optimal Design , Population Modeling
LIFE EXPECTANCY PREDICTION FOR POST THORACIC SURGERYIRJET Journal
This document discusses using machine learning algorithms to predict life expectancy after thoracic surgery. Researchers used attribute ranking and selection methods to identify the most important attributes from a dataset of patient health records. They evaluated algorithms like logistic regression and random forest on the reduced dataset. Logistic regression achieved the highest prediction accuracy of 85%. The goal was to more accurately predict mortality risk based on a patient's underlying health issues and attributes related to lung cancer.
IRJET-Automatic RBC And WBC Counting using Watershed Segmentation AlgorithmIRJET Journal
This document presents a method for automatically counting red blood cells (RBCs) and white blood cells (WBCs) using image processing techniques. It discusses the limitations of conventional manual counting methods and proposes a software-based watershed segmentation algorithm to segment and count blood cells from microscope images. The algorithm involves preprocessing the image, applying filters, segmenting cells using markers and boundaries, and counting the segmented cells. Experimental results found the automatic method took 14.43 seconds on average and achieved 94.58% accuracy, faster and more accurate than manual counting. This software-based solution provides a low-cost alternative for blood cell analysis in medical laboratories.
The Cytell Cell Imaging System is a compact, automated cell imaging system that combines digital microscopy, image cytometry and cell counting capabilities. It provides high quality cellular and subcellular image data and analysis through the use of preconfigured biological applications. These applications simplify routine assays and tasks while generating detailed experimental reports. The system allows non-experts to perform assays but also provides flexibility for experienced researchers. It captures multi-channel fluorescence and brightfield images to analyze cells in a variety of sample formats.
IRJET - IoT based Steroid Measurement in Milk ProductsIRJET Journal
The document discusses an IOT-based system to measure steroid levels in milk products. It aims to address the issue of dairy farmers illegally injecting large amounts of steroids into cattle to increase milk production, which can harm human health. The proposed system uses sensors to analyze steroid levels in cattle and milk at various stages, and indicates which organs may be affected based on the chemicals detected. It analyzes the data using machine learning algorithms and displays the results through a graphical interface. The system aims to more accurately detect illegal steroid usage and protect consumer health.
An introductory workshop about machine learning in chemistry. This workshop is a set of slides and jupyter notebooks intended to give an overview of machine learning in chemistry to graduate students in chemical sciences, which was originally presented during a research trip to Ben Gurion University and the Hebrew University in Jerusalem in February 2019. Part 1 of 2.
The workshop lives at https://github.com/jpjanet/ML-chem-workshop where it is maintained in an up-to-date fashion. Notebook examples can be obtained from the GitHub page.
Economy and forecast for 2020 3 key trends in the futureeSAT Journals
Abstract The article deals with 3 key trends in the future and their general implications including 3D, RFID, Business Intelligence and new managerial positions. 3D by 2020 could replace conventional mass-production. The basic trends in the RFID aplications will be: RFID Wearables,RFID On Merchandise, Host Card Emulation (HCE) Payment Solutions,Printed RFID Technology, RFID chip tracking everyone everywhere in the near future. Business intelligence will be transformed to the general intelligence.The contribution covers the the following topics: selected Aspects of economy and social Aspects of Information Systems, complex technological and human Issues in today’s globalized and interconnected World and presents new results in the diffused way. Key words: 3 Key trends, 3D, RFID, Business Intelligence, Computer Feudal Monarchy, New Managerial Positions JEL Classification: A10, A11, A19, E27, E69
Data analytics and software sensors for single use bioprocessingExputec
This document discusses data analytics and software sensors for single-use bioprocessing. It provides an overview of Exputec, which offers solutions to improve biopharmaceutical development and production processes using data analytics. Software sensors are described as using models to monitor process variables like substrates and metabolites, complementing hardware sensors. Challenges of data analytics include data integration from collaborators and dealing with outliers. The document advocates for using robust multivariate statistics rather than removing outliers to avoid incorrect results from data analytics.
Towards Non-invasive Labour Detection: A Free- Living EvaluationMarco Altini
Slides of my presentation at EMBC 2018, more info on this research is available here: https://www.researchgate.net/project/Bloomlife-improving-prenatal-health-through-longitudinal-physiological-monitoring-at-large-scale?_sg=pbraocCDNc2lJd9v5GESvRhkmffW99OTeeNMkalglCirK5r-ZECp2XRy_5Otk-_B-_dlCalxvKUVtex9MkAHUPFKhHT56GfrO6h3
Statistical quality control applied industrial and manufacturing operations. Case study regarding the use of these tools. Description of statistical tools used in quality control and inspection.
This document provides an introduction to mathematical modeling. It discusses key concepts such as dimensional analysis, the Buckingham Pi theorem, types of mathematical models including static vs dynamic, discrete vs continuous, deterministic vs probabilistic, and linear vs nonlinear models. Examples of mathematical modeling applications in various fields like physics, engineering, biology and combat modeling are provided. The modeling process from defining the real-world problem to model formulation, solution, evaluation and refinement is outlined. Dimensional analysis using Rayleigh, Buckingham and Bridgman methods is explained through a heat transfer example.
ReComp and P4@NU: Reproducible Data Science for HealthPaolo Missier
brief overview of the ReComp project (http://recomp.org.uk) on Selective recurring re-computation of complex analytics, and a brief outlook for the P4@NU project on seeking digital biomarkers for age-0related metabolic diseases
Similar to Improving Reproducibility and Reuse of Modelling Results in the Life Sciences (20)
Invited to introduce Docker to the Dept. for Information and Communication Services (Informations- und Kommunikationsdienste - IuK) at the University of Rostock.
This document discusses metadata for the COMBINE archive format. It begins by explaining that the COMBINE archive allows sharing of all information needed to reproduce a modeling project in a single file. It then provides examples of metadata for models, simulations, and results within a COMBINE archive. Finally, it proposes extending the metadata by integrating additional ontologies such as PROV and PAV to capture more detailed information about models, simulations, people, and relationships between these elements.
Characterising differences between model versionsMartin Scharm
The document discusses characterizing differences between versions of computational models. It describes a system called BiVeS that identifies changes by mapping elements between two model versions hierarchically, then evaluating the mapping to find inserts, deletes, updates and moves. An ontology called COMODI is being built to classify model changes by type, target, reason, intention and more to help understand how and why models evolve.
HandsOn: git (or version control in general...)Martin Scharm
The document discusses an invited talk on Git and version control approaches given in Rostock, Germany in 2015. It provides an overview of topics that were covered in the talk, including commits, branches, tags, undoing changes, diffs, logs, stashing, GitHub/Bitbucket, syncing across machines, files to ignore with .gitignore, Git modules, aliases, relative paths, common problems like detached heads and merge conflicts, collaboration techniques, and using Git in the cloud. Examples and images are linked to illustrate various points. The talk aimed to provide insights into national and international practices using version control systems.
The CellML models’ walk through the repositoryMartin Scharm
This document summarizes a retrospective study of the CellML model repository. It found that the repository grew from around 500 models in 2007 to over 2,500 models in 2013. Version control was implemented to track model development over time, revealing over 18,000 total versions and 15,000 differences between versions. Common operations on models in the repository included updates, deletes, inserts, and moves of components. The study provided insights into the evolution of models in the repository.
CombineArchiveWeb -- web based tool to handle files associated with modelling...Martin Scharm
The document discusses the CombineArchive Toolkit (CAT), a web-based tool for handling and sharing files associated with modeling results. CAT allows users to create, modify, explore, and share models and data through a single CombineArchive file format. It was created by researchers at the University of Rostock's Department of Systems Biology and Bioinformatics.
Improving the Management of Computational Models -- Invited talk at the EBIMartin Scharm
Improving the Management of Computational Models:
storage – retrieval & ranking – version control
More information and slides to download at http://sems.uni-rostock.de/2013/12/martin-visits-the-ebi/
Talk by Martin Scharm at the COMBINE meeting September 2013 in Paris.
Find more information and download the slides at http://sems.uni-rostock.de/2013/09/sems-at-the-combine-2013/
Talk at the IPK in Gatersleben to initiate collaborations with the VANTED/e!Dal crews. The slides are also available at our website http://sems.uni-rostock.de/2013/05/research-visit-at-the-ipk-gatersleben/
This document describes BiVeS and BudHat, tools for version control of computational models. BiVeS detects differences between versions of models encoded in standardized formats like SBML. It maps common elements and constructs an XML diff file. BudHat visualizes these diffs, highlighting changes in reaction networks or providing a readable report. The tools are open source and intended to extend existing model repositories with version control capabilities. A demonstration of BiVeS and BudHat in action is provided.
BiVeS & BudHat -- Version Control for Computational Models @ All hands PALs M...Martin Scharm
Talk about version control for computational models at the All hands PALs Meeting in Heidelberg by Martin Scharm, member of the SEMS project at the University of Rostock.
The slides are also available at our website http://sems.uni-rostock.de/2012/11/sems-at-the-pals-meeting-2012/
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
Infrastructure Challenges in Scaling RAG with Custom AI models
Improving Reproducibility and Reuse of Modelling Results in the Life Sciences
1. SYSTEMS BIOLOGY
BIOINFORMATICS
ROSTOCK
S E Ssimulation experiment management system
Improving Reproducibility and Reuse of
Modelling Results in the Life Sciences
MARTIN SCHARM
Department of Systems Biology & Bioinformatics, University of Rostock
https://sbi.uni-rostock.de/scharm
Rostock, 30 August 2018
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 1
2. Outline of my talk
Model 1
v1 v2 v3
Model 2
v1 v2 v3 v4
Model 3
v1 v2
Correct Hypothesis 1
Hypothesis 2
Improve
A method to characterise differences in
computational models (Chapters 2 and 3)
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 2
3. Outline of my talk
Search&Retrieve
Publish&Share
Model 1
v1 v2 v3
Model 2
v1 v2 v3 v4
Model 3
v1 v2
Correct Hypothesis 1
Hypothesis 2
Improve
A method to characterise differences in
computational models (Chapters 2 and 3)
Support for shareable and reproducible
simulation studies (Chapter 4)
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 2
4. Outline of my talk
Search&Retrieve
Publish&Share
Model 1
v1 v2 v3
Model 2
v1 v2 v3 v4
Model 3
v1 v2
Correct Hypothesis 1
Hypothesis 2
Improve
A method to characterise differences in
computational models (Chapters 2 and 3)
Support for shareable and reproducible
simulation studies (Chapter 4)
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 2
5. Computational models evolve
an example from real life
Novak and Tyson: Numerical analysis of a comprehensive model of M-phase control in Xenopus oocyte extracts and intact embryos.
In Journal of Cell Science 1993 106:1153-1168
Models evolve: I not just witnessed changes, but induced model evolution myself.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 3
6. Computational models evolve
an example from real life
Novak and Tyson: Numerical analysis of a comprehensive model of M-phase control in Xenopus oocyte extracts and intact embryos.
In Journal of Cell Science 1993 106:1153-1168
Models evolve: I not just witnessed changes, but induced model evolution myself.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 3
7. Computational models evolve
an example from real life
Novak and Tyson: Numerical analysis of a comprehensive model of M-phase control in Xenopus oocyte extracts and intact embryos.
In Journal of Cell Science 1993 106:1153-1168
Models evolve: I not just witnessed changes, but induced model evolution myself.
Cyclin
Cdc2
Cdc2Cyclin
SBGN
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 3
8. Computational models evolve
an example from real life
June 2007
The model with id BIOMD0000000107 was published in
BioModels Database. It encodes the biological system
described by Novak and Tyson in 1993.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 4
9. Computational models evolve
an example from real life
June 2007 June 2013
Along with many other models in BioModels Database, the model
was released in multiple versions between 2007 and 2013.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 4
10. Computational models evolve
an example from real life
Cyclin
Cdc2
Cdc2
Cyclin
June 2007
SBML code has changed
June 2013
differences
modifications
inserts
deletes
The changes however do not affect
the graphical representation of the
reaction.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 4
11. Computational models evolve
an example from real life
Cyclin
Cdc2
Cdc2
Cyclin
Cyclin
Cdc2
Cdc2
Cdc2Cyclin
June 2007
SBML code has changed
Model was corrected
June 2013
November 2013
differences
modifications
inserts
deletes
When studying the model
I found an error in that reaction
and corrected the model.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 4
12. Computational models evolve
an example from real life
Cyclin
Cdc2
Cdc2
Cyclin
Cyclin
Cdc2
Cdc2
Cyclin
Cyclin
Cdc2
Cdc2
Cdc2Cyclin
June 2007
SBML code has changed No modifications
Model was corrected
June 2013
November 2013
April 2015
differences
modifications
inserts
deletes
BioModels Database created
many more releases, but none of
them affected the reaction.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 4
13. Computational models evolve
an example from real life
Cyclin
Cdc2
Cdc2
Cyclin
Cyclin
Cdc2
Cdc2
Cyclin
Cyclin
Cdc2
Cdc2
Cdc2Cyclin
June 2007
SBML code has changed No modifications
Model was corrected
June 2013
November 2013
April 2015 Latest Version
in BioModels
differences
modifications
inserts
deletes
In 2016 my corrections were merged
into BioModels Database.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 4
14. Computational models evolve
even after publication
Scharm, Gebhardt, Touré, Bagnacani, Salehzadeh-Yazdi, Wolkenhauer and Waltemath: Evolution of computational models in
BioModels Database and the Physiome Model Repository. In BMC Systems Biology 2018 12:53.
BIOMD0001
BIOMD0021
BIOMD0041
BIOMD0061
BIOMD0081
BIOMD0101
BIOMD0121
BIOMD0141
BIOMD0161
BIOMD0181
BIOMD0201
BIOMD0221
BIOMD0241
BIOMD0261
BIOMD0281
BIOMD0301
BIOMD0321
BIOMD0341
BIOMD0361
BIOMD0381
BIOMD0401
BIOMD0421
BIOMD0441
BIOMD0461
BIOMD0481
BIOMD0501
BIOMD0521
BIOMD0541
BIOMD0561
BIOMD0581
BIOMD0601
Apr 05
Jul 05
Jan 06
Jun 06
Oct 06
Jan 07
Jun 07
Sep 07
Mar 08
Aug 08
Dec 08
Mar 09
Jun 09
Sep 09
Jan 10
Apr 10
Sep 10
Apr 11
Sep 11
Feb 12
May 12
Aug 12
Dec 12
Jun 13
Nov 13
Apr 14
Sep 14
Apr 15
May 16
Jun 17
0
5
10
50
100
500
1000
5000
17425
Numberofdifferences
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 5
15. Computational models evolve
changes need to be communicated
Waltemath, Henkel, Hälke, Scharm and Wolkenhauer: Improving the Reuse of Computational Models through Version Control.
In Bioinformatics 2013 29:6 742–748.
• Requirements of a version control system for computational models:
• it must be entailed to the structure of model documents,
• changes must be transparent and versions must be unambiguously identifiable,
addressable, and accessible,
• changes must be justified.
• Only with proper difference detection at hand users are able to grasp a model’s
history and to identify errors and inconsistencies.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 6
17. Versioning of models
June 2007 June 2013
November 2013
April 2015 Latest Version
in BioModels
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 7
18. Versioning of models
June 2007 June 2013
November 2013
April 2015 Latest Version
in BioModels
track versions ✓
BioModels Database collects
TM
Physiome Model Repository focuses on
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 7
19. Versioning of models
June 2007 June 2013
November 2013
April 2015 Latest Version
in BioModels
track versions ✓
What happened?
BioModels Database collects
TM
Physiome Model Repository focuses on
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 7
20. Differences in models
detection and communication
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
A method to detect and characterise differences in versions of computational models,
including
1. an algorithm to detect changes based on the ideas of the XyDiff algorithm,
2. output formats to communicate the changes,
3. an ontology to semantically describe changes.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 8
21. Differences in models
detecting changes
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
A
B
C D E
F
G
A
B
D H E
F
G
model version 1
model version 2
list of species list of reactions
C + D E D + H E
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 9
22. Differences in models
detecting changes
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
A
B
C D E
F
G
A
B
D H E
F
G
id=“cdc2” GO:0000123
id=“listOfReactions”
initialmappingpb
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 9
23. Differences in models
detecting changes
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
A
B
C D E
F
G
A
B
D H E
F
G
initialmappingpb
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 9
24. Differences in models
detecting changes
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
A
B
C D E
F
G
A
B
D H E
F
G
propagationpb
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 9
25. Differences in models
detecting changes
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
A
B
C D E
F
G
A
B
D H E
F
G
evaluationpb
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 9
26. Differences in models
detecting changes
Scharm, Wolkenhauer and Waltemath: An algorithm to detect and communicate the differences in computational models describing
biological systems. In Bioinformatics 2015 32:4 563-570
C
D
H E
communicationpb
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 9
28. Differences in models
semantic characterisation
• A good communication of model changes can increase the trust in a model.
• COMODI — COmputational MOdels DIffer.
• The COMODI ontology provides terms to characterise changes in models.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 11
29. Differences in models
the COMODI ontology
Scharm, Waltemath, Mendes and Wolkenhauer: COMODI: an ontology to characterise differences in versions of computational models in biology.
In Journal of Biomedical Semantics 2016 7:46.
wasTriggeredBy
affects
hasReason
hasIntention
appliesTo
Change
XmlEntity
Target
Intention
Reason
Correction
Simplification
Annotations
Setup
Reaction Network
Kinetics
Node
Attribute
Typo
Curation
insertion
deletion
update
65 classes
5 object properties
purl.uni-rostock.de/comodi
BioPortal::COMODI
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 12
30. Differences in models
an example
<parameter name=“Km1”
value=“0.3”
units=“molesperlitre” />
<parameter name=“Km1”
value=“0.7”
units=“molesperlitre” />
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 13
31. Differences in models
an example
<parameter name=“Km1”
value=“0.3”
units=“molesperlitre” />
<parameter name=“Km1”
value=“0.7”
units=“molesperlitre” />
<update>
<attribute id=“1” name=“value”
oldPath=“/sbml[1]/../parameter[1]”
oldValue=“0.3” newValue=“0.7”
[...] />
</update>
difference detection
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 13
32. Differences in models
an example
<parameter name=“Km1”
value=“0.3”
units=“molesperlitre” />
<parameter name=“Km1”
value=“0.7”
units=“molesperlitre” />
<update>
<attribute id=“1” name=“value”
oldPath=“/sbml[1]/../parameter[1]”
oldValue=“0.3” newValue=“0.7”
[...] />
</update>
difference detection
#1 a comodi:Update ;
comodi:appliesTo comodi:XmlAttribute ;
comodi:affects comodi:ParameterSetup ;
comodi:hasIntention comodi:Correction ;
comodi:hasReason comodi:MismatchWithPublication.
annotation
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 13
33. Applications
the BiVeS framework
BiVeS is a framework implementing the method to characterise differences in
computational models.
BiVeS
BiVeS-SBML BiVeS-CellML
BiVeS-Core
jCOMODI xmlutils
BiVeS-WebApp
BiVeS-WebApp-Client
HTTP
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 14
34. Applications
the BiVeS framework
BiVeS is a framework implementing the method to characterise differences in
computational models.
BiVeS
BiVeS-SBML BiVeS-CellML
BiVeS-Core
jCOMODI xmlutils
BiVeS-WebApp
BiVeS-WebApp-Client
HTTPJava API
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 14
35. Applications
the BiVeS framework
BiVeS is a framework implementing the method to characterise differences in
computational models.
BiVeS
BiVeS-SBML BiVeS-CellML
BiVeS-Core
jCOMODI xmlutils
BiVeS-WebApp
BiVeS-WebApp-Client
HTTP
Executable
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 14
36. Applications
the BiVeS framework
BiVeS is a framework implementing the method to characterise differences in
computational models.
BiVeS
BiVeS-SBML BiVeS-CellML
BiVeS-Core
jCOMODI xmlutils
BiVeS-WebApp
BiVeS-WebApp-Client
HTTP
Web interface
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 14
37. Applications
in the systems biology domain
SEEK and the
FAIRDOMHub
Physiome Model
Repository
Cardiac Electrophysiology
Web Lab
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 15
38. Outline of my talk
Search&Retrieve
Publish&Share
Model 1
v1 v2 v3
Model 2
v1 v2 v3 v4
Model 3
v1 v2
Correct Hypothesis 1
Hypothesis 2
Improve
A method to characterise differences in
computational models (Chapters 2 and 3)
Support for shareable and reproducible
simulation studies (Chapter 4)
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 16
39. Outline of my talk
Search&Retrieve
Publish&Share
Model 1
v1 v2 v3
Model 2
v1 v2 v3 v4
Model 3
v1 v2
Correct Hypothesis 1
Hypothesis 2
Improve
A method to characterise differences in
computational models (Chapters 2 and 3)
Support for shareable and reproducible
simulation studies (Chapter 4)
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 16
40. The COMBINE archive
one file to share them all
Bergmann, Adams, Moodie, Cooper, Glont and Scharm et al.: COMBINE archive and OMEX format: one file to share all information to reproduce
a modeling project. In BMC Bioinformatics 2014 15:369.
01001
00111
1101001
0010110
1110100
1010110
TM
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 17
41. The COMBINE archive
a fully featured demo
Scharm and Waltemath: A fully featured COMBINE archive of a simulation study on syncytial mitotic cycles in Drosophila embryos.
In F1000Research 2016 5:2421.
01001
00111
1101001
0010110
1110100
1010110
TM
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 18
42. Applications
employing COMBINE archives
CombineArchiveWeb
application – manage
modelling results
SED-ML web tools:
generate, modify and
export sim. studies
Extracting
reproducible
simulation studies
The Cardiac
Electrophysiology
Web Lab
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 19
43. Summary
Search&Retrieve
Publish&Share
Model 1
v1 v2 v3
Model 2
v1 v2 v3 v4
Model 3
v1 v2
Correct Hypothesis 1
Hypothesis 2
Improve
• Models are subject to changes.
• A novel method to identify and
characterise changes:
• entailed to the structure of models,
• changes become transparent,
• changes can be justified.
• Methods and tools support sharing and
reproducing of research results.
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 20
44. Many thanks to...
SYSTEMS BIOLOGY
BIOINFORMATICS
ROSTOCK
Olaf Wolkenhauer
Ali Salehzadeh-Yazdi
Holger Hennig
Tom Theile
Sherry Freiesleben
Markus Wolfien
Peggy Sterling
Ulf Liebal
Shailendra Gupta
S E Ssimulation experiment management system
Dagmar Waltemath
Martin Peters
Tom Gebhardt
Vasundra Touré
Mariam Nassar
Fabienne Lambusch
Ron Henkel
Vivek Garg
Srijana Kayastha
Jenny Fabian
Jonathan Cooper
Stian Soiland-Reyes
Natalie Stanford
Matthew Gamble
Gary Mirams
Carole Goble
Pedro Mendes
Tommy Yu
Frank Bergmann
Olga Krebs
Falk Schreiber
CellML
SBML
COPASI
FAIRDOM
Inkscape
LATEX
Texstudio
Protégé
Java
Maven
Docker
Debian, GNU/Linux
Stack Exchange
Git, SVN, Hg
Python, Perl, PHP
Trac, Git(Hub|Lab), Wiki*
Icedove
Iceweasel
OpenSSH
Tomcat
Eclipse
Okular
Zsh
R ...the audience!
30 August 2018 Improving Reproducibility and Reuse of Modelling Results in the Life Sciences | Martin Scharm 21