DisGeNET: A discovery platform for the dynamical exploration of human disease...Núria Queralt Rosinach
The document describes DisGeNET, a discovery platform for exploring gene-disease associations through integrated data sources including expert-curated databases and text mining of biomedical literature. DisGeNET contains over 17,000 genes and 14,000 diseases with 429,000 associations, integrating information from various sources through normalization. It provides tools for users to explore gene-disease relationships and supports biomedical research.
Making it Easier, Possibly Even Pleasant, to Author Rich Experimental MetadataMichel Dumontier
Biomedical researchers will remain stymied in their ability to take full advantage of the Big Data revolution if they can never find the datasets that they need to analyze, if there is lack of clarity about what particular datasets contain, and if data are insufficiently described.
CEDAR, an NIH BD2K Center of Excellence, aims to develop methods and tools to vastly ease the burden of authoring good experimental metadata, and to maximally use this information to zero in on datasets of interest.
The DisGeNET R package (disgenet2r) allows users to query and analyze gene-disease association data from the DisGeNET knowledge platform within the R environment. Disgenet2r provides functions to retrieve associations between genes and diseases, variants and diseases, and disease-disease relationships. It also enables visualization of results as networks and heatmaps. The package integrates with other R/Bioconductor packages and can be used for knowledge discovery and generating hypotheses to study gene-disease relationships and support precision medicine.
dkNET is a portal that provides researchers access to diverse research resources and tools to improve rigor and reproducibility. It supports the use of Research Resource Identifiers (RRIDs) to properly identify research tools in publications. dkNET also provides Resource Reports that give detailed information on tools, their usage metrics, ratings, and alerts about potential issues. Additionally, it offers services like Reproducibility Reports and a Hypothesis Center to facilitate rigorous research.
Powering Scientific Discovery with the Semantic Web (VanBUG 2014)Michel Dumontier
The document discusses how the semantic web can help power scientific discovery. It proposes building a massive network of interconnected data and software using web standards to 1) generate and test hypotheses by discovering associations in the data, 2) gather evidence to support or dispute hypotheses, and 3) contribute new knowledge back to the global network. This network, called the semantic web, treats data as a web of facts that can be shared and queried using semantic web standards. The document provides examples of how linked open data in the life sciences is being created and used via semantic web technologies to integrate data from multiple sources and answer complex queries.
Access to consistent, high-quality metadata is critical to finding, understanding, and reusing scientific data. This document describes a consensus among participating stakeholders in the Health Care and the Life Sciences domain on the description of datasets using the Resource Description Framework (RDF). This specification meets key functional requirements, reuses existing vocabularies to the extent that it is possible, and addresses elements of data description, versioning, provenance, discovery, exchange, query, and retrieval.
Citing data in research articles: principles, implementation, challenges - an...FAIRDOM
Prepared and presented by Jo McEntyre (EMBL_EBI) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
Making Data FAIR (Findable, Accessible, Interoperable, Reusable)Tom Plasterer
What to do About FAIR…
In the experience of most pharma professionals, FAIR remains fairly abstract, bordering on inconclusive. This session will outline specific case studies – real problems with real data, and address opportunities and real concerns.
·
Why making data Findable, Actionable, Interoperable and Reusable is important.
Talk presented at the Data Driven Drug Development (D4) conference on March 20th, 2019.
DisGeNET: A discovery platform for the dynamical exploration of human disease...Núria Queralt Rosinach
The document describes DisGeNET, a discovery platform for exploring gene-disease associations through integrated data sources including expert-curated databases and text mining of biomedical literature. DisGeNET contains over 17,000 genes and 14,000 diseases with 429,000 associations, integrating information from various sources through normalization. It provides tools for users to explore gene-disease relationships and supports biomedical research.
Making it Easier, Possibly Even Pleasant, to Author Rich Experimental MetadataMichel Dumontier
Biomedical researchers will remain stymied in their ability to take full advantage of the Big Data revolution if they can never find the datasets that they need to analyze, if there is lack of clarity about what particular datasets contain, and if data are insufficiently described.
CEDAR, an NIH BD2K Center of Excellence, aims to develop methods and tools to vastly ease the burden of authoring good experimental metadata, and to maximally use this information to zero in on datasets of interest.
The DisGeNET R package (disgenet2r) allows users to query and analyze gene-disease association data from the DisGeNET knowledge platform within the R environment. Disgenet2r provides functions to retrieve associations between genes and diseases, variants and diseases, and disease-disease relationships. It also enables visualization of results as networks and heatmaps. The package integrates with other R/Bioconductor packages and can be used for knowledge discovery and generating hypotheses to study gene-disease relationships and support precision medicine.
dkNET is a portal that provides researchers access to diverse research resources and tools to improve rigor and reproducibility. It supports the use of Research Resource Identifiers (RRIDs) to properly identify research tools in publications. dkNET also provides Resource Reports that give detailed information on tools, their usage metrics, ratings, and alerts about potential issues. Additionally, it offers services like Reproducibility Reports and a Hypothesis Center to facilitate rigorous research.
Powering Scientific Discovery with the Semantic Web (VanBUG 2014)Michel Dumontier
The document discusses how the semantic web can help power scientific discovery. It proposes building a massive network of interconnected data and software using web standards to 1) generate and test hypotheses by discovering associations in the data, 2) gather evidence to support or dispute hypotheses, and 3) contribute new knowledge back to the global network. This network, called the semantic web, treats data as a web of facts that can be shared and queried using semantic web standards. The document provides examples of how linked open data in the life sciences is being created and used via semantic web technologies to integrate data from multiple sources and answer complex queries.
Access to consistent, high-quality metadata is critical to finding, understanding, and reusing scientific data. This document describes a consensus among participating stakeholders in the Health Care and the Life Sciences domain on the description of datasets using the Resource Description Framework (RDF). This specification meets key functional requirements, reuses existing vocabularies to the extent that it is possible, and addresses elements of data description, versioning, provenance, discovery, exchange, query, and retrieval.
Citing data in research articles: principles, implementation, challenges - an...FAIRDOM
Prepared and presented by Jo McEntyre (EMBL_EBI) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
Making Data FAIR (Findable, Accessible, Interoperable, Reusable)Tom Plasterer
What to do About FAIR…
In the experience of most pharma professionals, FAIR remains fairly abstract, bordering on inconclusive. This session will outline specific case studies – real problems with real data, and address opportunities and real concerns.
·
Why making data Findable, Actionable, Interoperable and Reusable is important.
Talk presented at the Data Driven Drug Development (D4) conference on March 20th, 2019.
The document outlines plans to transition the cBioPortal cancer genomics platform to an open source model with coordinated development between Memorial Sloan Kettering Cancer Center, Dana-Farber Cancer Institute, and Princess Margaret Cancer Centre. It discusses expanding usage, new features, funding options, and establishing an advisory committee. The goal is to build a sustainable open source community through collaborative development, additional funding, and engagement with users and potential contributors.
Semantic web technologies offer a potential mechanism for the representation and integration of thousands of biomedical databases. Many of these databases offer cross-references to other data sources, but these are generally incomplete and prone to error. In this paper, we conduct an empirical analysis of the link structure of life science Linked Data, obtained from the Bio2RDF project. Three different link graphs for datasets, entities and terms are characterized by degree, connectivity, and clustering metrics, and their correlation is measured as well. Furthermore, we utilize the symmetry and transitivity of entity links to build a benchmark and evaluate several popular entity matching approaches. Our findings indicate that the life science data network can help find hidden links, can be used to validate links, and may offer a mechanism to integrate a wider set of resources to support biomedical knowledge discovery.
This document discusses opportunities for using the open source cBioPortal platform in a commercial setting. It summarizes The Hyve's experiences supporting cBioPortal for the Center for Translational Molecular Medicine's TraIT project. The Hyve provides professional support for open source bioinformatics software like cBioPortal through software development, data services, consultancy, and hosting. For translation projects, The Hyve employs a phased approach including definition, pilot, implementation, and evaluation phases to implement cBioPortal and demonstrate its capabilities for data integration and analysis.
Generating Biomedical Hypotheses Using Semantic Web TechnologiesMichel Dumontier
With its focus on investigating the nature and basis for the sustained existence of living systems, modern biology has always been a fertile, if not challenging, domain for formal knowledge representation and automated reasoning. Over the past 15 years, hundreds of projects have developed or leveraged ontologies for entity recognition and relation extraction, semantic annotation, data integration, query answering, consistency checking, association mining and other forms of knowledge discovery. In this talk, I will discuss our efforts to build a rich foundational network of ontology-annotated linked data, discover significant biological associations across these data using a set of partially overlapping ontologies, and identify new avenues for drug discovery by applying measures of semantic similarity over phenotypic descriptions. As the portfolio of Semantic Web technologies continue to mature in terms of functionality, scalability and an understanding of how to maximize their value, increasing numbers of biomedical researchers will be strategically poised to pursue increasingly sophisticated KR projects aimed at improving our overall understanding of the capability and behavior of biological systems.
Beacon Network: A System for Global Genomic Data SharingMiro Cupak
The Beacon Network provides a system for global genomic data sharing by allowing users to query a network of genetic data sources to determine if a particular genetic variant or mutation exists in their databases. It began as a web service called Beacon that responds with "yes" or "no" to questions about genetic mutations. The Beacon Network expands this by distributing queries across multiple beacons and aggregating the results. It currently includes over 25 genomic organizations with access to over 2 million samples and 2 billion genetic variants, serving hundreds of thousands of queries from users around the world. The goal is to facilitate discovery of new links between genetic data and health conditions.
Beacon Network: A System for Global Genomic Data SharingMiro Cupak
A global federated network called the Beacon Network was created to share genomic data in order to drive discoveries and applications in medicine. The Beacon Network allows users to query beacons from various genomic organizations to discover if they have information about specific genetic mutations. It translates queries and intelligently distributes them across over 60 beacons from 25 organizations. Since launching, the Beacon Network has served over 400,000 queries from users in over 100 countries, resulting in over 2 million queries to participants of the network.
Beacon: A Protocol for Federated Discovery and Sharing of Genomic DataMiro Cupak
The document summarizes Beacon, a protocol for federated discovery and sharing of genomic data across institutions. It allows institutions to share whether they have information on specific genetic mutations through a standardized web service API. Over 25 genomic organizations representing over 160 datasets totaling over 2 million queries participate in the Beacon Network, which searches across participating beacons and aggregates the results.
dkNET Webinar: Creating and Sustaining a FAIR Biomedical Data Ecosystem 10/09...dkNET
Abstract
In this presentation, Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health, will share the NIH’s vision for a modernized, integrated FAIR biomedical data ecosystem and the strategic roadmap that NIH is following to achieve this vision. Dr. Gregurick will highlight projects being implemented by team members across the NIH’s 27 institutes and centers and will ways that industry, academia, and other communities can help NIH enable a FAIR data ecosystem. Finally, she will weave in how this strategy is being leveraged to address the COVID-19 pandemic.
Presenter: Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health
dkNET Webinar Information: https://dknet.org/about/webinar
dkNET Webinar: "The Microphysiology Systems Database (MPS-Db): A Platform For...dkNET
This document discusses harnessing quantitative systems pharmacology to deliver personalized medicine through the Microphysiology Systems Database (MPS-Db). The MPS-Db is presented as a solution for aggregating, analyzing, sharing, and modeling in vitro data from microphysiology systems (MPS) to accelerate drug development and implementation of personalized medicine approaches. Key features and capabilities of the MPS-Db are described, including supporting various model types, integrating with other databases, and performing data analysis, reproducibility evaluation, and computational modeling. Commercial and non-profit versions are discussed.
Reproducible and citable data and models: an introduction.FAIRDOM
Prepared and presented by Carole Goble (University of Manchester), Wolfgang Mueller (HITS), Dagmar Waltermath (University of Rostock), at the Reproducible and Citable Data and Models Workshop, Warnemünde, Germany. September 14th - 16th 2015.
2015-02-10 The Open PHACTS Discovery Platform: Semantic Data Integration for ...open_phacts
The Open PHACTS Discovery Platform integrates multiple biomedical data resources into a single open access point using semantic web technology. It is guided by business questions from pharmaceutical companies to integrate data from sources like ChEMBL, DrugBank, UniProt, and more. The platform is run as a public-private partnership through 2021 to support drug discovery.
FAIR data and model management for systems biology.FAIRDOM
Written and presented by Carole Goble (University of Manchester) as part of Intelligent Systems for Molecular Biology (ISMB), Dublin. July 10th - 14th 2015.
FAIR Data and Model Management for Systems Biology(and SOPs too!)Carole Goble
MultiScale Biology Network Springboard meeting, Nottingham, UK, 1 June 2015
FAIR Data and model management for Systems Biology
Over the past 5 years we have seen a change in expectations for the management of all the outcomes of research – that is the “assets” of data, models, codes, SOPs and so forth. Don’t stop reading. Yes, data management isn’t likely to win anyone a Nobel prize. But publications should be supported and accompanied by data, methods, procedures, etc. to assure reproducibility of results. Funding agencies expect data (and increasingly software) management retention and access plans as part of the proposal process for projects to be funded. Journals are raising their expectations of the availability of data and codes for pre- and post- publication. And the multi-component, multi-disciplinary nature of Systems Biology demands the interlinking and exchange of assets and the systematic recording of metadata for their interpretation.
Data and model management for the Systems Biology community is a multi-faceted one including: the development and adoption appropriate community standards (and the navigation of the standards maze); the sustaining of international public archives capable of servicing quantitative biology; and the development of the necessary tools and know-how for researchers within their own institutes so that they can steward their assets in a sustainable, coherent and credited manner while minimizing burden and maximising personal benefit.
The FAIRDOM (Findable, Accessible, Interoperable, Reusable Data, Operations and Models) Initiative has grown out of several efforts in European programmes (SysMO and EraSysAPP ERANets and the ISBE ESRFI) and national initiatives (de.NBI, German Virtual Liver Network, SystemsX, UK SynBio centres). It aims to support Systems Biology researchers with data and model management, with an emphasis on standards smuggled in by stealth.
This talk will use the FAIRDOM Initiative to discuss the FAIR management of data, SOPs, and models for Sys Bio, highlighting the challenges multi-scale biology presents.
http://www.fair-dom.org
http://www.fairdomhub.org
http://www.seek4science.org
This document introduces FAIRDOM, a consortium that provides a platform and services to help researchers organize, manage, share, and preserve research outputs according to FAIR principles. FAIRDOM has been in operation for 10 years and has over 50 installations supporting over 118 projects. It provides tools and services to help researchers collaborate better and integrate their data, models, publications and other research objects. FAIRDOM also works with other organizations and infrastructure providers to support broader research initiatives.
Written and presented by Carole Goble (University of Manchester) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
This document summarizes the BioAssay Research Database (BARD), a public database developed to provide access to bioassay data from the NIH Molecular Libraries Program (MLP). BARD has curated and migrated data from over 600 MLP projects, standardizing the metadata using a controlled vocabulary. This allows for systematic cross-assay analysis. BARD supports data depositors, data miners accessing and querying the database, and software developers building new tools using the BARD application programming interface.
FAIRy stories: tales from building the FAIR Research CommonsCarole Goble
Plenary Lecture Presented at INCF Neuroinformatics 2019 https://www.neuroinformatics2019.org
Title: FAIRy stories: tales from building the FAIR Research Commons
Findable Accessable Interoperable Reusable. The “FAIR Principles” for research data, software, computational workflows, scripts, or any kind of Research Object is a mantra; a method; a meme; a myth; a mystery. For the past 15 years I have been working on FAIR in a range of projects and initiatives in the Life Sciences as we try to build the FAIR Research Commons. Some are top-down like the European Research Infrastructures ELIXIR, ISBE and IBISBA, and the NIH Data Commons. Some are bottom-up, supporting FAIR for investigator-led projects (FAIRDOM), biodiversity analytics (BioVel), and FAIR drug discovery (Open PHACTS, FAIRplus). Some have become movements, like Bioschemas, the Common Workflow Language and Research Objects. Others focus on cross-cutting approaches in reproducibility, computational workflows, metadata representation and scholarly sharing & publication. In this talk I will relate a series of FAIRy tales. Some of them are Grimm. There are villains and heroes. Some have happy endings; all have morals.
COMBINE 2019, EU-STANDS4PM, Heidelberg, Germany 18 July 2019
FAIR: Findable Accessable Interoperable Reusable. The “FAIR Principles” for research data, software, computational workflows, scripts, or any other kind of Research Object one can think of, is now a mantra; a method; a meme; a myth; a mystery. FAIR is about supporting and tracking the flow and availability of data across research organisations and the portability and sustainability of processing methods to enable transparent and reproducible results. All this is within the context of a bottom up society of collaborating (or burdened?) scientists, a top down collective of compliance-focused funders and policy makers and an in-the-middle posse of e-infrastructure providers.
Making the FAIR principles a reality is tricky. They are aspirations not standards. They are multi-dimensional and dependent on context such as the sensitivity and availability of the data and methods. We already see a jungle of projects, initiatives and programmes wrestling with the challenges. FAIR efforts have particularly focused on the “last mile” – “FAIRifying” destination community archive repositories and measuring their “compliance” to FAIR metrics (or less controversially “indicators”). But what about FAIR at the first mile, at source and how do we help Alice and Bob with their (secure) data management? If we tackle the FAIR first and last mile, what about the FAIR middle? What about FAIR beyond just data – like exchanging and reusing pipelines for precision medicine?
Since 2008 the FAIRDOM collaboration [1] has worked on FAIR asset management and the development of a FAIR asset Commons for multi-partner researcher projects [2], initially in the Systems Biology field. Since 2016 we have been working with the BioCompute Object Partnership [3] on standardising computational records of HTS precision medicine pipelines.
So, using our FAIRDOM and BioCompute Object binoculars let’s go on a FAIR safari! Let’s peruse the ecosystem, observe the different herds and reflect what where we are for FAIR personalised medicine.
References
[1] http://www.fair-dom.org
[2] http://www.fairdomhub.org
[3] http://www.biocomputeobject.org
How are we Faring with FAIR? (and what FAIR is not)Carole Goble
Keynote presented at the workshop FAIRe Data Infrastructures, 15 October 2020
https://www.gmds.de/aktivitaeten/medizinische-informatik/projektgruppenseiten/faire-dateninfrastrukturen-fuer-die-biomedizinische-informatik/workshop-2020/
Remarkably it was only in 2016 that the ‘FAIR Guiding Principles for scientific data management and stewardship’ appeared in Scientific Data. The paper was intended to launch a dialogue within the research and policy communities: to start a journey to wider accessibility and reusability of data and prepare for automation-readiness by supporting findability, accessibility, interoperability and reusability for machines. Many of the authors (including myself) came from biomedical and associated communities. The paper succeeded in its aim, at least at the policy, enterprise and professional data infrastructure level. Whether FAIR has impacted the researcher at the bench or bedside is open to doubt. It certainly inspired a great deal of activity, many projects, a lot of positioning of interests and raised awareness. COVID has injected impetus and urgency to the FAIR cause (good) and also highlighted its politicisation (not so good).
In this talk I’ll make some personal reflections on how we are faring with FAIR: as one of the original principles authors; as a participant in many current FAIR initiatives (particularly in the biomedical sector and for research objects other than data) and as a veteran of FAIR before we had the principles.
- The document discusses Harry Hochheiser's research in translational bioinformatics and challenges in data sharing.
- It describes the FaceBase project, which aims to compile biological data related to craniofacial development across multiple organisms and datasets.
- Effective data sharing is challenging due to the diversity of data types and projects involved; metadata and ontologies could help but have not been fully leveraged.
tranSMART Community Meeting 5-7 Nov 13 - Session 1: Translational Drug Disco...David Peyruc
This document summarizes Andy Plump's presentation on translational drug discovery at Sanofi. It discusses two pillars of Sanofi's strategy: translational medicine and open innovation. Translational medicine focuses on human genetics, biology and disease to select targets and design clinical trials, moving from patients to research and back. Four success stories are highlighted: PCSK9 for heart disease, TrkA for pain, P53 for cancer, and glycolipids for Gaucher's disease. The presentation emphasizes applying lessons from human genetics and biology throughout the drug development process.
The document outlines plans to transition the cBioPortal cancer genomics platform to an open source model with coordinated development between Memorial Sloan Kettering Cancer Center, Dana-Farber Cancer Institute, and Princess Margaret Cancer Centre. It discusses expanding usage, new features, funding options, and establishing an advisory committee. The goal is to build a sustainable open source community through collaborative development, additional funding, and engagement with users and potential contributors.
Semantic web technologies offer a potential mechanism for the representation and integration of thousands of biomedical databases. Many of these databases offer cross-references to other data sources, but these are generally incomplete and prone to error. In this paper, we conduct an empirical analysis of the link structure of life science Linked Data, obtained from the Bio2RDF project. Three different link graphs for datasets, entities and terms are characterized by degree, connectivity, and clustering metrics, and their correlation is measured as well. Furthermore, we utilize the symmetry and transitivity of entity links to build a benchmark and evaluate several popular entity matching approaches. Our findings indicate that the life science data network can help find hidden links, can be used to validate links, and may offer a mechanism to integrate a wider set of resources to support biomedical knowledge discovery.
This document discusses opportunities for using the open source cBioPortal platform in a commercial setting. It summarizes The Hyve's experiences supporting cBioPortal for the Center for Translational Molecular Medicine's TraIT project. The Hyve provides professional support for open source bioinformatics software like cBioPortal through software development, data services, consultancy, and hosting. For translation projects, The Hyve employs a phased approach including definition, pilot, implementation, and evaluation phases to implement cBioPortal and demonstrate its capabilities for data integration and analysis.
Generating Biomedical Hypotheses Using Semantic Web TechnologiesMichel Dumontier
With its focus on investigating the nature and basis for the sustained existence of living systems, modern biology has always been a fertile, if not challenging, domain for formal knowledge representation and automated reasoning. Over the past 15 years, hundreds of projects have developed or leveraged ontologies for entity recognition and relation extraction, semantic annotation, data integration, query answering, consistency checking, association mining and other forms of knowledge discovery. In this talk, I will discuss our efforts to build a rich foundational network of ontology-annotated linked data, discover significant biological associations across these data using a set of partially overlapping ontologies, and identify new avenues for drug discovery by applying measures of semantic similarity over phenotypic descriptions. As the portfolio of Semantic Web technologies continue to mature in terms of functionality, scalability and an understanding of how to maximize their value, increasing numbers of biomedical researchers will be strategically poised to pursue increasingly sophisticated KR projects aimed at improving our overall understanding of the capability and behavior of biological systems.
Beacon Network: A System for Global Genomic Data SharingMiro Cupak
The Beacon Network provides a system for global genomic data sharing by allowing users to query a network of genetic data sources to determine if a particular genetic variant or mutation exists in their databases. It began as a web service called Beacon that responds with "yes" or "no" to questions about genetic mutations. The Beacon Network expands this by distributing queries across multiple beacons and aggregating the results. It currently includes over 25 genomic organizations with access to over 2 million samples and 2 billion genetic variants, serving hundreds of thousands of queries from users around the world. The goal is to facilitate discovery of new links between genetic data and health conditions.
Beacon Network: A System for Global Genomic Data SharingMiro Cupak
A global federated network called the Beacon Network was created to share genomic data in order to drive discoveries and applications in medicine. The Beacon Network allows users to query beacons from various genomic organizations to discover if they have information about specific genetic mutations. It translates queries and intelligently distributes them across over 60 beacons from 25 organizations. Since launching, the Beacon Network has served over 400,000 queries from users in over 100 countries, resulting in over 2 million queries to participants of the network.
Beacon: A Protocol for Federated Discovery and Sharing of Genomic DataMiro Cupak
The document summarizes Beacon, a protocol for federated discovery and sharing of genomic data across institutions. It allows institutions to share whether they have information on specific genetic mutations through a standardized web service API. Over 25 genomic organizations representing over 160 datasets totaling over 2 million queries participate in the Beacon Network, which searches across participating beacons and aggregates the results.
dkNET Webinar: Creating and Sustaining a FAIR Biomedical Data Ecosystem 10/09...dkNET
Abstract
In this presentation, Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health, will share the NIH’s vision for a modernized, integrated FAIR biomedical data ecosystem and the strategic roadmap that NIH is following to achieve this vision. Dr. Gregurick will highlight projects being implemented by team members across the NIH’s 27 institutes and centers and will ways that industry, academia, and other communities can help NIH enable a FAIR data ecosystem. Finally, she will weave in how this strategy is being leveraged to address the COVID-19 pandemic.
Presenter: Susan Gregurick, Ph.D., Associate Director of Data Science and Director, Office of Data Science Strategy at the National Institutes of Health
dkNET Webinar Information: https://dknet.org/about/webinar
dkNET Webinar: "The Microphysiology Systems Database (MPS-Db): A Platform For...dkNET
This document discusses harnessing quantitative systems pharmacology to deliver personalized medicine through the Microphysiology Systems Database (MPS-Db). The MPS-Db is presented as a solution for aggregating, analyzing, sharing, and modeling in vitro data from microphysiology systems (MPS) to accelerate drug development and implementation of personalized medicine approaches. Key features and capabilities of the MPS-Db are described, including supporting various model types, integrating with other databases, and performing data analysis, reproducibility evaluation, and computational modeling. Commercial and non-profit versions are discussed.
Reproducible and citable data and models: an introduction.FAIRDOM
Prepared and presented by Carole Goble (University of Manchester), Wolfgang Mueller (HITS), Dagmar Waltermath (University of Rostock), at the Reproducible and Citable Data and Models Workshop, Warnemünde, Germany. September 14th - 16th 2015.
2015-02-10 The Open PHACTS Discovery Platform: Semantic Data Integration for ...open_phacts
The Open PHACTS Discovery Platform integrates multiple biomedical data resources into a single open access point using semantic web technology. It is guided by business questions from pharmaceutical companies to integrate data from sources like ChEMBL, DrugBank, UniProt, and more. The platform is run as a public-private partnership through 2021 to support drug discovery.
FAIR data and model management for systems biology.FAIRDOM
Written and presented by Carole Goble (University of Manchester) as part of Intelligent Systems for Molecular Biology (ISMB), Dublin. July 10th - 14th 2015.
FAIR Data and Model Management for Systems Biology(and SOPs too!)Carole Goble
MultiScale Biology Network Springboard meeting, Nottingham, UK, 1 June 2015
FAIR Data and model management for Systems Biology
Over the past 5 years we have seen a change in expectations for the management of all the outcomes of research – that is the “assets” of data, models, codes, SOPs and so forth. Don’t stop reading. Yes, data management isn’t likely to win anyone a Nobel prize. But publications should be supported and accompanied by data, methods, procedures, etc. to assure reproducibility of results. Funding agencies expect data (and increasingly software) management retention and access plans as part of the proposal process for projects to be funded. Journals are raising their expectations of the availability of data and codes for pre- and post- publication. And the multi-component, multi-disciplinary nature of Systems Biology demands the interlinking and exchange of assets and the systematic recording of metadata for their interpretation.
Data and model management for the Systems Biology community is a multi-faceted one including: the development and adoption appropriate community standards (and the navigation of the standards maze); the sustaining of international public archives capable of servicing quantitative biology; and the development of the necessary tools and know-how for researchers within their own institutes so that they can steward their assets in a sustainable, coherent and credited manner while minimizing burden and maximising personal benefit.
The FAIRDOM (Findable, Accessible, Interoperable, Reusable Data, Operations and Models) Initiative has grown out of several efforts in European programmes (SysMO and EraSysAPP ERANets and the ISBE ESRFI) and national initiatives (de.NBI, German Virtual Liver Network, SystemsX, UK SynBio centres). It aims to support Systems Biology researchers with data and model management, with an emphasis on standards smuggled in by stealth.
This talk will use the FAIRDOM Initiative to discuss the FAIR management of data, SOPs, and models for Sys Bio, highlighting the challenges multi-scale biology presents.
http://www.fair-dom.org
http://www.fairdomhub.org
http://www.seek4science.org
This document introduces FAIRDOM, a consortium that provides a platform and services to help researchers organize, manage, share, and preserve research outputs according to FAIR principles. FAIRDOM has been in operation for 10 years and has over 50 installations supporting over 118 projects. It provides tools and services to help researchers collaborate better and integrate their data, models, publications and other research objects. FAIRDOM also works with other organizations and infrastructure providers to support broader research initiatives.
Written and presented by Carole Goble (University of Manchester) as part of the Reproducible and Citable Data and Models Workshop in Warnemünde, Germany. September 14th - 16th 2015.
This document summarizes the BioAssay Research Database (BARD), a public database developed to provide access to bioassay data from the NIH Molecular Libraries Program (MLP). BARD has curated and migrated data from over 600 MLP projects, standardizing the metadata using a controlled vocabulary. This allows for systematic cross-assay analysis. BARD supports data depositors, data miners accessing and querying the database, and software developers building new tools using the BARD application programming interface.
FAIRy stories: tales from building the FAIR Research CommonsCarole Goble
Plenary Lecture Presented at INCF Neuroinformatics 2019 https://www.neuroinformatics2019.org
Title: FAIRy stories: tales from building the FAIR Research Commons
Findable Accessable Interoperable Reusable. The “FAIR Principles” for research data, software, computational workflows, scripts, or any kind of Research Object is a mantra; a method; a meme; a myth; a mystery. For the past 15 years I have been working on FAIR in a range of projects and initiatives in the Life Sciences as we try to build the FAIR Research Commons. Some are top-down like the European Research Infrastructures ELIXIR, ISBE and IBISBA, and the NIH Data Commons. Some are bottom-up, supporting FAIR for investigator-led projects (FAIRDOM), biodiversity analytics (BioVel), and FAIR drug discovery (Open PHACTS, FAIRplus). Some have become movements, like Bioschemas, the Common Workflow Language and Research Objects. Others focus on cross-cutting approaches in reproducibility, computational workflows, metadata representation and scholarly sharing & publication. In this talk I will relate a series of FAIRy tales. Some of them are Grimm. There are villains and heroes. Some have happy endings; all have morals.
COMBINE 2019, EU-STANDS4PM, Heidelberg, Germany 18 July 2019
FAIR: Findable Accessable Interoperable Reusable. The “FAIR Principles” for research data, software, computational workflows, scripts, or any other kind of Research Object one can think of, is now a mantra; a method; a meme; a myth; a mystery. FAIR is about supporting and tracking the flow and availability of data across research organisations and the portability and sustainability of processing methods to enable transparent and reproducible results. All this is within the context of a bottom up society of collaborating (or burdened?) scientists, a top down collective of compliance-focused funders and policy makers and an in-the-middle posse of e-infrastructure providers.
Making the FAIR principles a reality is tricky. They are aspirations not standards. They are multi-dimensional and dependent on context such as the sensitivity and availability of the data and methods. We already see a jungle of projects, initiatives and programmes wrestling with the challenges. FAIR efforts have particularly focused on the “last mile” – “FAIRifying” destination community archive repositories and measuring their “compliance” to FAIR metrics (or less controversially “indicators”). But what about FAIR at the first mile, at source and how do we help Alice and Bob with their (secure) data management? If we tackle the FAIR first and last mile, what about the FAIR middle? What about FAIR beyond just data – like exchanging and reusing pipelines for precision medicine?
Since 2008 the FAIRDOM collaboration [1] has worked on FAIR asset management and the development of a FAIR asset Commons for multi-partner researcher projects [2], initially in the Systems Biology field. Since 2016 we have been working with the BioCompute Object Partnership [3] on standardising computational records of HTS precision medicine pipelines.
So, using our FAIRDOM and BioCompute Object binoculars let’s go on a FAIR safari! Let’s peruse the ecosystem, observe the different herds and reflect what where we are for FAIR personalised medicine.
References
[1] http://www.fair-dom.org
[2] http://www.fairdomhub.org
[3] http://www.biocomputeobject.org
How are we Faring with FAIR? (and what FAIR is not)Carole Goble
Keynote presented at the workshop FAIRe Data Infrastructures, 15 October 2020
https://www.gmds.de/aktivitaeten/medizinische-informatik/projektgruppenseiten/faire-dateninfrastrukturen-fuer-die-biomedizinische-informatik/workshop-2020/
Remarkably it was only in 2016 that the ‘FAIR Guiding Principles for scientific data management and stewardship’ appeared in Scientific Data. The paper was intended to launch a dialogue within the research and policy communities: to start a journey to wider accessibility and reusability of data and prepare for automation-readiness by supporting findability, accessibility, interoperability and reusability for machines. Many of the authors (including myself) came from biomedical and associated communities. The paper succeeded in its aim, at least at the policy, enterprise and professional data infrastructure level. Whether FAIR has impacted the researcher at the bench or bedside is open to doubt. It certainly inspired a great deal of activity, many projects, a lot of positioning of interests and raised awareness. COVID has injected impetus and urgency to the FAIR cause (good) and also highlighted its politicisation (not so good).
In this talk I’ll make some personal reflections on how we are faring with FAIR: as one of the original principles authors; as a participant in many current FAIR initiatives (particularly in the biomedical sector and for research objects other than data) and as a veteran of FAIR before we had the principles.
- The document discusses Harry Hochheiser's research in translational bioinformatics and challenges in data sharing.
- It describes the FaceBase project, which aims to compile biological data related to craniofacial development across multiple organisms and datasets.
- Effective data sharing is challenging due to the diversity of data types and projects involved; metadata and ontologies could help but have not been fully leveraged.
tranSMART Community Meeting 5-7 Nov 13 - Session 1: Translational Drug Disco...David Peyruc
This document summarizes Andy Plump's presentation on translational drug discovery at Sanofi. It discusses two pillars of Sanofi's strategy: translational medicine and open innovation. Translational medicine focuses on human genetics, biology and disease to select targets and design clinical trials, moving from patients to research and back. Four success stories are highlighted: PCSK9 for heart disease, TrkA for pain, P53 for cancer, and glycolipids for Gaucher's disease. The presentation emphasizes applying lessons from human genetics and biology throughout the drug development process.
This document outlines the planning and resources for a digital graphic narrative project. It considers costs, available resources like computers and software, audience, quality factors, codes of practice, regulation, copyright, ethics, and health and safety. A production schedule is included that breaks the project into 10 half-day sessions focused on creating backgrounds, characters, and assembling pages. Risks like eyestrain are addressed by taking regular breaks, and tripping hazards are prevented by taping down wires.
Se describe un nuevo método y producto llamado MOHOSTOP TM para prevenir el deterioro visual de materias primas alimenticias causado por el crecimiento de hongos, mohos y bacterias. El método implica la formulación de una mezcla de ácido aminoacético, ácido acético glacial y 3,4-dihidroxifenilacético, los cuales se aplicarían a los alimentos para neutralizar dichos microorganismos. El método podría usarse en alimentos crudos, cocinados u otros procesados para extender su vida útil.
Los periféricos son dispositivos auxiliares conectados a la unidad central de procesamiento de una computadora. Los periféricos de entrada proporcionan datos y señales de control al sistema, como teclados y escáneres. Los periféricos de salida muestran los resultados de las operaciones del computador al usuario, como monitores e impresoras. Los periféricos de entrada y salida sirven para comunicar la computadora con el exterior y transferir información en ambas direcciones.
Establishing validity, reproducibility, and utility of highly scalable geneti...Human Variome Project
Background: New technologies and increased competition have, and will continue to improve the cost-effectiveness of genetic testing, making genetic analysis more accessible to medical practices worldwide. However, challenges remain to establishing the validity of such tests. Moreover many patients harbor rare or novel variants and classification is likely to remain a bottleneck in broader deployment of genetic medicine.
The document reports statistics from a study of 6893 patients. It found that 247 patients, or less than 20% of the total, had a condition called "KD". It also found that 7544 patients had a different condition described as "______", while 6744 patients had a third condition described as "______". The document does not provide any other details about the conditions, patients, or purpose of collecting these statistics.
Este documento describe un producto llamado MohoStop TM Fórmula panificación que ayuda a prevenir la aparición de hongos y mohos en productos de panificación y repostería. El producto utiliza ingredientes naturales y orgánicos clasificados como seguros por autoridades alimentarias. Ayuda a prolongar la vida útil de los productos, mantener su sabor y aspecto, y protege la marca del negocio al reducir devoluciones y reclamaciones. La compañía ND Pharma & Biotech se enfoca en ofrecer soluciones
Este decreto reforma diversas disposiciones de la Ley General de Educación de México. Las reformas incluyen garantizar la educación de calidad para toda la población, establecer que la educación impartida por el Estado será gratuita, y prohibir el pago de cuotas que condicionen o impidan la prestación del servicio educativo. También se reforman artículos relacionados con los principios orientadores de la educación, las instituciones y autoridades educativas, y sus funciones.
The Controlled Natural Language of Randall Munroe’s Thing Explainer Tobias Kuhn
It is rare that texts or entire books written in a Controlled Natural Language (CNL) become very popular, but exactly this has happened with a book that has been published last year. Randall Munroe's Thing Explainer uses only the 1'000 most often used words of the English language together with drawn pictures to explain complicated things such as nuclear reactors, jet engines, the solar system, and dishwashers. This restricted language is a very interesting new case for the CNL community. I describe here its place in the context of existing approaches on Controlled Natural Languages, and I provide a first analysis from a scientific perspective, covering the word production rules and word distributions.
Current corporate analysis (aviation and automobile industry)1Subhashree Mishra
The document provides an overview of the aviation and automobile industries in India. For aviation, it notes that India is the 9th largest aviation market and is experiencing growth driven by factors such as low-cost carriers and expanding airports. The automobile industry overview highlights that India is one of the largest auto producers in the world, led by two-wheelers which have an 81% market share. It also discusses the growth opportunities in both industries as the economy and disposable incomes rise in India.
This short document promotes the creation of presentations using Haiku Deck, an online presentation tool. It includes three stock photos without captions or additional context. The document encourages the viewer to get started making their own Haiku Deck presentation and sharing it on SlideShare.
Albert Kamel Said Shamroukh is seeking a position as a chemist utilizing his educational background which includes bachelor's and master's degrees in chemistry, physics, and nuclear physics from Alexandria University in Egypt. He has over 10 years of experience in quality control and analytical chemistry roles for pharmaceutical companies in Egypt and New Jersey, performing tasks like HPLC, GC, UV spectroscopy, dissolution testing, and calibration according to cGMP standards. He also has teaching experience and is bilingual in English and Arabic.
Iniciativa con proyecto de ley del programa de derechos humanos para el estad...ricardomejiaberdeja
Los suscritos Diputados Ricardo Mejía Berdeja, Silvano Blanco Deaquino y Magdalena Camacho Díaz, Integrantes de la Fracción Parlamentaria de Movimiento Ciudadano de la Sexagésima Primera Legislatura al Honorable Congreso del Estado Libre y Soberano de Guerrero, en uso de las facultades que nos confieren los artículos 65 fracción I de la Constitución Política del Estado Libre y Soberano de Guerrero, 126 fracción II, 127 párrafo tercero, y demás aplicables de la Ley Orgánica del Poder Legislativo del Estado de Guerrero Número 286, ponemos a la consideración de esta Soberanía Popular, la presente Iniciativa con Proyecto de Ley del Programa de Derechos Humanos para el Estado de Guerrero, al tenor de la siguiente
This document summarizes a presentation on new sources of big data for precision medicine. It discusses how new data sources like genomics, the human microbiome, epigenomics, and the exposome are generating large amounts of data. It then covers the evolution of precision medicine from concepts like personalized medicine and how strategic initiatives in the UK and US are supporting precision medicine research through funding programs and projects like the Cancer Genome Atlas, eMERGE, and exposome studies. The presentation raises the question of whether we are ready for precision medicine given these new data sources and research efforts.
CINECA webinar slides: Open science through fair health data networks dream o...CINECAProject
Since the FAIR data principles were published in 2016, many organizations including science funders and governments have adopted these principles to promote and foster true open science collaborations. However, to define a vision and create a video of a Personal Health Train that leverages worldwide FAIR health data in a federated manner is one step. To actually make this happen at scale and be able to show new scientific and medical insights for it is quite another!
In this webinar, we will dive into the basics of FAIR health data, but also take stock of the current situation in health data networks: after a year of frantic research and collaborations and many open datasets and hackathons on COVID-19, has the situation actually improved? Are we sharing health data on a global scale to improve medical practice, or is quality medical data still only accessible to researchers with the right credentials and deep pockets?
This webinar is part of the “How FAIR are you” webinar series and hackathon, which aim at increasing and facilitating the uptake of FAIR approaches into software, training materials and cohort data, to facilitate responsible and ethical data and resource sharing and implementation of federated applications for data analysis.
The CINECA webinar series aims to discuss ways to address common challenges and share best practices in the field of cohort data analysis, as well as distribute CINECA project results. All CINECA webinars include an audience Q&A session during which attendees can ask questions and make suggestions. Please note that all webinars are recorded and available for posterior viewing. CINECA webinars include an audience Q&A session during which attendees can ask questions and make suggestions.
This webinar took place on 21st January 2021 and is part of the CINECA webinar series.
For previous and upcoming CINECA webinars see:
https://www.cineca-project.eu/webinars
The Human Variome Database in Australia in 2014 - Graham TaylorHuman Variome Project
There are a number of genetics and genomics initiatives underway in Australia, including the Australian node of the Human Variome Project (HVPA), as well as many active research collaborations including familial cancer, endocrine disease, and developmental delay. Most of these projects work with disease-specific databases on a research basis, with the risk that such archives may be ephemeral. HVPA is the only database that is directly integrated with accredited clinical reporting of variants. As such it is designed to capture variants that have passed scrutiny as diagnostically robust, and have therefore already been curated by qualified staff. Registered users access the HVPA database via a secure Internet portal.
I will describe three recent developments of the HVPA database and portal: the upgraded search interface, linkage to other datasets via BioGrid using hash-based de-identified case matching, and the introduction of a genome wide database using LOVD3. Finally I will discuss the future direction of the HVPA and the questions of utility, quality control and sustainability of genetic variation databases.
Search interface
The search interface has to provide useful tools for clinicians and lab scientists so that the HPVA project offers them direct benefits and incentivises them to participate. Following a request for feedback from users, a series of improvements were implemented, initially on a demonstration server and then on the live server following review by the Steering Committee. The highest priorities were for more information about numbers of times particular variants were
recorded, the ability to search by range and to filter by pathogenicity. There was also interest in enabling direct uploading of VCF files and the automated calculation of pathogenicity scores. Many of these features are now implemented and examples will be presented.
Linkage to other datasets
We have implemented the hash key algorithm and work is in progress with BioGrid to link variation data to clinical data sets.
Genome wide database
We have established an HVPA LOVD3 database and are working with the Human Genetics Society of Australasia on a pilot study to sequence the exomes of two trios and review the data using this database.
Building a Network of Interoperable and Independently Produced Linked and Ope...Michel Dumontier
Over 15 years ago, Sir Tim Berners Lee proclaimed the founding of an exciting new future involving intelligent agents operating over smarter data in order to perform complex tasks at the behest of their human controllers. At the heart of this vision lies an uneasy alliance between tedious formal knowledge representations and powerful analytics over big, but often messy data. Bio2RDF, our decade old open source project to create Linked Data for the life sciences, has weaved emergent Semantic Web technologies such as ontologies and Linked Data to generate FAIR - Findable, Accessible, Interoperable, and Reusable - data in the form of billions of machine accessible statements for use in downstream biomedical discovery.
This revolution in data publication has been strengthened by action from global bioinformatics institutions such as the NCBI, NCBO, EBI, and DBCLS. Notably, NCBI's PubChem has successfully coupled large scale data integration with community-based standards to offer a remakable biochemical knowledge resource amenable to data hungry discovery tools. Yet, in the face of increasing pressure from researchers, funders, and publishers, will these approaches be sufficient for growing and maintaining a comprehensive knowledge graph that is inclusive of all biomedical research?
IRIDA: Canada’s federated platform for genomic epidemiology William Hsiao
This document summarizes the IRIDA platform, a federated genomic epidemiology platform for Canada. IRIDA aims to (1) build a user-friendly analysis platform to process genomic data, (2) enable more efficient information sharing between public health agencies, and (3) standardize inconsistent information representation through the use of ontologies. The platform is a partnership between various public health and academic institutions to bridge gaps between genomic research and applications in public health outbreak investigations.
IRIDA: Canada’s federated platform for genomic epidemiology, ABPHM 2015 WHsiaoIRIDA_community
This document summarizes the IRIDA platform, a federated genomic epidemiology platform for Canada. IRIDA aims to bridge gaps between advances in genomic epidemiology and real-time application in public health. It is developing solutions such as building a user-friendly analysis platform, implementing security and role-based sharing of genomic data, and using ontologies to standardize inconsistent information representation and address the complexity of genomic data interpretation. The IRIDA platform is in beta testing and plans continued development and training workshops.
Tutorial on the DisGeNET Discovery Platform, with especial focus on its exploitation in the Semantic Web showing how to retrieve and integrate DisGeNET data with other RDF linked datasets.
HANDI Summit 18 - Introducing HANDI-HOPD - Ewan DavisHANDI HEALTH
NHS England hosted the HANDI-HOPD Summit in London on the 18th September. This was attended by an invited audience of around 40 people to discuss plans to take the HANDI-HOPD platform forward to the NHS England Open Source Open Day on the 26th of November in Newcastle-Upon-Tyne where is will be launched as the Platform for NHS Code4Health.
HANDI-HOPD The HANDI Open Platform Demonstrator provides an experimental platform to demonstrate the power of emerging open standards and APIs to deliver the transformational power of the Internet to support digital health and care.
Ewan Davis introduced the HOPD, described where it fitted in the global development of open health platforms what had already been deployed and our plans for it’s development.
The document discusses the challenges of implementing electronic health records (EHR) in Slovenia and the benefits of using an openEHR approach. It describes how Slovenia created a Smart Healthcare & Wellbeing Cluster to deliver value through an open data platform based on openEHR standards. This has resulted in a vendor-neutral clinical data repository being used at a children's hospital in Slovenia and as part of the national health interoperability backbone. The openEHR approach is now also being used for EHR systems in Moscow, Russia.
Public Laboratory LOINC Workshop and Committee Meeting documents the origins and growth of LOINC as a universal standard for clinical observations and laboratory results. It discusses how LOINC provides a common language for information exchange and how its open model has led to widespread international adoption and translations. Large healthcare organizations around the world have implemented LOINC to facilitate interoperability across hundreds of systems.
Stratergies for the intergration of information (IPI_ConfEX)Ben Gardner
The document discusses approaches to integrating internal and external data across pharmaceutical research. It describes utilizing a data warehousing strategy through a Research Information Factory (RIF) to create a single global repository for research data. However, integrating external data from various sources poses additional challenges. Tools like PharmaMatrix provide a pre-indexed mine of scientific literature linking drug targets to indications, but result sets can be large. The document suggests that Web 2.0 technologies like wikis, blogs and tagging could help turn integrated information into knowledge by enabling collaboration and sharing. Industry-wide data standards and common ontologies would also help facilitate external data integration.
The document provides an overview of linked data, including what it is, how to make linked data, and what can be done with linked data. It discusses several existing linked datasets for health care and life sciences like EntrezGene, UniProt, KEGG Pathway, and DrugBank. It describes how these datasets are published as linked data using HTTP URIs and RDF, and how they are interlinked. It also presents an example SPARQL query run across multiple linked datasets and the results. The document aims to promote the benefits of linked open drug data for applications like drug discovery.
Toward F.A.I.R. Pharma. PhUSE Linked Data Initiatives Past and PresentTim Williams
Abstract:
In recent years, the PhUSE organization has supported several Linked Data initiatives. The CDISC Foundational Standards as RDF is an early example of one such initiative. The results are available on the CDISC website. Subsequent proof of concept projects enjoyed marginal success at a time when pharma’s familiarity with the technology was still very limited. A recent surge in interest in F.A.I.R. data and Knowledge Graphs has sparked renewed interest in Linked Data within PhUSE and the industry at large. The recently completed “Clinical Trials Data as RDF (CTDasRDF)” spawned a new project, “Going Translational With Linked Data (GoTWLD).” GoTWLD extends the project scope of its predecessor beyond SDTM into the non-clinical domain.
Educational initiatives at PhUSE include an introductory, interactive workshop at the annual European conference (EU-Connect) and at the US Computational Science Symposium (CSS). A side-project of GoTWLD is investigating the potential use of URIs as study identifiers to promote adoption of Linked Data. Challenges remain, including the need for demonstrable return on investment and the development of user-friendly, intuitive interfaces for graph data. These challenges can be overcome if pharmaceutical companies cooperate in the pre-competitive space.
Presented at Semantics@Roche, Basel 2019-04-04
Medical innovation calls for new models for collaborations that facilitates, government, academia and industry.
Barriers to research and ultimate commercialization will be lowered by bringing best practices from industry and academic settings.
Hippocrates platform facilitates early drug development extending from basic research to drug invention and commercialization significantly saving time and money.
The platform is designed in such way to facilitate collaboration amongst stakeholders as well as taking advantage of the vast resources currently available on the web to generate and aggregate content based on the needs of the research of the end-user.
The document discusses healthcare informatics and big data in healthcare. It provides an introduction to healthcare informatics, the advantages and disciplines involved. It then discusses big data in healthcare, including the sources and types of healthcare data, challenges in big data analytics, and conceptual architectures. Tools for big data analytics are also outlined, including Hadoop, Pig, Hive and others. Finally, it provides an example case study of a systematic review on the effectiveness of mobile health technology interventions.
Pistoia Alliance US Conference 2015 - 1.3.2 New member introductions - DNAnexusPistoia Alliance
The document discusses The Global Network For Genomic Medicine, which provides a cloud-based platform for large-scale genomic analysis and collaboration. The platform allows various groups like academic centers, developers, and regulators to securely share genomic data, reference data, and analysis pipelines. It has supported massive projects like analyzing over 14,000 individual genomes from the Baylor College of Medicine and hosting thousands of datasets from the ENCODE project.
This document discusses openEHR, an open specification for health information modeling that supports an open digital care ecosystem. OpenEHR allows clinical data to remain fully interoperable and queryable across systems and technologies through archetypes and templates defined by clinicians. It provides a standards-based approach using normal technical specifications to define how clinical content and health data are represented separately from programming languages or databases. This enables apps and systems to integrate detailed clinical models directly without proprietary constraints.
Open science and medical evidence generation - Kees van Bochove - The HyveKees van Bochove
Presentation about open science, the FAIR principles, and medical evidence generation with the OHDSI COVID-19 study-a-thon as an example. I've used variations on this deck in a couple of classroom and online courses for PhD and master students early 2020.
SCOPE Summit - Applying the OMOP data model & OHDSI software to national Euro...Kees van Bochove
Talk from Kees van Bochove, The Hyve at SCOPE Summit, Real World Data track, Jan 26, 2017, Miami
A large open source initiative for standardisation and epidemiological analysis for real world data is OHDSI: Observational Health Data Sciences and Informatics. OHDSI leverages the OMOP common data model for observational data, and provides data analysis tools for a broad range of use cases. This talk will explain OMOP and OHDSI with case study IMI EMIF, in which health data from over 50 million patients from 13 national and regional European registries is brought together.
Similar to DisGeNET: a discovery platform to support translational research and drug discovery (20)
- Video recording of this lecture in English language: https://youtu.be/Pt1nA32sdHQ
- Video recording of this lecture in Arabic language: https://youtu.be/uFdc9F0rlP0
- Link to download the book free: https://nephrotube.blogspot.com/p/nephrotube-nephrology-books.html
- Link to NephroTube website: www.NephroTube.com
- Link to NephroTube social media accounts: https://nephrotube.blogspot.com/p/join-nephrotube-on-social-media.html
Travel vaccination in Manchester offers comprehensive immunization services for individuals planning international trips. Expert healthcare providers administer vaccines tailored to your destination, ensuring you stay protected against various diseases. Conveniently located clinics and flexible appointment options make it easy to get the necessary shots before your journey. Stay healthy and travel with confidence by getting vaccinated in Manchester. Visit us: www.nxhealthcare.co.uk
Nano-gold for Cancer Therapy chemistry investigatory projectSIVAVINAYAKPK
chemistry investigatory project
The development of nanogold-based cancer therapy could revolutionize oncology by providing a more targeted, less invasive treatment option. This project contributes to the growing body of research aimed at harnessing nanotechnology for medical applications, paving the way for future clinical trials and potential commercial applications.
Cancer remains one of the leading causes of death worldwide, prompting the need for innovative treatment methods. Nanotechnology offers promising new approaches, including the use of gold nanoparticles (nanogold) for targeted cancer therapy. Nanogold particles possess unique physical and chemical properties that make them suitable for drug delivery, imaging, and photothermal therapy.
PGx Analysis in VarSeq: A User’s PerspectiveGolden Helix
Since our release of the PGx capabilities in VarSeq, we’ve had a few months to gather some insights from various use cases. Some users approach PGx workflows by means of array genotyping or what seems to be a growing trend of adding the star allele calling to the existing NGS pipeline for whole genome data. Luckily, both approaches are supported with the VarSeq software platform. The genotyping method being used will also dictate what the scope of the tertiary analysis will be. For example, are your PGx reports a standalone pipeline or would your lab’s goal be to handle a dual-purpose workflow and report on PGx + Diagnostic findings.
The purpose of this webcast is to:
Discuss and demonstrate the approaches with array and NGS genotyping methods for star allele calling to prep for downstream analysis.
Following genotyping, explore alternative tertiary workflow concepts in VarSeq to handle PGx reporting.
Moreover, we will include insights users will need to consider when validating their PGx workflow for all possible star alleles and options you have for automating your PGx analysis for large number of samples. Please join us for a session dedicated to the application of star allele genotyping and subsequent PGx workflows in our VarSeq software.
STUDIES IN SUPPORT OF SPECIAL POPULATIONS: GERIATRICS E7shruti jagirdar
Unit 4: MRA 103T Regulatory affairs
This guideline is directed principally toward new Molecular Entities that are
likely to have significant use in the elderly, either because the disease intended
to be treated is characteristically a disease of aging ( e.g., Alzheimer's disease) or
because the population to be treated is known to include substantial numbers of
geriatric patients (e.g., hypertension).
Giloy in Ayurveda - Classical Categorization and SynonymsPlanet Ayurveda
Giloy, also known as Guduchi or Amrita in classical Ayurvedic texts, is a revered herb renowned for its myriad health benefits. It is categorized as a Rasayana, meaning it has rejuvenating properties that enhance vitality and longevity. Giloy is celebrated for its ability to boost the immune system, detoxify the body, and promote overall wellness. Its anti-inflammatory, antipyretic, and antioxidant properties make it a staple in managing conditions like fever, diabetes, and stress. The versatility and efficacy of Giloy in supporting health naturally highlight its importance in Ayurveda. At Planet Ayurveda, we provide a comprehensive range of health services and 100% herbal supplements that harness the power of natural ingredients like Giloy. Our products are globally available and affordable, ensuring that everyone can benefit from the ancient wisdom of Ayurveda. If you or your loved ones are dealing with health issues, contact Planet Ayurveda at 01725214040 to book an online video consultation with our professional doctors. Let us help you achieve optimal health and wellness naturally.
Know the difference between Endodontics and Orthodontics.Gokuldas Hospital
Your smile is beautiful.
Let’s be honest. Maintaining that beautiful smile is not an easy task. It is more than brushing and flossing. Sometimes, you might encounter dental issues that need special dental care. These issues can range anywhere from misalignment of the jaw to pain in the root of teeth.
Know the difference between Endodontics and Orthodontics.
DisGeNET: a discovery platform to support translational research and drug discovery
1. DisGeNET-RDF: a GDA Linked Open Data resource
DisGeNET: a discovery platform to support translational research and drug discovery
Janet Piñero, Núria Queralt-Rosinach, Àlex Bravo, Ferran Sanz and Laura I. Furlong
Integrative Biomedical Informatics Group, Research Programme on Biomedical Informatics; Hospital del Mar Medical Research Institute; Pompeu Fabra University
Acknowledgements
The authors thank the Open PHACTSpartners, MichelDumontierand the OpenLinkstaff for their input, collaborationand help.
Funding: We received support from ISCIII-FEDER (PI13/00082, CP10/00524), from the IMI-JU under grants agreements nº 115002 (eTOX), nº
115191 (Open PHACTS)], nº 115372 (EMIF) and nº 115735 (iPiE), resources of which are composed of financial con-tribution from the European
Union's Seventh Framework Pro-gramme (FP7/2007-2013) and EFPIA companies’ in kind contribu-tion, and the EU H2020 Programme 2014-2020
under grant agreements no. 634143 (MedBioinformatics) and no. 676559 (Elixir-Excelerate). The Research Programme on Biomedical Informatics
(GRIB)is a node of the Spanish National Institute of Bioinformatics(INB).
DisGeNET: Disease-Gene NETwork of relations for discovery
DATA
DISCOVERY
KNOWLEDGE BASE
TOOLS FOR EXPLORATION AND ANALYSIS
Motivation: Better understanding of human gene component and disease mechanisms for translational research and
drug discovery and development.
Challenge: One of the major current bottlenecks for knowledge discovery on the genetic component of diseases is
that the information is fragmented. The vast amount of biomedical information about genotype-phenotype relations
is distributed in several databases, represented and annotated using different data models, vocabularies and
standards, and it is domain and technology-specific, which hampers their access, integration, analysis, and
interpretation.
Approach: DisGeNET Discovery Platform1 collects and integrates the available information on gene-disease
associations (GDAs), covering the whole spectrum of human diseases, and using standards for their annotation and
representation.
DisGeNET in the LOD cloud for translational research
• DisGeNET + external multidomain
sources in LOD.
• It is interlinked to other biomedical
databases to answer scientific
questions that need the interrogation
of cross-domain resources.
• It aims to support the development
of bioinformatic Semantic Web
applications to extract key knowledge
on the molecular mechanisms of
diseases.
Implementation: The platform is composed of a knowledge base and a set of tools for data analysis and interpretation.
EVIDENCE-BASED DISCOVERY
CLINICIAN
INTEROPERABILITY
METADATA
DATABASES&
LITERATURE STANDARDS
INTEGRATION
OPEN
http://www.disgenet.org/
RESEARCHER
CURATOR
BIOINFORMATICIAN &
DEVELOPER
DISCOVERABILITY COMMUNITY USE
LARGE-SCALE EXTRACTION AND INTEGRATION
DIGITAL PUBLICATION,
SHARING AND LINKING
Usage stats (Ago2014-Ago2015):
• 12,040 users, 22,696 sessions
• 14,494 downloads
• DisGeNET used in +20 publications,
cited in +60 articles
• Other Projects: PubAnnotation,
OpenLifeData
Registered:
• biosharing
• OMICtools
• NeuroLex
• Datahub
Present in the Semantic Web:
• URI/RDF/nanpublications
• Machine-processable
• Semantic integration
• Links to the Linked Open Data (LOD)
cloud
• Data analysis across domains
SEMANTIC WEB
What is the tissue expression pattern of the genes associated to Obesity?
• Large-scaleintegration across domains
• 17,181 Genes
• PANTHER class
• 14,610 Diseases
• MeSH class
60% complex,36%rare/Mendelian,
and 4% infectiousdiseases
DO MSH OMIM NCI ORDO ICD9
19 58 38 33 13 12
TRACK OF EVIDENCE
S = WCURATED + WPREDICTED + WLITERATURE
• Provenance(PubMed ID, source)
• DisGeNETscore (evidence)
Web:
http://www.disgenet.org/
RDF:
http://rdf.disgenet.org/
SPARQL:
http://rdf.disgenet.org/sparql/
Open PHACTS API:
https://dev.openphacts.org
ACCESS
Open Database License:
http://opendatacommons.org/licenses/odbl/1.0/
Downloads:
• Tab separated plain text
• SQLite
• RDF
• Trusty nanopublications
Webinterface
SPARQL endpoint / Linked Data browser
Open PHACTS Discovery Platform
Nanopublication network
disGeNET2R R package
DIFFERENT USER PROFILES AVAILABILITY
Metadata:
• data-item description
• dataset description
Programmatic access:
• Automatic analysis
• Higher speed
• Reduce error
• Share results
• Embed in workflows
REPRODUCIBILITY
Several formats and
models
Transparency and
Validation
SOURCES
Recentfindings
429,111 Gene-Disease Associations
Sentence description
NORMALIZATION
HARMONIZATION
• NCBI Gene ID
• UMLS CUIs.
DisGeNET association type ontology
INTEROPERABILITY
SYNTACTIC
COMMON IDs and ONTOLOGIES
SEMANTIC
• 11 common ontologies in
• RDF2
• Nanopublications3
• GENE:
• DISEASE
STANDARDIZATION
Digital objects
DisGeNET association type ontology
Semanticscience Integrated Ontology
(SIO)4
• Normalized Identification Scheme
http://rdf.disgenet.org/resource/gda/ + ID
http://lod-cloud.net/;Aug2014
4,962,315 RDF links to RDF datasets in the LOD
https://datahub.io/dataset/disgenet
(morestatistics)
LOD cloud RDFIZATION
METADATA
RDF
INTERLINKING
• Dataset (Open PHACTS + )
• Linksets (Open PHACTS + )
• Use Open PHACTS guidelines
• Dereferenceable URIs (primary or
)
• SIO
•
• )
OWL
• NCBI Gene ID
• PANTHER Classification
• UMLS CUIs
• MeSH Classification
• Data providers
• Disease annotation in the Open PHACTS Discovery Platform5
• OMIM included
• > 20 000 000 of triples
RDF SCHEMA METADATA INTERLINKING
• Linksets providers
• > 70 000 number of linksets
FUTURE
New data:
• Disease-phenotype associations (HPO)
• New use cases
• New API calls
Score:
• Add to API calls
EXPLORER KNIME
More @ http://www.disgenet.org/web/DisGeNET/menu/rdf#sparql-queries-2
MAPPINGS TO OTHER DISEASE TERMINOLOGIES
DRUG
TARGETPATHWAY
DISEASE
DISEASE PHENOTYPE
DISEASE GENE
GDA
EVIDENCE SNPSCORE
Gene-disease association as entity
• Data item
• Dataset
<disease>
<void:inDataset><dgn-void:disease-dataset>
http://www.myexperiment.org/groups/1125.html
API
References
1. Piñero, J., Queralt-Rosinach, N., Bravo, A., Deu-Pons, J., Bauer-Mehren, A., Baron, M., … Furlong, L. I. (2015). DisGeNET: a
discovery platform for the dynamical exploration of human diseases and their genes. Database, 2015(0),bav028–bav028.
2. Queralt-Rosinach, N., Piñero,J. , Bravo, À, Sanz, F. and Furlong, L.I. DisGeNET-RDF: harnessing the innovative power of the
Semantic Web to explore the genetic basis of diseases, 2015 (submitted).
3. Queralt-Rosinach, N., Kuhn, T., Chichester, C., Dumontier, M., Sanz, F., and Furlong, L.I., Publishing DisGeNET as
Nanopublications. Semantic Web Journal, (to appear), 1-10, 2015.
4. Dumontier, M., Baker, C. J., Baran, J., Callahan, A., Chepelev, L., Cruz-Toledo, J., … Hoehndorf, R. (2014). The Semanticscience
Integrated Ontology (SIO) for biomedical research and knowledge discovery. Journal of Biomedical Semantics, 5(1), 2014.
5. Gray, A. J. G., Groth, P., Loizou, A., Askjaer, S., Brenninkmeijer, C., Burger, K., … Williams, A. J. (2014, January 1). Applying linked
data approaches to pharmacology: Architectural decisions and implementation. Semantic Web. IOS Press. doi:10.3233/SW-
2012-0088
• GDAs described by SIO
https://dev.openphacts.org
/disease/getTargets
http://rdf.disgenet.org/void-v3.0.0.ttl
Which compounds target proteins associated with Parkinson's disease or Alzheimer's disease?
DisGeNET in the Open PHACTS Discovery Platform for drug discovery and development