Allotrope foundation vanderwall_and_little_bio_it_world_2016OSTHUS
The document discusses the Allotrope Foundation's efforts to drive improved data modeling and management in life sciences research through the use of semantic technologies. It outlines current challenges with data silos and lack of standards. The Foundation is developing the Allotrope Data Format and taxonomies to standardize metadata and facilitate data integration and sharing. Several pharmaceutical companies are now implementing the framework in areas like small molecule CMC and biotherapeutics development. The Foundation aims to enable smarter laboratories of the future with integrated, sharable, and analyzable data.
Allotrope Foundation & OSTHUS at SmartLab Exchange 2015: Update on the Allotr...OSTHUS
During SmartLab Exchange 2015, Allotrope Foundation and OSTHUS presented the latest update on the Allotrope Framework. To learn more, please view the slides below.
Presented by:
Dana Vanderwall, BMS Research IT & Automation Patrick Chin, Merck Research Laboratories IT Wolfgang Colsman, OSTHUS
OSTHUS-Allotrope presents "Laboratory Informatics Strategy" at SmartLab 2015OSTHUS
Building your laboratory informatics strategy: The benefit of reference architectures & data standardization.
Presented by:
Wolfgang Colsman, OSTHUS
Dana Vanderwall, Bristol-Myers Squibb
Semantics for integrated laboratory analytical processes - The Allotrope Pers...OSTHUS
The software environment currently found in the analytical community consists of a patchwork of incompatible software, proprietary and non-standardized file formats,
which is further complicated by incomplete, inconsistent and potentially inaccurate metadata. To overcome these issues, the Allotrope Foundation develops a
comprehensive and innovative Framework consisting of metadata dictionaries, data standards, and class libraries for managing analytical data throughout its lifecycle. The
talk describes how laboratory data and their semantic metadata descriptions are brought together to ease the management of vast amount of data that underpin almost
every aspect of drug discovery and development.
Semantics for Integrated Analytical Laboratory Processes – the Allotrope Pers...OSTHUS
The software environment currently found in the analytical community consists of a patchwork of incompatible software, proprietary and non-standardized file formats, which is further complicated by incomplete, inconsistent and potentially inaccurate metadata. To overcome these issues, Allotrope Foundation is developing a comprehensive and innovative framework consisting of metadata dictionaries, data standards, and class libraries for managing analytical data throughout its life cycle. In this talk we describe how laboratory data and semantic metadata descriptions are brought together to ease the management of a vast amount of data that underpins almost every aspect of drug discovery and development.
IQPC’s 5th Forum on Laboratory Informatics will provide strategies for overcoming challenges, including:
- In-depth regulatory compliance guidance
- Extensive ELN deployment and roll out projects, focusing on ROI maximization and impact on business performance
- Informatics systems in the biobanking environment
- Proactive approaches to address challenges of integration and interfacing
- Integrating and embracing knowledge management and social media tools
Revolutionizing Laboratory Instrument Data for the Pharmaceutical Industry:...OSTHUS
The Allotrope Foundation is a consortium of major pharmaceutical companies and a partner network whose goal is to address challenges in the pharmaceutical industry by providing a set of public, non-proprietary standards for using and integrating analytical laboratory data. Current challenges in data management within the pharmaceutical industry often center around inconsistent or incomplete data and metadata and proprietary data formats. Because of a lack of standardization, several operations (e.g. integration of instruments/applications, transfer of methods or results, archiving for regulatory purposes) require unnecessary efforts. Further, higher level aggregation of data, e.g. regulatory filings, that are derived from multiple sources of laboratory data are costly to create. These unnecessary costs impact operations within a company’s laboratories, between partnering companies, and between a company and contract research organizations (CROs). Finally, the accelerating transition of laboratories from hybrid (paper + electronic) to purely electronic data streams, coupled with an ever-increasing regulatory scrutiny of electronic data management practices, further require a comprehensive solution. This talk will discuss how The Allotrope Foundation is providing a new framework for data standards through collaboration between numerous stakeholders.
Reinventing Laboratory Data To Be Bigger, Smarter & FasterOSTHUS
• Big Data technologies, especially Data Lakes are spreading across many industries at the moment with the hopes that they will provide unprecedented capabilities for data integration and data analytics
• In spite of the popularity and promise of these technology approaches, many early adopters are not seeking immediate solutions to their complex problems. Answers are not simply appearing – this talk will explore this issue more thoroughly
• Of the 4 V’s of Big Data, Data Variety and Data Veracity (uncertainty) are of increasing importance. These can cause barriers to successful integration strategies , which, in turn, can lead to poorly performing analytics.
• The problems of Variety and Veracity can be tackled using a new form of Data Science which combines formal ontologies with statistical heuristics. This talk will explore some key features of these approaches and how they can be developed together in symbiosis – leading to complex models that allow for improved analytics – or as we call it Big Analysis.
• The end result is improved capture of data types/sources, from laboratory instrument data, to clinical data, to regulatory rules & submissions, all the way to business drivers for the enterprise. In the end providing advanced analytics capabilities that can be built as modules and expand across an enterprise.
Allotrope foundation vanderwall_and_little_bio_it_world_2016OSTHUS
The document discusses the Allotrope Foundation's efforts to drive improved data modeling and management in life sciences research through the use of semantic technologies. It outlines current challenges with data silos and lack of standards. The Foundation is developing the Allotrope Data Format and taxonomies to standardize metadata and facilitate data integration and sharing. Several pharmaceutical companies are now implementing the framework in areas like small molecule CMC and biotherapeutics development. The Foundation aims to enable smarter laboratories of the future with integrated, sharable, and analyzable data.
Allotrope Foundation & OSTHUS at SmartLab Exchange 2015: Update on the Allotr...OSTHUS
During SmartLab Exchange 2015, Allotrope Foundation and OSTHUS presented the latest update on the Allotrope Framework. To learn more, please view the slides below.
Presented by:
Dana Vanderwall, BMS Research IT & Automation Patrick Chin, Merck Research Laboratories IT Wolfgang Colsman, OSTHUS
OSTHUS-Allotrope presents "Laboratory Informatics Strategy" at SmartLab 2015OSTHUS
Building your laboratory informatics strategy: The benefit of reference architectures & data standardization.
Presented by:
Wolfgang Colsman, OSTHUS
Dana Vanderwall, Bristol-Myers Squibb
Semantics for integrated laboratory analytical processes - The Allotrope Pers...OSTHUS
The software environment currently found in the analytical community consists of a patchwork of incompatible software, proprietary and non-standardized file formats,
which is further complicated by incomplete, inconsistent and potentially inaccurate metadata. To overcome these issues, the Allotrope Foundation develops a
comprehensive and innovative Framework consisting of metadata dictionaries, data standards, and class libraries for managing analytical data throughout its lifecycle. The
talk describes how laboratory data and their semantic metadata descriptions are brought together to ease the management of vast amount of data that underpin almost
every aspect of drug discovery and development.
Semantics for Integrated Analytical Laboratory Processes – the Allotrope Pers...OSTHUS
The software environment currently found in the analytical community consists of a patchwork of incompatible software, proprietary and non-standardized file formats, which is further complicated by incomplete, inconsistent and potentially inaccurate metadata. To overcome these issues, Allotrope Foundation is developing a comprehensive and innovative framework consisting of metadata dictionaries, data standards, and class libraries for managing analytical data throughout its life cycle. In this talk we describe how laboratory data and semantic metadata descriptions are brought together to ease the management of a vast amount of data that underpins almost every aspect of drug discovery and development.
IQPC’s 5th Forum on Laboratory Informatics will provide strategies for overcoming challenges, including:
- In-depth regulatory compliance guidance
- Extensive ELN deployment and roll out projects, focusing on ROI maximization and impact on business performance
- Informatics systems in the biobanking environment
- Proactive approaches to address challenges of integration and interfacing
- Integrating and embracing knowledge management and social media tools
Revolutionizing Laboratory Instrument Data for the Pharmaceutical Industry:...OSTHUS
The Allotrope Foundation is a consortium of major pharmaceutical companies and a partner network whose goal is to address challenges in the pharmaceutical industry by providing a set of public, non-proprietary standards for using and integrating analytical laboratory data. Current challenges in data management within the pharmaceutical industry often center around inconsistent or incomplete data and metadata and proprietary data formats. Because of a lack of standardization, several operations (e.g. integration of instruments/applications, transfer of methods or results, archiving for regulatory purposes) require unnecessary efforts. Further, higher level aggregation of data, e.g. regulatory filings, that are derived from multiple sources of laboratory data are costly to create. These unnecessary costs impact operations within a company’s laboratories, between partnering companies, and between a company and contract research organizations (CROs). Finally, the accelerating transition of laboratories from hybrid (paper + electronic) to purely electronic data streams, coupled with an ever-increasing regulatory scrutiny of electronic data management practices, further require a comprehensive solution. This talk will discuss how The Allotrope Foundation is providing a new framework for data standards through collaboration between numerous stakeholders.
Reinventing Laboratory Data To Be Bigger, Smarter & FasterOSTHUS
• Big Data technologies, especially Data Lakes are spreading across many industries at the moment with the hopes that they will provide unprecedented capabilities for data integration and data analytics
• In spite of the popularity and promise of these technology approaches, many early adopters are not seeking immediate solutions to their complex problems. Answers are not simply appearing – this talk will explore this issue more thoroughly
• Of the 4 V’s of Big Data, Data Variety and Data Veracity (uncertainty) are of increasing importance. These can cause barriers to successful integration strategies , which, in turn, can lead to poorly performing analytics.
• The problems of Variety and Veracity can be tackled using a new form of Data Science which combines formal ontologies with statistical heuristics. This talk will explore some key features of these approaches and how they can be developed together in symbiosis – leading to complex models that allow for improved analytics – or as we call it Big Analysis.
• The end result is improved capture of data types/sources, from laboratory instrument data, to clinical data, to regulatory rules & submissions, all the way to business drivers for the enterprise. In the end providing advanced analytics capabilities that can be built as modules and expand across an enterprise.
Managing Complex Data Packages with Labcore SDMS Scientific Data Management...msf4566
The document discusses Labcore SDMS, a scientific data management system that helps laboratories manage complex data packages. It reduces costs and risks for laboratories by automating manual paper processes, expediting search and retrieval of data packages, and providing full audit trails and version history. Key benefits include significant reduction in labor costs to produce and store data packages, faster delivery of data packages to customers, and less errors and liability risks. It seamlessly integrates with LIMS, laboratory instruments, and other data systems.
Introduction of BJU-BMR-RG and use case study of Applying openEHR archetypes ...openEHR-Japan
The document discusses applying openEHR archetypes to implement a clinical data repository (CDR) in China. It analyzes existing EMR data schemas, identifies 892 relevant data items, and maps them to 62 clinical concepts guided by openEHR. Most concepts were mapped directly to existing archetypes, while some required extension or specialization to fully represent Chinese CDR requirements. Implementing a CDR based on openEHR archetypes allows clinical experts to define, retrieve, and query necessary data flexibly.
How to Rapidly Configure Oracle Life Sciences Data Hub (LSH) to Support the M...Perficient
This document outlines best practices for rapidly configuring Oracle Life Sciences Data Hub (LSH) to support patient data management. It discusses data flows, conforming data to standards, necessary utilities and tools, infrastructure requirements, and implementation process. The presentation recommends hosting the LSH environment with BioPharm for a turnkey solution, reducing time and risks compared to a custom implementation.
Enabling Clinical Data Reuse with openEHR Data Warehouse EnvironmentsLuis Marco Ruiz
Modern medicine needs methods to enable access to data,
captured during health care, for research, surveillance,
decision support and other reuse purposes. Initiatives like the
National Patient Centered Clinical Research Network in the
US and the Electronic Health Records for Clinical Research
in the EU are facilitating the reuse of Electronic Health
Record (EHR) data for clinical research. One of the barriers
for data reuse is the integration and interoperability of
different Healthcare Information Systems (HIS). The reason is
the differences among the HIS information and terminology
models. The use of EHR standards like openEHR can alleviate
these barriers providing a standard, unambiguous,
semantically enriched representation of clinical data to
enable semantic interoperability and data integration. Few
works have been published describing how to drive
proprietary data stored in EHRs into standard openEHR
repositories. This tutorial provides an overview of the key
concepts, tools and techniques necessary to implement an
openEHR-based Data Warehouse (DW) environment to reuse
clinical data. We aim to provide insights into data extraction
from proprietary sources, transformation into openEHR
compliant instances to populate a standard repository and
enable access to it using standard query languages and
services
Automated and Explainable Deep Learning for Clinical Language Understanding a...Databricks
Unstructured free-text medical notes are the only source for many critical facts in healthcare. As a result, accurate natural language processing is a critical component of many healthcare AI applications like clinical decision support, clinical pathway recommendation, cohort selection, patient risk or abnormality detection.
The presentation talks about the components of FHIR, its distribution and use. Scenarios for the introduction of FHIR in the country using HL7 V3 are also offered.
ICIC 2014 Increasing the efficiency of pharmaceutical research through data i...Dr. Haxel Consult
The pressures of pharmaceutical research and development demand increasing efficiency from scientists. High-quality decisions must be made faster and encompass all available information. At the same time there is a growing desire to better utilize the multi-billion dollar research investment recorded in laboratory notebooks and bioassay databases. Key values for data integration in a data exploration environment include gathering data from disparate E-notebooks and bioassay databases into a single searchable “virtual” system and increased discoverability by accessing data through a system designed for exploration. Key benefits are better chemistry decisions through easier access to broader data and reduced time for preparing patent filings. The ability to interlink in-house and reported assay data with in-house and published chemistry provides a data-rich environment for developing insights and predictive models. We will discuss our experience with integrating information from journals, patents, bio-assay databases, and E-lab notebooks to address these needs.
Building linked data large-scale chemistry platform - challenges, lessons and...Valery Tkachenko
Chemical databases have been around for decades, but in recent years we have observed a qualitative change from rather small in-house built proprietary databases to large-scale, open and increasingly complex chemistry knowledgebases. This tectonic shift has imposed new requirements for database design and system architecture as well as the implementation of completely new components and workflows which did not exist in chemical databases before. Probably the most profound change is being caused by the linked nature of modern resources - individual databases are becoming nodes and hubs of a huge and truly distributed web of knowledge. This change has important aspects such as data and format standards, interoperability, provenance, security, quality control and metainformation standards.
ChemSpider at the Royal Society of Chemistry was first public chemical database which incorporated rigorous quality control by introducing both community curation and automated quality checks at the scale of tens of millions of records. Yet we have come to realize that this approach may now be incomplete in a quickly changing world of linked data. In this presentation we will talk about challenges associated with building modern public and private chemical databases as well as lessons that we have learned from our past and present experience. We will also talk about solutions for some common problems.
The OntoChem IT Solutions GmbH ...
... was founded in 2015 as a purely IT-oriented offshoot of the OntoChem GmbH. Even before we had many years of experience and it has always been our mission to provide added value to our customers by helping them to navigate today’s complex information world by developing cognitive computing solutions, indexing intranet and internet data and applying semantic search solutions for pharmaceutical, material sciences and technology driven businesses.
We strive to support our customers with the most useful tools for knowledge discovery possible, encompassing up-to-date data sources, optimized ontologies and high-throughput semantic document processing and annotation techniques.
We create new knowledge from structured and unstructured data by extracting relationships thereby exploiting the full potential of full-text documents & databases while also scanning social media, news flows and analyzing web-pages.
We aim at an unprecedented, machine understanding of text and subsequent knowledge extraction and inference. The application of our methods towards chemical compounds and their properties supports our customers in generating intellectual property and their use as novel therapeutics, agrochemical products, nutraceuticals, cosmetics and in the field of novel materials.
It's our mission to provide added value to customers by:
developing and applying cognitive computing solutions
creating intranet and internet data indexing and semantic search solutions
Big Data analytics for technology driven businesses
supporting product development and surveillance.
We deliver useful tools for knowledge discovery for:
creating background knowledge ontologies
high-throughput semantic document processing and annotation
knowledge mining by extracting relationships
exploiting the full potential of full-text documents & databases while also scanning social media, news flows and analyzing web-pages.
This document discusses using Wiley's Chemistry Toolkit to plan syntheses of target molecules. It describes an approach using computer-aided synthesis design (CASD) to perform retrosynthetic analysis back to available starting materials, generating many alternatives and supporting them with literature examples. CASD works by extracting reaction rules from source reactions in databases like CIRX and applying chemical perception to identify compatible functional groups and regio- and stereoselectivity. The goal is to provide an integrated solution for data mining and synthesis planning.
The Open PHACTS project delivers an online platform integrating a wide variety of data from across chemistry and the life sciences and an ecosystem of tools and services to query this data in support of pharmacological research, turning the semantic web from a research project into something that can be used by practising medicinal chemists in both academia and industry. In the summer of 2015 it was the first winner of the European Linked Data Award. At the Royal Society of Chemistry we have provided the chemical underpinnings to this system and in this talk we review its development over the past five years. We cover both our early work on semantic modelling of chemistry data for the Open PHACTS triplestore and more recent work building an all-purpose data platform, for which the Open PHACTS data has been an important test case, what has worked well, what's missing and where this is is likely to go in future.
Alice: "What version of ChEMBL are we using?"
Bob: "Er…let me check. It's going to take a while, I'll get back to you."
This simple question took us the best part of a month to resolve and involved several individuals. Knowing the provenance of your data is essential, especially when using large complex systems that process multiple datasets.
The underlying issues of this simple question motivated us to improve the provenance data in the Open PHACTS project. We developed a guideline for dataset descriptions where the metadata is carried with the data. In this talk I will highlight the challenges we faced and give an overview of our metadata guidelines.
Presentation given to the W3C Semantic Web for Health Care and Life Sciences Interest Group on 14 January 2013.
Over the past decade, CDISC Standards have been widely accepted and implemented in clinical research. The FDA’s final “Guidance for Industry on electronic submission” mandates that submission data conform to CDISC standards such as SDTM, ADaM and SEND. This presentation will discuss how life sciences organizations can use Standards metadata to manage the regulatory compliance process. It will introduce how standards metadata management not only ensures regulatory compliance, but also supports process efficiency in clinical trial artefacts (e.g., protocol, CDASH, SDMT and ADaM) development and standards governance, and enables efficient communication between organizational units.
It will also introduce metadata management system and discuss how metadata management system will create, store, govern and manage standards. It will also show how standards metadata management system interacts with ETL system and dictates standards-driven clinical artefacts development.
Why ICT Fails in Healthcare: Software Maintenance and MaintainabilityKoray Atalag
This presentation was for a SERG seminar at the University of Auckland Department of Computer Science. I present why software maintenance is a barrier for adoption of IT in healthcare and the maintainability aspects based on ISO/IEC 9126 software quality standard quality model. I then present the preliminary results of my research here.
This document provides an overview of how LOINC (Logical Observation Identifiers Names and Codes) codes can be used with FHIR (Fast Healthcare Interoperability Resources). It discusses how LOINC codes are represented and used in various FHIR resources like Observation, Questionnaire, DiagnosticReport, etc. It also describes how FHIR terminology services can be used to retrieve information about LOINC codes and structures like parts, answer lists, and properties to build value sets. The document demonstrates how LOINC enhances interoperability when clinical data is coded with LOINC and accessible via FHIR.
FHIR Developer Days 2015. Study on db implementations for FHIR serverIgor Bossenko
Presentation describes different approaches for implementing database for FHIR server, that were considered during implementation of Nortal National Healthcare System (NHS) for Lithuania.
This document provides an overview of the openEHR CDR open source project called EHRbase. EHRbase aims to provide an open standard-compliant backend platform for electronic health records and clinical applications using the openEHR specification. It has a team of developers across multiple continents and uses modern development practices like Scrum and BDD. EHRbase provides a REST API and SDK for creating, querying, and managing openEHR objects in a clinical data repository, and also integrates with FHIR through a FHIR bridge. It is being used as the backend platform for a national COVID-19 system in Germany.
How to Manage APIs in your Enterprise for Maximum Reusability and GovernanceWSO2
This document discusses API management and governance in an enterprise. It covers the challenges of APIs existing in silos across an organization, and addresses how an API manager and governance registry can help manage the API lifecycle from planning to sharing. It also demonstrates how the WSO2 API management platform can be used to publish, protect, analyze usage of APIs across different environments.
Data Integration for Big Data (OOW 2016, Co-Presented With Oracle)Rittman Analytics
Oracle Data Integration Platform is a cornerstone for big data solutions that provides five core capabilities: business continuity, data movement, data transformation, data governance, and streaming data handling. It includes eight core products that can operate in the cloud or on-premise, and is considered the most innovative in areas like real-time/streaming integration and extract-load-transform capabilities with big data technologies. The platform offers a comprehensive architecture covering key areas like data ingestion, preparation, streaming integration, parallel connectivity, and governance.
Managing Complex Data Packages with Labcore SDMS Scientific Data Management...msf4566
The document discusses Labcore SDMS, a scientific data management system that helps laboratories manage complex data packages. It reduces costs and risks for laboratories by automating manual paper processes, expediting search and retrieval of data packages, and providing full audit trails and version history. Key benefits include significant reduction in labor costs to produce and store data packages, faster delivery of data packages to customers, and less errors and liability risks. It seamlessly integrates with LIMS, laboratory instruments, and other data systems.
Introduction of BJU-BMR-RG and use case study of Applying openEHR archetypes ...openEHR-Japan
The document discusses applying openEHR archetypes to implement a clinical data repository (CDR) in China. It analyzes existing EMR data schemas, identifies 892 relevant data items, and maps them to 62 clinical concepts guided by openEHR. Most concepts were mapped directly to existing archetypes, while some required extension or specialization to fully represent Chinese CDR requirements. Implementing a CDR based on openEHR archetypes allows clinical experts to define, retrieve, and query necessary data flexibly.
How to Rapidly Configure Oracle Life Sciences Data Hub (LSH) to Support the M...Perficient
This document outlines best practices for rapidly configuring Oracle Life Sciences Data Hub (LSH) to support patient data management. It discusses data flows, conforming data to standards, necessary utilities and tools, infrastructure requirements, and implementation process. The presentation recommends hosting the LSH environment with BioPharm for a turnkey solution, reducing time and risks compared to a custom implementation.
Enabling Clinical Data Reuse with openEHR Data Warehouse EnvironmentsLuis Marco Ruiz
Modern medicine needs methods to enable access to data,
captured during health care, for research, surveillance,
decision support and other reuse purposes. Initiatives like the
National Patient Centered Clinical Research Network in the
US and the Electronic Health Records for Clinical Research
in the EU are facilitating the reuse of Electronic Health
Record (EHR) data for clinical research. One of the barriers
for data reuse is the integration and interoperability of
different Healthcare Information Systems (HIS). The reason is
the differences among the HIS information and terminology
models. The use of EHR standards like openEHR can alleviate
these barriers providing a standard, unambiguous,
semantically enriched representation of clinical data to
enable semantic interoperability and data integration. Few
works have been published describing how to drive
proprietary data stored in EHRs into standard openEHR
repositories. This tutorial provides an overview of the key
concepts, tools and techniques necessary to implement an
openEHR-based Data Warehouse (DW) environment to reuse
clinical data. We aim to provide insights into data extraction
from proprietary sources, transformation into openEHR
compliant instances to populate a standard repository and
enable access to it using standard query languages and
services
Automated and Explainable Deep Learning for Clinical Language Understanding a...Databricks
Unstructured free-text medical notes are the only source for many critical facts in healthcare. As a result, accurate natural language processing is a critical component of many healthcare AI applications like clinical decision support, clinical pathway recommendation, cohort selection, patient risk or abnormality detection.
The presentation talks about the components of FHIR, its distribution and use. Scenarios for the introduction of FHIR in the country using HL7 V3 are also offered.
ICIC 2014 Increasing the efficiency of pharmaceutical research through data i...Dr. Haxel Consult
The pressures of pharmaceutical research and development demand increasing efficiency from scientists. High-quality decisions must be made faster and encompass all available information. At the same time there is a growing desire to better utilize the multi-billion dollar research investment recorded in laboratory notebooks and bioassay databases. Key values for data integration in a data exploration environment include gathering data from disparate E-notebooks and bioassay databases into a single searchable “virtual” system and increased discoverability by accessing data through a system designed for exploration. Key benefits are better chemistry decisions through easier access to broader data and reduced time for preparing patent filings. The ability to interlink in-house and reported assay data with in-house and published chemistry provides a data-rich environment for developing insights and predictive models. We will discuss our experience with integrating information from journals, patents, bio-assay databases, and E-lab notebooks to address these needs.
Building linked data large-scale chemistry platform - challenges, lessons and...Valery Tkachenko
Chemical databases have been around for decades, but in recent years we have observed a qualitative change from rather small in-house built proprietary databases to large-scale, open and increasingly complex chemistry knowledgebases. This tectonic shift has imposed new requirements for database design and system architecture as well as the implementation of completely new components and workflows which did not exist in chemical databases before. Probably the most profound change is being caused by the linked nature of modern resources - individual databases are becoming nodes and hubs of a huge and truly distributed web of knowledge. This change has important aspects such as data and format standards, interoperability, provenance, security, quality control and metainformation standards.
ChemSpider at the Royal Society of Chemistry was first public chemical database which incorporated rigorous quality control by introducing both community curation and automated quality checks at the scale of tens of millions of records. Yet we have come to realize that this approach may now be incomplete in a quickly changing world of linked data. In this presentation we will talk about challenges associated with building modern public and private chemical databases as well as lessons that we have learned from our past and present experience. We will also talk about solutions for some common problems.
The OntoChem IT Solutions GmbH ...
... was founded in 2015 as a purely IT-oriented offshoot of the OntoChem GmbH. Even before we had many years of experience and it has always been our mission to provide added value to our customers by helping them to navigate today’s complex information world by developing cognitive computing solutions, indexing intranet and internet data and applying semantic search solutions for pharmaceutical, material sciences and technology driven businesses.
We strive to support our customers with the most useful tools for knowledge discovery possible, encompassing up-to-date data sources, optimized ontologies and high-throughput semantic document processing and annotation techniques.
We create new knowledge from structured and unstructured data by extracting relationships thereby exploiting the full potential of full-text documents & databases while also scanning social media, news flows and analyzing web-pages.
We aim at an unprecedented, machine understanding of text and subsequent knowledge extraction and inference. The application of our methods towards chemical compounds and their properties supports our customers in generating intellectual property and their use as novel therapeutics, agrochemical products, nutraceuticals, cosmetics and in the field of novel materials.
It's our mission to provide added value to customers by:
developing and applying cognitive computing solutions
creating intranet and internet data indexing and semantic search solutions
Big Data analytics for technology driven businesses
supporting product development and surveillance.
We deliver useful tools for knowledge discovery for:
creating background knowledge ontologies
high-throughput semantic document processing and annotation
knowledge mining by extracting relationships
exploiting the full potential of full-text documents & databases while also scanning social media, news flows and analyzing web-pages.
This document discusses using Wiley's Chemistry Toolkit to plan syntheses of target molecules. It describes an approach using computer-aided synthesis design (CASD) to perform retrosynthetic analysis back to available starting materials, generating many alternatives and supporting them with literature examples. CASD works by extracting reaction rules from source reactions in databases like CIRX and applying chemical perception to identify compatible functional groups and regio- and stereoselectivity. The goal is to provide an integrated solution for data mining and synthesis planning.
The Open PHACTS project delivers an online platform integrating a wide variety of data from across chemistry and the life sciences and an ecosystem of tools and services to query this data in support of pharmacological research, turning the semantic web from a research project into something that can be used by practising medicinal chemists in both academia and industry. In the summer of 2015 it was the first winner of the European Linked Data Award. At the Royal Society of Chemistry we have provided the chemical underpinnings to this system and in this talk we review its development over the past five years. We cover both our early work on semantic modelling of chemistry data for the Open PHACTS triplestore and more recent work building an all-purpose data platform, for which the Open PHACTS data has been an important test case, what has worked well, what's missing and where this is is likely to go in future.
Alice: "What version of ChEMBL are we using?"
Bob: "Er…let me check. It's going to take a while, I'll get back to you."
This simple question took us the best part of a month to resolve and involved several individuals. Knowing the provenance of your data is essential, especially when using large complex systems that process multiple datasets.
The underlying issues of this simple question motivated us to improve the provenance data in the Open PHACTS project. We developed a guideline for dataset descriptions where the metadata is carried with the data. In this talk I will highlight the challenges we faced and give an overview of our metadata guidelines.
Presentation given to the W3C Semantic Web for Health Care and Life Sciences Interest Group on 14 January 2013.
Over the past decade, CDISC Standards have been widely accepted and implemented in clinical research. The FDA’s final “Guidance for Industry on electronic submission” mandates that submission data conform to CDISC standards such as SDTM, ADaM and SEND. This presentation will discuss how life sciences organizations can use Standards metadata to manage the regulatory compliance process. It will introduce how standards metadata management not only ensures regulatory compliance, but also supports process efficiency in clinical trial artefacts (e.g., protocol, CDASH, SDMT and ADaM) development and standards governance, and enables efficient communication between organizational units.
It will also introduce metadata management system and discuss how metadata management system will create, store, govern and manage standards. It will also show how standards metadata management system interacts with ETL system and dictates standards-driven clinical artefacts development.
Why ICT Fails in Healthcare: Software Maintenance and MaintainabilityKoray Atalag
This presentation was for a SERG seminar at the University of Auckland Department of Computer Science. I present why software maintenance is a barrier for adoption of IT in healthcare and the maintainability aspects based on ISO/IEC 9126 software quality standard quality model. I then present the preliminary results of my research here.
This document provides an overview of how LOINC (Logical Observation Identifiers Names and Codes) codes can be used with FHIR (Fast Healthcare Interoperability Resources). It discusses how LOINC codes are represented and used in various FHIR resources like Observation, Questionnaire, DiagnosticReport, etc. It also describes how FHIR terminology services can be used to retrieve information about LOINC codes and structures like parts, answer lists, and properties to build value sets. The document demonstrates how LOINC enhances interoperability when clinical data is coded with LOINC and accessible via FHIR.
FHIR Developer Days 2015. Study on db implementations for FHIR serverIgor Bossenko
Presentation describes different approaches for implementing database for FHIR server, that were considered during implementation of Nortal National Healthcare System (NHS) for Lithuania.
This document provides an overview of the openEHR CDR open source project called EHRbase. EHRbase aims to provide an open standard-compliant backend platform for electronic health records and clinical applications using the openEHR specification. It has a team of developers across multiple continents and uses modern development practices like Scrum and BDD. EHRbase provides a REST API and SDK for creating, querying, and managing openEHR objects in a clinical data repository, and also integrates with FHIR through a FHIR bridge. It is being used as the backend platform for a national COVID-19 system in Germany.
How to Manage APIs in your Enterprise for Maximum Reusability and GovernanceWSO2
This document discusses API management and governance in an enterprise. It covers the challenges of APIs existing in silos across an organization, and addresses how an API manager and governance registry can help manage the API lifecycle from planning to sharing. It also demonstrates how the WSO2 API management platform can be used to publish, protect, analyze usage of APIs across different environments.
Data Integration for Big Data (OOW 2016, Co-Presented With Oracle)Rittman Analytics
Oracle Data Integration Platform is a cornerstone for big data solutions that provides five core capabilities: business continuity, data movement, data transformation, data governance, and streaming data handling. It includes eight core products that can operate in the cloud or on-premise, and is considered the most innovative in areas like real-time/streaming integration and extract-load-transform capabilities with big data technologies. The platform offers a comprehensive architecture covering key areas like data ingestion, preparation, streaming integration, parallel connectivity, and governance.
OGF actively collaborates with other standards organizations through cooperative agreements to develop standards for distributed computing. OGF has relationships with groups like DMTF, ISO, SNIA, ETSI, ITU-T, and NIST to jointly develop standards for areas like cloud computing, identity management, and data formats. These collaborations help drive innovation while avoiding duplication of efforts between organizations.
How to Manage APIs in your Enterprise for Maximum Reusability and GovernanceHARMAN Services
Trying to form an API/service strategy to keep pace with the IoT revolution? Know how you can address issues and challenges your enterprise might face while implementing it and know how you can address the same.
This webinar will also explains how WSO2 API Manager and WSO2 Governance Registry have helped enterprises overcome the following challenges:
1. How the number of services and their users affect service discoverability, catalog, and re-usability.
2. Mistrust among producers and consumers
3. Reliability, stability, and availability of services
4. How externally built common and reusable services meet requirements (anti-patterns - NIH)
This document summarizes a presentation about the Federation Lab and OpenID Connect. The Federation Lab is an identity toolkit that automates testing of identity software to increase interoperability between providers and consumers using SAML and OpenID Connect. It is a GÉANT project in collaboration with industry and research partners. The presentation discusses challenges like interoperability issues that can arise from complex identity systems with many implementations and deployments. Federation Lab addresses this by performing over 100 automated test flows on identity providers to discover errors. It also provides debugging tools. The presentation contrasts identity flows and attribute returning between SAML and OpenID Connect. In closing, the Federation Lab testing tool is made available for participants to use.
Metrics Monitoring Is So Critical - What's Your Best Approach? Wavefront
Metrics monitoring is so critical for modern cloud applications. But can you do it with APM, with a log monitor, or with a specialized metrics platform? Open source or commercial? How are SaaS leaders monitoring their environments with metrics today?
Learn about unified metrics monitoring with real-time analytics, and why it’s the preferred methodology for assuring cloud application environments.
There are several approaches to implementing a metrics-monitoring platform. Depending on where you are on the metrics maturity curve, some platforms are better than others. Learn how to pick the approach that's best for you.
Whither the Hadoop Developer Experience, June Hadoop Meetup, Nitin MotgiFelicia Haggarty
The document discusses challenges with building operational data applications on Hadoop and introduces the Cask Data Application Platform (CDAP) as a solution. It provides an agenda that covers data applications, challenges, CDAP motivation and goals, use cases, and an introduction and architecture overview of CDAP. The document aims to demonstrate how CDAP provides a unified platform that simplifies application development and lifecycle while supporting reusable data and processing patterns.
Pivotal Digital Transformation Forum: Data Science Technical OverviewVMware Tanzu
This document provides an overview of Pivotal's data science capabilities and tools. It discusses how Pivotal uses an Agile approach to data science projects, focusing on frequent interactions with customers. Pivotal's software stack is designed to enable data science work by supporting real-time, interactive, and batch operations on data through tools like Spring XD, Pivotal HD, and GemFire. Examples are provided of how Pivotal has used these tools for applications like connected cars and scalable video analytics.
Technical deep dive on Java Micro Edition (ME) 8 (apologies for the partially messed up colors and slides - SlideShare is doing that during the conversion process)
From allotrope to reference master data management OSTHUS
We will present the updated Allotrope framework and cover .adf files and how they are used. We’ll demonstrate semantic modeling in .adf (OWL models + the SHACL constraint language). We’ll show how the data description layer in .adf can be extended via a “semantic hub” that we call Reference Master Data Management, which can be used across the enterprise. RMDM provides a means to integrate metadata about any data source within your enterprise – including structured, semi-structured and unstructured data. Customer examples from current project work will be given where possible. Last we’ll show scalability of this approach using data science techniques can be employed beyond just the metadata – we refer to this as Big Analysis.
Data proliferation from 7+ billion humans and 20+ billion devices from every walk of life has been the focus in the last decade. With the velocity, variety and volume of data, every data organization’s goal shifted to protecting and monetizing data from rapidly growing network of IOT embedded objects and sensors.
One of the true and tried business continuity methodology of storing and retrieving vast amount of data has been through replication of Hadoop systems on hybrid clouds and in geographically distributed data centers. Replication is similar to Blockchain using autonomous smart contracts instantiated on the metadata and data so that the replicated data follows a single source of truth.
Replicas can be maintained across geographically distributed data centers giving greater risk tolerance capabilities to the businesses continuity plan for the data-sets. With intelligent predictive analytics based on usage patterns, dynamic tiering policies can be triggered on the data sets to provide true value-add to the data. The temperature of the data is used to move data between hot/warm/cold/archival storage based on configurable policies leading to greater reduction in total cost of ownership.
Users in 2018 and beyond demand absolute availability of data as and when they desire. The dynamic data access management is fundamental concept to satisfy the business continuity plan. Seamless enterprise-grade disaster recovery to support business continuity use case has significant challenges around replicating security and governance on data-sets. In this talk we will discuss how the above challenge can be addressed for supporting seamless replication and disaster recovery for Hadoop-scale data. NIRU ANISETI, Product Manager, Hortonworks
The document discusses establishing a National Digital Repository System (NDRS) in India using a harvesting model. It analyzes different technical models (centralized, distributed, harvesting), and recommends adopting the harvesting model. The harvesting model would involve individual institutional repositories exposing their metadata using OAI-PMH for a central searchable server to harvest and provide enhanced discovery services. Benefits of the NDRS for various stakeholders are discussed. Current scenarios of institutional repositories in India and potential organizations to contribute to the proposed NDRS are also outlined.
Richard Bolton (GSK and Pistoia's ELN query services workstream coordinator) discusses the Alliance's chemistry strategy, which includes ELN query standards, hosted ELN, and chemistry externalization faciliation
1) NetApp is a leader in open source technologies like OpenStack, Docker, Kubernetes, and Mesosphere. It contributes code to many open source projects and supports customers using these technologies.
2) NetApp supports OpenStack through technologies like Cinder for block storage and Manila for shared file systems. It has contributed a significant amount of code to these projects.
3) NetApp helps customers use containers through technologies like Trident, its container storage orchestrator. It was one of the first vendors certified for Docker volumes and developed early dynamic provisioning for Kubernetes.
Machine Learning to Turbo-Charge the Ops Portion of DevOpsDeborah Schalm
Already on a continuous or short-cycle delivery? Constantly rewiring your apps with microservice and similar architectures? Maintaining visibility and maximizing service levels once this stuff gets into production could be a regular nightmare. Coding instrumentation into your apps is time-consuming and error-prone. Instead, let machine learning do the work of adapting your monitoring to your fast-moving application environments. In this webcast learn about various types of machine learning that are optimized for operational data, and see in a demo how this could be leveraged to ensure your ops move as fast as rest of your DevOps pipeline.
Contexti / Oracle - Big Data : From Pilot to ProductionContexti
The document discusses challenges in moving big data projects from pilots to production. It highlights that pilots have loose SLAs and focus on a few use cases and demonstrated insights, while production requires enforced SLAs, supporting many use cases and delivering actionable insights. Key challenges in the transition include establishing governance, skills, funding models and integrating insights into operations. The document also provides examples of technology considerations and common operating models for big data analytics.
OData External Data Integration Strategies for SaaSSumit Sarkar
This document discusses OData integration strategies for SaaS applications. It provides an overview of the OData standard and why SaaS vendors are adopting it. It then describes how Oracle Service Cloud uses OData accelerators to integrate with external data sources like Salesforce and Siebel. These accelerators allow agents to access and edit external data without leaving the Service Cloud interface.
MPLS/SDN 2013 Intercloud Standardization and Testbeds - SillAlan Sill
This talk givens an overview of several multi-SDO and cross-SDO activities to promote and spur innovation in cloud computing. The focus is on API development and standardization, including testbeds, test use cases, and collaborative activities between organizations to create and carry out development and testing in this area. The focus is on work being pursued through the Cloud and Autonomic Computing Center at Texas Tech University, which is part of the US National Science Foundation's Industry/University Cooperative Research Center, and on work being done by standards organizations such as the Open Grid Forum, Distributed Management Task Force, and Telecommunications Management Forum in which the CAC@TTU is involved. A summary is also given of work to produce a new round of more detailed use cases suitable for testing by the US National Institute of Standards and Technology's Standards Acceleration to Jumpstart Adoption of Cloud Computing (SAJACC) working group, with brief mention also given to other related work going on in this area in other parts of the world. Background and other standards work is also mentioned.
Apcera reviews the good, bad and the amazing, based on feedback collected from 250+ early adopters, of emerging microservices platforms and best practices.
You can learn more about The Trusted Cloud Platform at: https://www.apcera.com/
Similar to Pistoia Alliance European Conference 2015 - Gerhard Noelken / Allotrope Foundation (20)
Fairification experience clarifying the semantics of data matricesPistoia Alliance
This webinar presents the Statistics Ontology, STATO which is a semantic framework to support the creation of standardized analysis reports to help with review of results in the form of data matrices. STATO includes a hierarchy of classes and a vocabulary for annotating statistical methods used in life, natural and biomedical sciences investigations, text mining and statistical analyses.
This webinar discusses driving adoption of microphysiological systems (MPS) in drug R&D. The webinar agenda includes presentations on multi-organ chips for safety and efficacy assessment from TissUse, current applications and future perspectives of organ-on-chips in pharmaceutical industry from AstraZeneca, and driving adoption of MPS from ToxRox Consulting. A panel discussion will be moderated by Mary Ellen Cosenza. The presentations will cover benefits of MPS for reducing drug failures and animal testing, applications across drug discovery and development, challenges for adoption, and perspectives from industry.
Federated Learning (FL) is a learning paradigm that enables collaborative learning without centralizing datasets. In this webinar, NVIDIA present the concept of FL and discuss how it can help overcome some of the barriers seen in the development of AI-based solutions for pharma, genomics and healthcare. Following the presentation, the panel debate on other elements that could drive the adoption of digital approaches more widely and help answer currently intractable science and business questions.
It seems that AI is also becoming a buzzword, like design thinking. Everyone is talking about AI or wants to have AI, and sees all the ideas and benefits – that’s fine, but how do you get started? But what’s different now? Three innovations have finally put AI on the fast track: Big Data, with the internet and sensors everywhere; massive computing power, especially through the Cloud; and the development of breakthrough algorithms, so computers can be trained to accomplish more sophisticated tasks on their own with deep learning. If you use new technology, you need to explore and know what’s possible. With design thinking, it aids to outline the steps and define the ways in which you’re going to create the solution. Starting with mapping the customer journey, defining who will be using that service enhanced with intelligent technology, or who will benefit and gain value from it. We discuss how these two worlds are coming together, and how you get started to transform your venture with Artificial Intelligence using Design Thinking.
Speaker: Claudio Mirti, Principal Solution Specialist – Data & AI, Microsoft
Themes and objectives:
To position FAIR as a key enabler to automate and accelerate R&D process workflows
FAIR Implementation within the context of a use case
Grounded in precise outcomes (e.g. faster and bigger science / more reuse of data to enhance value / increased ability to share data for collaboration and partnership)
To make data actionable through FAIR interoperability
Speakers:
Mathew Woodwark,Head of Data Infrastructure and Tools, Data Science & AI, AstraZeneca
Erik Schultes, International Science Coordinator, GO-FAIR
Georges Heiter, Founder & CEO, Databiology
Knowledge graphs ilaria maresi the hyve 23apr2020Pistoia Alliance
Data for drug discovery and healthcare is often trapped in silos which hampers effective interpretation and reuse. To remedy this, such data needs to be linked both internally and to external sources to make a FAIR data landscape which can power semantic models and knowledge graphs.
2020.04.07 automated molecular design and the bradshaw platform webinarPistoia Alliance
This presentation described how data-driven chemoinformatics methods may automate much of what has historically been done by a medicinal chemist. It explored what is reasonable to expect “AI” approaches might achieve, and what is best left with a human expert. The implications of automation for the human-machine interface were explored and illustrated with examples from Bradshaw, GSK’s experimental automated design environment.
This presentation reviewed the challenges in identifying, acquiring and utilizing research data in relation to an evolving data market. Strategic solutions were examined in which the FAIR principles play a key role in the future of data management.
Dr. Dennis Wang discusses possible ways to enable ML methods to be more powerful for discovery and to reduce ambiguity within translational medicine, allowing data-informed decision-making to deliver the next generation of diagnostics and therapeutics to patients quicker, at lowered costs, and at scale.
The talk by Dr. Dennis Wang was followed by a panel discussion with Mr. Albert Wang, M. Eng., Head, IT Business Partner, Translational Research & Technologies, Bristol-Myers Squibb.
With the explosion of interest in both enhanced knowledge management and open science, the past few years have seen considerable discussion about making scientific data “FAIR” — findable, accessible, interoperable, and reusable. The problem is that most scientific datasets are not FAIR. When left to their own devices, scientists do an absolutely terrible job creating the metadata that describe the experimental datasets that make their way in online repositories. The lack of standardization makes it extremely difficult for other investigators to locate relevant datasets, to re-analyse them, and to integrate those datasets with other data. The Center for Expanded Data Annotation and Retrieval (CEDAR) has the goal of enhancing the authoring of experimental metadata to make online datasets more useful to the scientific community. The CEDAR work bench for metadata management will be presented in this webinar. CEDAR illustrates the importance of semantic technology to driving open science. It also demonstrates a means for simplifying access to scientific data sets and enhancing the reuse of the data to drive new discoveries.
Open interoperability standards, tools and services at EMBL-EBIPistoia Alliance
In this webinar Dr Henriette Harmse from EMBL-EBI presents how they are using their ontology services at EMBL-EBI to scale up the annotation of data and deliver added value through ontologies and semantics to their users.
Fair webinar, Ted slater: progress towards commercial fair data products and ...Pistoia Alliance
Elsevier is a global information analytics business that helps institutions and professional’s
advance healthcare and open science to improve performance for the benefit of humanity.
In this webinar, we discuss how Elsevier is increasingly leveraging the FAIR Guiding Principles to improve its products and services to better serve the scientific community.
Application of recently developed FAIR metrics to the ELIXIR Core Data ResourcesPistoia Alliance
The FAIR (Findable, Accessible, Interoperable and Reusable) principles aim to maximize the discovery and reuse of digital resources. Using recently developed software and metrics to assess FAIRness and supported through an ELIXIR Implementation Study, Michel worked with a subset of ELIXIR Core Data Resources to apply these technologies. In this webinar, he will discuss their approach, findings, and lessons learned towards the understanding and promotion of the FAIR principles.
Implementing Blockchain applications in healthcarePistoia Alliance
Blockchain technology can revolutionise the way information is exchanged between parties by bringing an unprecedented level of security and trust to these transactions. The technology is finding its way into multiple use cases but we are yet to see full adoption and real-world business implementation in the Healthcare industry.
In this webinar we will explore the main challenges and considerations for the implementation of Blockchain technology in Healthcare use cases. This is the third webinar in our Blockchain Education series.
Building trust and accountability - the role User Experience design can play ...Pistoia Alliance
In this webinar our panel of UX specialists give a brief introduction to User Experience before presenting the design opportunities UX can bring to AI. We all know that AI has great potential but has some significant hurdles to overcome not least so the human aspect of trust and ethical considerations when designing in the life sciences.
This document summarizes a webinar on using machine learning and data mining techniques to predict drug repurposing opportunities for chronic pancreatitis. Specifically:
1. Ensemble learning techniques like kernel-based models were used to analyze drug and disease target interaction data from multiple sources to identify potential drug candidates for repurposing.
2. The top 5 repurposing candidates identified through this process were being evaluated further by the partner organization Mission-Cure with the goal of beginning patient trials by January 2020.
3. Additional techniques discussed included using compressed sensing to analyze drug-disease networks and predict side effects to help evaluate candidate drugs identified for repurposing opportunities.
PA webinar on benefits & costs of FAIR implementation in life sciences Pistoia Alliance
The slides from the Pistoia Alliance Debates Webinar where a panel of experts from technology support providers and the biopharma industry, who have been invited to share their views on the "Benefits and costs of FAIR Implementation for life science industry".
Creating novel drugs is an extraordinarily hard and complex problem.
One of the many challenges in drug design is the sheer size of the search space for novel chemical compounds. Scientists need to find molecules that are active toward a biological target or pathway and at the same time have acceptable ADMET properties.
There is now considerable research going on using various AI and ML approaches to tackle these challenges.
Our distinguished speakers, Drs. Alex Tropsha and Ola Engkvist, will discuss their recent work in Drug Design involving Deep Reinforcement Learning and Neural Networks, and will answer questions from the audience on the current state of the research in the field.
Speakers:
Prof Alex Tropsha, Professor at University of North Carolina at Chapel Hill, USA
Dr. Ola Engkvist, Associate Director at AstraZeneca R&D, Gothenburg, Sweden
Alexander Tropsha presented on using AI and machine learning for drug design and discovery. He discussed using QSAR models to predict properties and activity of molecules based on their structural descriptors. He also introduced ReLeaSE, a new method using deep reinforcement learning to generate novel drug-like molecules and guide chemical library design through a thought cycle of molecule generation, model building, and iterative improvement. If successful, this approach could disrupt traditional computational drug discovery pipelines.
The cost of acquiring information by natural selectionCarl Bergstrom
This is a short talk that I gave at the Banff International Research Station workshop on Modeling and Theory in Population Biology. The idea is to try to understand how the burden of natural selection relates to the amount of information that selection puts into the genome.
It's based on the first part of this research paper:
The cost of information acquisition by natural selection
Ryan Seamus McGee, Olivia Kosterlitz, Artem Kaznatcheev, Benjamin Kerr, Carl T. Bergstrom
bioRxiv 2022.07.02.498577; doi: https://doi.org/10.1101/2022.07.02.498577
PPT on Alternate Wetting and Drying presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
Immersive Learning That Works: Research Grounding and Paths ForwardLeonel Morgado
We will metaverse into the essence of immersive learning, into its three dimensions and conceptual models. This approach encompasses elements from teaching methodologies to social involvement, through organizational concerns and technologies. Challenging the perception of learning as knowledge transfer, we introduce a 'Uses, Practices & Strategies' model operationalized by the 'Immersive Learning Brain' and ‘Immersion Cube’ frameworks. This approach offers a comprehensive guide through the intricacies of immersive educational experiences and spotlighting research frontiers, along the immersion dimensions of system, narrative, and agency. Our discourse extends to stakeholders beyond the academic sphere, addressing the interests of technologists, instructional designers, and policymakers. We span various contexts, from formal education to organizational transformation to the new horizon of an AI-pervasive society. This keynote aims to unite the iLRN community in a collaborative journey towards a future where immersive learning research and practice coalesce, paving the way for innovative educational research and practice landscapes.
Authoring a personal GPT for your research and practice: How we created the Q...Leonel Morgado
Thematic analysis in qualitative research is a time-consuming and systematic task, typically done using teams. Team members must ground their activities on common understandings of the major concepts underlying the thematic analysis, and define criteria for its development. However, conceptual misunderstandings, equivocations, and lack of adherence to criteria are challenges to the quality and speed of this process. Given the distributed and uncertain nature of this process, we wondered if the tasks in thematic analysis could be supported by readily available artificial intelligence chatbots. Our early efforts point to potential benefits: not just saving time in the coding process but better adherence to criteria and grounding, by increasing triangulation between humans and artificial intelligence. This tutorial will provide a description and demonstration of the process we followed, as two academic researchers, to develop a custom ChatGPT to assist with qualitative coding in the thematic data analysis process of immersive learning accounts in a survey of the academic literature: QUAL-E Immersive Learning Thematic Analysis Helper. In the hands-on time, participants will try out QUAL-E and develop their ideas for their own qualitative coding ChatGPT. Participants that have the paid ChatGPT Plus subscription can create a draft of their assistants. The organizers will provide course materials and slide deck that participants will be able to utilize to continue development of their custom GPT. The paid subscription to ChatGPT Plus is not required to participate in this workshop, just for trying out personal GPTs during it.
(June 12, 2024) Webinar: Development of PET theranostics targeting the molecu...Scintica Instrumentation
Targeting Hsp90 and its pathogen Orthologs with Tethered Inhibitors as a Diagnostic and Therapeutic Strategy for cancer and infectious diseases with Dr. Timothy Haystead.
Evidence of Jet Activity from the Secondary Black Hole in the OJ 287 Binary S...Sérgio Sacani
Wereport the study of a huge optical intraday flare on 2021 November 12 at 2 a.m. UT in the blazar OJ287. In the binary black hole model, it is associated with an impact of the secondary black hole on the accretion disk of the primary. Our multifrequency observing campaign was set up to search for such a signature of the impact based on a prediction made 8 yr earlier. The first I-band results of the flare have already been reported by Kishore et al. (2024). Here we combine these data with our monitoring in the R-band. There is a big change in the R–I spectral index by 1.0 ±0.1 between the normal background and the flare, suggesting a new component of radiation. The polarization variation during the rise of the flare suggests the same. The limits on the source size place it most reasonably in the jet of the secondary BH. We then ask why we have not seen this phenomenon before. We show that OJ287 was never before observed with sufficient sensitivity on the night when the flare should have happened according to the binary model. We also study the probability that this flare is just an oversized example of intraday variability using the Krakow data set of intense monitoring between 2015 and 2023. We find that the occurrence of a flare of this size and rapidity is unlikely. In machine-readable Tables 1 and 2, we give the full orbit-linked historical light curve of OJ287 as well as the dense monitoring sample of Krakow.
JAMES WEBB STUDY THE MASSIVE BLACK HOLE SEEDSSérgio Sacani
The pathway(s) to seeding the massive black holes (MBHs) that exist at the heart of galaxies in the present and distant Universe remains an unsolved problem. Here we categorise, describe and quantitatively discuss the formation pathways of both light and heavy seeds. We emphasise that the most recent computational models suggest that rather than a bimodal-like mass spectrum between light and heavy seeds with light at one end and heavy at the other that instead a continuum exists. Light seeds being more ubiquitous and the heavier seeds becoming less and less abundant due the rarer environmental conditions required for their formation. We therefore examine the different mechanisms that give rise to different seed mass spectrums. We show how and why the mechanisms that produce the heaviest seeds are also among the rarest events in the Universe and are hence extremely unlikely to be the seeds for the vast majority of the MBH population. We quantify, within the limits of the current large uncertainties in the seeding processes, the expected number densities of the seed mass spectrum. We argue that light seeds must be at least 103 to 105 times more numerous than heavy seeds to explain the MBH population as a whole. Based on our current understanding of the seed population this makes heavy seeds (Mseed > 103 M⊙) a significantly more likely pathway given that heavy seeds have an abundance pattern than is close to and likely in excess of 10−4 compared to light seeds. Finally, we examine the current state-of-the-art in numerical calculations and recent observations and plot a path forward for near-future advances in both domains.
When I was asked to give a companion lecture in support of ‘The Philosophy of Science’ (https://shorturl.at/4pUXz) I decided not to walk through the detail of the many methodologies in order of use. Instead, I chose to employ a long standing, and ongoing, scientific development as an exemplar. And so, I chose the ever evolving story of Thermodynamics as a scientific investigation at its best.
Conducted over a period of >200 years, Thermodynamics R&D, and application, benefitted from the highest levels of professionalism, collaboration, and technical thoroughness. New layers of application, methodology, and practice were made possible by the progressive advance of technology. In turn, this has seen measurement and modelling accuracy continually improved at a micro and macro level.
Perhaps most importantly, Thermodynamics rapidly became a primary tool in the advance of applied science/engineering/technology, spanning micro-tech, to aerospace and cosmology. I can think of no better a story to illustrate the breadth of scientific methodologies and applications at their best.
Travis Hills of MN is Making Clean Water Accessible to All Through High Flux ...Travis Hills MN
By harnessing the power of High Flux Vacuum Membrane Distillation, Travis Hills from MN envisions a future where clean and safe drinking water is accessible to all, regardless of geographical location or economic status.
The root of the problem is the absence of standards in the current laboratory software environment.
Substantial funds are invested into treating the symptoms of our data management problems
We patch local gaps in the software, fix specific problems, and investment in large integration efforts,
But are left with the fundamental, underlying root problems
Lack of standard data format: Each instrument may have its own format created. Conversion is needed in order to read the data.
Lack of standards for metadata: The contextual metadata describing analytical methods, instruments, processes, and even the reasons for doing an experiment are inconsistent, incomplete, and often spread over multiple applications in the analytical workflow
Lack of standard interface between software applications: Finally, because the software we use across the analytical workflow is often from different sources, they are written using different technologies, and read and write different formats, so don’t often talk to one another without a custom integration effort
Transition to next slide: These are a few examples of issues that affect our entire analytical data lifecycle
While the Pharmaceutical Company member reps make fund the work, and provide subject matter experts, the Foundation and Project benefit from excellent pofessional partnership and support of
DBR for the Project and Consortium management, as well as legal, logistical and scientific advise. They are not only professionals at running an organization like this, (so do a better job than if it were left up to the scientists), it allows the members to focus on the technical and strategic issues.
Osthus is the professional data and systems integrator we have partnered with to engineer and build the framework.
As we will discuss more later- we created a partner program to enable the collaboration with instrument and software vendors
Data standards are necessary but NOT sufficient to solve the problems- the solution requires a holistic approach to build standards into the software used throughout the data lifecycle.
Without adoption, a data standard alone is an abstract that cannot alone create change, the standard needs to be adopted, which ultimately means it has to be integrated into the software used to generate and manage data. The Allotrope Framework embodies a three pronged approach to driving adoption:
Document standard
Store data and metadata in a vendor agnostic, common, non-proprietary file format
Ready for Archiving
Easy data sharing & retrieval
Taxonomies- Metadata repository
Ensures accurate, complete, & consistent experiment context is stored along with data
Reusable software components
Provide access to data, metadata, and business objects
Available for integration with vendor or in-house software
No single public standard that covers them all
2 imporetant concepts in the ADF design- first is the landscape of technicques
Over the course of that workflow, important context, or metadata is created- starting with things like the purpose, project, whether or not it’s a GxP process, etc
In the planning stage experimental and instrument parameters are established, details of the analysis and any fitting algorithm contextualize results.
All of this is context we would like to have to be to find and analyze the data later, but it’s often incompletely captured, captured as free text, or spread over multiple software applications
Ultimately the collective meta data is EVIDENCE that supports a DECISION about your MANUFACTURING PROCESS or MATERIAL…
That standardization of the interface between software applications also creates a more plug-and-play environment, making it much easier to substitute one brand of instrument for another, use a different analysis application to access and alternative processing algorithm, or simply avoid having to rebuild custom interfaces when one of your pieces of software requires an upgrade.
Finally, reporting is more powerful because it’s based on meaningful criteria, and uses an index based on standardized terminology, and an index which has been built using the complete and consistent captured context of everything that happened along the workflow
HDF5 is a data model, library, and file format for storing and managing data. It supports an unlimited variety of data types, and is designed for flexible and efficient I/O and for high volume and complex data. HDF5 is portable and is extensible
Only the start! ADF design & practices being developed facilitate rapid taxonomy development with SMES, and thus extension to additional techniques
Standards
Evaluated > 100 public standards against scientific and business requirements across the full data lifecycle, from creation to archiving
Developed reference architecture for data archiving based on public standards
Federated select standards and ontologies for use by the Framework
Development
Created first version of Framework (pre-release), with class libraries for ADF, metadata repository and data archive
Created proof-of-concept software and delivered to all members
Benchmarked ADF performance using MS data
Launched the Allotrope Partner Network to partner with instrument and software vendors to facilitate adoption
Initiated interactions with FDA
Second important concept in ADF design is the dimension of the process in which analytical measurements are used
Drivers of adoption
Adoption of the standards by way of the Allotrope Framework standardize the data format, the contextual metadata (the orange box) and the inputs and outputs between software applications (the orange arrows). The metadata will be stored with the data.
This additionally opens up significantly more opportunity for automating the workflow- for example when the instrument parameters can be automatically sent to the instrument from the LIMS or ELN where they are first selected, removing the need for a human to read a document created in one system, and reenter the parameters in another.