This presentation was provided by Patricia Brennan of The National Library of Medicine, during the NISO event "Defining The Library: The Mission, Role and Community," held on March 13, 2019.
1. The document summarizes Kerstin Forsberg's presentation on semantics and linked data at AstraZeneca R&D. It discusses (1) an internal competitive intelligence tool called CI360, (2) public pre-competitive projects like Open PHACTS and standards bodies, (3) AstraZeneca's Linked Data Community of Practice, and (4) ongoing work on study identifiers and APIs.
2. It provides an overview of Kerstin Forsberg's background and goal of improving the utility of clinical trial data through semantic interoperability. It also outlines some of AstraZeneca's collaborations and contributions to linked data initiatives.
3. The presentation highlights AstraZeneca
Connecting Dataverse with the Research Life CycleMerce Crosas
This document outlines how data repositories can connect with the research lifecycle. It discusses Dataverse, an open source data repository platform developed by IQSS at Harvard University. It describes several application programming interfaces (APIs) that can be used for interoperating with Dataverse, including OAI-PMH, SWORD, search, and data access APIs. It also lists several integrations that are implemented or in progress to connect Dataverse with other systems like the Open Science Framework (OSF) and support functions like depositing data management plans, data curation, assigning DOIs, and more.
BibBase Linked Data Triplification Challenge 2010 PresentationReynold Xin
The document summarizes BibBase Triplified, a system that publishes bibliographic data from BibTeX files as structured data on the semantic web. It takes BibTeX files maintained by scientists, detects and resolves duplicates, and publishes the data as HTML pages and RDF triples. It also links entries to external datasets like DBLP and DBpedia. As of September 2010, the system had over 4,500 publications and 100 active users. Future work includes improving duplicate detection, linking to more external sources, and broadening the user base.
CrossMark: Standardizing Funding Information in Scholarly Journal Articles 20...Crossref
H. Frederick Dylla discusses standardizing funding information in scholarly articles through CrossRef's CrossMark initiative. He notes that CrossRef contributed to the US Scholarly Publishing Roundtable discussions on access and interoperability. CrossMark allows inclusion of funding agency attribution, which benefits researchers, funders, and publishers by increasing transparency, consistency, and ability to monitor compliance. Next steps include finalizing the CrossRef standard for funding data and encouraging adoption.
1. The FAIRsharing registry curates records on standards, databases, and data policies to improve their discoverability and reuse.
2. FAIRsharing provides recommendations to guide users and producers in selecting and describing these resources for data sharing.
3. The registry is used by projects like ELIXIR, GO-FAIR, and NIH to support their work in research data management.
Reading Group: From Database to DataspacesJürgen Umbrich
The document discusses the concept of dataspaces and dataspace support systems (DSSPs) as a new approach to data management. It describes dataspaces as loosely connected data sources of various formats that are not fully integrated but exist together. DSSPs are proposed to offer services like search, querying, monitoring, and discovery across heterogeneous dataspace participants with varying degrees of control and consistency guarantees. Key challenges discussed include data modeling and querying across different formats, automated discovery of relationships between data sources, and developing theoretical foundations.
Presentation slides from a talk by Gareth Knight which discussed the need to consider data sharing activities in academic citizenship, different approaches that may be taken to publish data associated with publications, and the opportunities presented by data journals
This presentation was provided by Patricia Brennan of The National Library of Medicine, during the NISO event "Defining The Library: The Mission, Role and Community," held on March 13, 2019.
1. The document summarizes Kerstin Forsberg's presentation on semantics and linked data at AstraZeneca R&D. It discusses (1) an internal competitive intelligence tool called CI360, (2) public pre-competitive projects like Open PHACTS and standards bodies, (3) AstraZeneca's Linked Data Community of Practice, and (4) ongoing work on study identifiers and APIs.
2. It provides an overview of Kerstin Forsberg's background and goal of improving the utility of clinical trial data through semantic interoperability. It also outlines some of AstraZeneca's collaborations and contributions to linked data initiatives.
3. The presentation highlights AstraZeneca
Connecting Dataverse with the Research Life CycleMerce Crosas
This document outlines how data repositories can connect with the research lifecycle. It discusses Dataverse, an open source data repository platform developed by IQSS at Harvard University. It describes several application programming interfaces (APIs) that can be used for interoperating with Dataverse, including OAI-PMH, SWORD, search, and data access APIs. It also lists several integrations that are implemented or in progress to connect Dataverse with other systems like the Open Science Framework (OSF) and support functions like depositing data management plans, data curation, assigning DOIs, and more.
BibBase Linked Data Triplification Challenge 2010 PresentationReynold Xin
The document summarizes BibBase Triplified, a system that publishes bibliographic data from BibTeX files as structured data on the semantic web. It takes BibTeX files maintained by scientists, detects and resolves duplicates, and publishes the data as HTML pages and RDF triples. It also links entries to external datasets like DBLP and DBpedia. As of September 2010, the system had over 4,500 publications and 100 active users. Future work includes improving duplicate detection, linking to more external sources, and broadening the user base.
CrossMark: Standardizing Funding Information in Scholarly Journal Articles 20...Crossref
H. Frederick Dylla discusses standardizing funding information in scholarly articles through CrossRef's CrossMark initiative. He notes that CrossRef contributed to the US Scholarly Publishing Roundtable discussions on access and interoperability. CrossMark allows inclusion of funding agency attribution, which benefits researchers, funders, and publishers by increasing transparency, consistency, and ability to monitor compliance. Next steps include finalizing the CrossRef standard for funding data and encouraging adoption.
1. The FAIRsharing registry curates records on standards, databases, and data policies to improve their discoverability and reuse.
2. FAIRsharing provides recommendations to guide users and producers in selecting and describing these resources for data sharing.
3. The registry is used by projects like ELIXIR, GO-FAIR, and NIH to support their work in research data management.
Reading Group: From Database to DataspacesJürgen Umbrich
The document discusses the concept of dataspaces and dataspace support systems (DSSPs) as a new approach to data management. It describes dataspaces as loosely connected data sources of various formats that are not fully integrated but exist together. DSSPs are proposed to offer services like search, querying, monitoring, and discovery across heterogeneous dataspace participants with varying degrees of control and consistency guarantees. Key challenges discussed include data modeling and querying across different formats, automated discovery of relationships between data sources, and developing theoretical foundations.
Presentation slides from a talk by Gareth Knight which discussed the need to consider data sharing activities in academic citizenship, different approaches that may be taken to publish data associated with publications, and the opportunities presented by data journals
The Pistoia Alliance: Update on Strategy and ProgressPistoia Alliance
Ramesh Durvasula, Pistoia Alliance board member, discusses the Pistoia Alliance mission and recaps activities in 2011-12, with particular emphasis on the successful completion of the Sequence Squeeze Competition and Sequence Services Phase 2. The presentation was delivered at BioITWorld in Boston in April 2012.
The Pistoia Alliance Conference in April 2011 included a series of 10-minute "lightning talks" from vendors about what they think pharma will look like in 2020. This presentation was delivered by John P. McAlister of CambridgeSoft.
Emerging Life Sciences Collaboration on Common Service SpecificationPistoia Alliance
Presentation by Pistoia Alliance reps Ian Harrow (Pfizer) and Nick Lynch (AstraZeneca) at the International Conference on Trends for Scientific Information Professionals, October 2010.
Pablo Castro, inventor of the OData protocol at Microsoft, describes the application of OData to the management of life science data in a presentation delivered in an open meeting and webinar of the Pistoia Alliance Technical Committee.
The document discusses the challenges of complying with legislation controlling substances across different countries and regions. It describes the Controlled Substance Compliance Service (CSCS) project led by Pistoia Alliance to address this issue. The project defined requirements, issued a request for proposals, and selected two vendors - ChemAxon/Patcore and Scitegrity - to develop controlled substance identification solutions using different technologies. The project helped establish a need for such solutions and a community of experts to address the global controlled substance compliance problem.
This presentation outlines a mechanism for using the power of "Big Data", social networking and technology infrastructure to speed the process of curing a horrible disease.
Edge Informatics and FAIR (Findable, Accessible, Interoperable and Reusable) ...Tom Plasterer
Edge Informatics is an approach to accelerate collaboration in the BioPharma pipeline. By combining technical and social solutions knowledge can be shared and leveraged across the multiple internal and external silos participating in the drug development process. This is accomplished by making data assets findable, accessible, interoperable and reusable (FAIR). Public consortia and internal efforts embracing FAIR data and Edge Informatics are highlighted, in both preclinical and clinical domains.
This talk was presented at the Molecular Medicine Tri-Conference in San Francisco, CA on February 20, 2017
This document discusses the shift from broad consent to dynamic consent for biobanks. It argues that broad consent, which allows a wide range of future research, has limitations including ongoing risks to individuals from identifiable data and an inability to withdraw consent from specific studies. Dynamic consent, which involves ongoing communication and allows individuals to set preferences for how their data is used, addresses these issues and fosters trust between researchers and participants. While technology is needed to implement dynamic consent, it could increase participation and benefit both research and individuals.
Project Data Sphere was started in 2013 by the CEO Roundtable on Cancer to facilitate sharing of clinical trial data in order to accelerate cancer research efforts. It provides a platform for sharing comparator arm data from cancer clinical trials conducted by pharmaceutical companies and academic organizations. The initial data includes comparator arm information from 10 clinical trials contributed by 7 companies and 3 academic groups, with a goal of including 50 trials by the end of 2014. The project aims to source novel analyses from the research community and show that responsible open access data sharing is possible, while addressing issues of privacy, access, and establishing data standards.
Project Data Sphere was started in 2013 by the CEO Roundtable on Cancer to facilitate sharing of clinical trial data in order to accelerate cancer research efforts. It provides a platform for sharing comparator arm data from cancer clinical trials conducted by pharmaceutical companies and academic organizations. The initial data includes comparator arm information from 10 clinical trials contributed by 7 companies and 3 academic groups, with a goal of including 50 trials by the end of 2014. The project aims to source novel analyses from the research community and show that responsible open access data sharing is possible, while addressing issues of privacy, access, and establishing data standards.
A FAIR Data Sharing Framework for Large-Scale Human Cancer ProteogenomicsBrett Tully
A FAIR Data Sharing Framework for Large-Scale Human Cancer Proteogenomics
Islam M1,2, Christiansen J3, Mahboob S4, Valova V4, Baker M4, Capes-Davis D4, Hains P4, Balleine R1,4, Zhong Q1,4, Reddel R1,4, Robinson P1,4, Tully B4
1 The University of Sydney, Camperdown, Sydney, NSW, 2050, Australia
2 Intersect, Level 13/50 Carrington St, Sydney, NSW, 2000, Australia
3 Queensland Cyber Infrastructure Foundation Ltd, Axon Building 47, University of Queensland, St Lucia, Brisbane, QLD, 4072, Australia
4 Children’s Medical Research Institute, Westmead, NSW, 2145, Australia
Background
The ACRF International Centre for the Proteome of Cancer (ProCan) at Children’s Medical Research Institute (CMRI) is an “industrial scale” program specialising in small-sample proteomics analysis from human cancer tissue.
ProCan seeks to generate both a wide and deep analytics pipeline and requires an enabling data framework. The framework must accommodate initial analysis and proteomic profiling of a large number of tumor samples, along with the clinical and demographic information, subsequent multi-omics studies, and any previously recorded responses to treatment. The curated datasets will provide a valuable resource beyond their primary use and ProCan is committed to making its data accessible to collaborators and the wider scientific community.
Objectives
The objective of the project is to an establish efficient, reliable, secure and ethical data sharing and publication framework based on the best practice data sharing principles, such as the FAIR principle. The framework must address various challenges that stem from the scale and complexity of the program, and ProCan’s focus on human-derived data and associated challenges presented in sharing these data while maintaining the privacy of any research participants.
Method
The project adopted a requirements-driven methodology and engaged with a wide range of ProCan stakeholders nationally and internationally. Together, various industrial-scale proteomics data management and sharing scenarios were explored such that robust and ethical sharing of the data would be achieved.
Results
The project developed a data sharing framework based on the FAIR principle that currently forms the basis of ongoing implementation work within the ProCan program.
John Koch presents on Merck's creation of an agile and scalable data science platform to improve information findability and accessibility for research and development. The platform addresses the challenges of thousands of people, information types, and repositories and hundreds of teams and decisions. It combines internal and external data using NoSQL tools on a flexible platform. This enables faster, integrated analysis across data sets compared to previous fragmented systems. The first analytics built on the platform were a scientific search application and a pharmacology knowledgebase. The flexible platform and integrated multi-disciplinary teams allow adding more data sources and analytics over time to further improve decision making.
Will Biomedical Research Fundamentally Change in the Era of Big Data?Philip Bourne
This document discusses how biomedical research may fundamentally change in the era of big data. It notes that biomedical research has always been data-driven, but the scope, variety, complexity and volume of data is now much greater. It also discusses the need for more open data sharing and new tools and methods for large-scale analysis. The document suggests biomedical research may move towards a more collaborative "platform" model, as seen with companies like Airbnb, with the goal of improving data access, reuse and reproducibility of research. However, overcoming challenges like incentives, trust and work practices will be important for any new platform to succeed.
Whitepaper developed with Pharma Exec magazine on how EIM- Enterprise Information Management- can provide efficiency and kick start innovation by ensuring information flows correctly inside- and outside- the company
With the recent announcement that GlaxoSmithKline have released a huge tranche of whole cell malaria screening data to the public domain, accompanied by a corresponding publication, this raises some issues for consideration before this exemplar instance becomes a trend. We have examined the data from a high level, by studying the molecular properties, and consider the various alerts presently in use by major pharma companies. We acknowledge the potential value of such data but also raise the issue of the actual value of such datasets released into the public domain. We also suggest approaches that could enhance the value of such datasets to the community and theoretically offer more immediate benefit to the search for leads for other neglected diseases.
The Path to Open Science with Illustrations from Computational Biology - A presentation made at the Microsoft 2011 Latin America Faculty Summit Cartagena, Columbia, May 18, 2011.
Quorum Review's April 2014 Institution Bulletin includes a letter from CEO, Cami Gearhart, JD, discussing the need to reform the regulations around biobanking and repositories, HIPAA templates and procedures, and the lessons IRBs can learn from 2013 FDA Warning Letters.
The document summarizes a webinar discussing health and medical data sharing. It covered the following key points:
1) Representatives from ANDS, NHMRC, and Wiley discussed policies around sharing health and medical research data. NHMRC encourages data sharing to maximize benefits from limited resources and ensure ethical use of data.
2) NHMRC's guidelines in its National Statement aim to minimize risks of unauthorized data use while maximizing ethical data sharing. Guidelines prompt consideration of risks but don't mandate specific mechanisms.
3) Wiley is committed to open science through open access publishing, improving data reproducibility, and using open standards to increase research discoverability.
In the world of pharma precompetitive efforts are increasing. These developments have created a dynamic ecosystem with pharma as smaller nodes in a complex network, in which collaborations have become an important business model.
The Pistoia Alliance: Update on Strategy and ProgressPistoia Alliance
Ramesh Durvasula, Pistoia Alliance board member, discusses the Pistoia Alliance mission and recaps activities in 2011-12, with particular emphasis on the successful completion of the Sequence Squeeze Competition and Sequence Services Phase 2. The presentation was delivered at BioITWorld in Boston in April 2012.
The Pistoia Alliance Conference in April 2011 included a series of 10-minute "lightning talks" from vendors about what they think pharma will look like in 2020. This presentation was delivered by John P. McAlister of CambridgeSoft.
Emerging Life Sciences Collaboration on Common Service SpecificationPistoia Alliance
Presentation by Pistoia Alliance reps Ian Harrow (Pfizer) and Nick Lynch (AstraZeneca) at the International Conference on Trends for Scientific Information Professionals, October 2010.
Pablo Castro, inventor of the OData protocol at Microsoft, describes the application of OData to the management of life science data in a presentation delivered in an open meeting and webinar of the Pistoia Alliance Technical Committee.
The document discusses the challenges of complying with legislation controlling substances across different countries and regions. It describes the Controlled Substance Compliance Service (CSCS) project led by Pistoia Alliance to address this issue. The project defined requirements, issued a request for proposals, and selected two vendors - ChemAxon/Patcore and Scitegrity - to develop controlled substance identification solutions using different technologies. The project helped establish a need for such solutions and a community of experts to address the global controlled substance compliance problem.
This presentation outlines a mechanism for using the power of "Big Data", social networking and technology infrastructure to speed the process of curing a horrible disease.
Edge Informatics and FAIR (Findable, Accessible, Interoperable and Reusable) ...Tom Plasterer
Edge Informatics is an approach to accelerate collaboration in the BioPharma pipeline. By combining technical and social solutions knowledge can be shared and leveraged across the multiple internal and external silos participating in the drug development process. This is accomplished by making data assets findable, accessible, interoperable and reusable (FAIR). Public consortia and internal efforts embracing FAIR data and Edge Informatics are highlighted, in both preclinical and clinical domains.
This talk was presented at the Molecular Medicine Tri-Conference in San Francisco, CA on February 20, 2017
This document discusses the shift from broad consent to dynamic consent for biobanks. It argues that broad consent, which allows a wide range of future research, has limitations including ongoing risks to individuals from identifiable data and an inability to withdraw consent from specific studies. Dynamic consent, which involves ongoing communication and allows individuals to set preferences for how their data is used, addresses these issues and fosters trust between researchers and participants. While technology is needed to implement dynamic consent, it could increase participation and benefit both research and individuals.
Project Data Sphere was started in 2013 by the CEO Roundtable on Cancer to facilitate sharing of clinical trial data in order to accelerate cancer research efforts. It provides a platform for sharing comparator arm data from cancer clinical trials conducted by pharmaceutical companies and academic organizations. The initial data includes comparator arm information from 10 clinical trials contributed by 7 companies and 3 academic groups, with a goal of including 50 trials by the end of 2014. The project aims to source novel analyses from the research community and show that responsible open access data sharing is possible, while addressing issues of privacy, access, and establishing data standards.
Project Data Sphere was started in 2013 by the CEO Roundtable on Cancer to facilitate sharing of clinical trial data in order to accelerate cancer research efforts. It provides a platform for sharing comparator arm data from cancer clinical trials conducted by pharmaceutical companies and academic organizations. The initial data includes comparator arm information from 10 clinical trials contributed by 7 companies and 3 academic groups, with a goal of including 50 trials by the end of 2014. The project aims to source novel analyses from the research community and show that responsible open access data sharing is possible, while addressing issues of privacy, access, and establishing data standards.
A FAIR Data Sharing Framework for Large-Scale Human Cancer ProteogenomicsBrett Tully
A FAIR Data Sharing Framework for Large-Scale Human Cancer Proteogenomics
Islam M1,2, Christiansen J3, Mahboob S4, Valova V4, Baker M4, Capes-Davis D4, Hains P4, Balleine R1,4, Zhong Q1,4, Reddel R1,4, Robinson P1,4, Tully B4
1 The University of Sydney, Camperdown, Sydney, NSW, 2050, Australia
2 Intersect, Level 13/50 Carrington St, Sydney, NSW, 2000, Australia
3 Queensland Cyber Infrastructure Foundation Ltd, Axon Building 47, University of Queensland, St Lucia, Brisbane, QLD, 4072, Australia
4 Children’s Medical Research Institute, Westmead, NSW, 2145, Australia
Background
The ACRF International Centre for the Proteome of Cancer (ProCan) at Children’s Medical Research Institute (CMRI) is an “industrial scale” program specialising in small-sample proteomics analysis from human cancer tissue.
ProCan seeks to generate both a wide and deep analytics pipeline and requires an enabling data framework. The framework must accommodate initial analysis and proteomic profiling of a large number of tumor samples, along with the clinical and demographic information, subsequent multi-omics studies, and any previously recorded responses to treatment. The curated datasets will provide a valuable resource beyond their primary use and ProCan is committed to making its data accessible to collaborators and the wider scientific community.
Objectives
The objective of the project is to an establish efficient, reliable, secure and ethical data sharing and publication framework based on the best practice data sharing principles, such as the FAIR principle. The framework must address various challenges that stem from the scale and complexity of the program, and ProCan’s focus on human-derived data and associated challenges presented in sharing these data while maintaining the privacy of any research participants.
Method
The project adopted a requirements-driven methodology and engaged with a wide range of ProCan stakeholders nationally and internationally. Together, various industrial-scale proteomics data management and sharing scenarios were explored such that robust and ethical sharing of the data would be achieved.
Results
The project developed a data sharing framework based on the FAIR principle that currently forms the basis of ongoing implementation work within the ProCan program.
John Koch presents on Merck's creation of an agile and scalable data science platform to improve information findability and accessibility for research and development. The platform addresses the challenges of thousands of people, information types, and repositories and hundreds of teams and decisions. It combines internal and external data using NoSQL tools on a flexible platform. This enables faster, integrated analysis across data sets compared to previous fragmented systems. The first analytics built on the platform were a scientific search application and a pharmacology knowledgebase. The flexible platform and integrated multi-disciplinary teams allow adding more data sources and analytics over time to further improve decision making.
Will Biomedical Research Fundamentally Change in the Era of Big Data?Philip Bourne
This document discusses how biomedical research may fundamentally change in the era of big data. It notes that biomedical research has always been data-driven, but the scope, variety, complexity and volume of data is now much greater. It also discusses the need for more open data sharing and new tools and methods for large-scale analysis. The document suggests biomedical research may move towards a more collaborative "platform" model, as seen with companies like Airbnb, with the goal of improving data access, reuse and reproducibility of research. However, overcoming challenges like incentives, trust and work practices will be important for any new platform to succeed.
Whitepaper developed with Pharma Exec magazine on how EIM- Enterprise Information Management- can provide efficiency and kick start innovation by ensuring information flows correctly inside- and outside- the company
With the recent announcement that GlaxoSmithKline have released a huge tranche of whole cell malaria screening data to the public domain, accompanied by a corresponding publication, this raises some issues for consideration before this exemplar instance becomes a trend. We have examined the data from a high level, by studying the molecular properties, and consider the various alerts presently in use by major pharma companies. We acknowledge the potential value of such data but also raise the issue of the actual value of such datasets released into the public domain. We also suggest approaches that could enhance the value of such datasets to the community and theoretically offer more immediate benefit to the search for leads for other neglected diseases.
The Path to Open Science with Illustrations from Computational Biology - A presentation made at the Microsoft 2011 Latin America Faculty Summit Cartagena, Columbia, May 18, 2011.
Quorum Review's April 2014 Institution Bulletin includes a letter from CEO, Cami Gearhart, JD, discussing the need to reform the regulations around biobanking and repositories, HIPAA templates and procedures, and the lessons IRBs can learn from 2013 FDA Warning Letters.
The document summarizes a webinar discussing health and medical data sharing. It covered the following key points:
1) Representatives from ANDS, NHMRC, and Wiley discussed policies around sharing health and medical research data. NHMRC encourages data sharing to maximize benefits from limited resources and ensure ethical use of data.
2) NHMRC's guidelines in its National Statement aim to minimize risks of unauthorized data use while maximizing ethical data sharing. Guidelines prompt consideration of risks but don't mandate specific mechanisms.
3) Wiley is committed to open science through open access publishing, improving data reproducibility, and using open standards to increase research discoverability.
In the world of pharma precompetitive efforts are increasing. These developments have created a dynamic ecosystem with pharma as smaller nodes in a complex network, in which collaborations have become an important business model.
Disruptive Strategies for Removing Drug Discovery Bottlenecks Sean Ekins
This document discusses strategies for removing bottlenecks in drug discovery. It notes that high-throughput screening has shifted from large pharmaceutical companies to academic institutions, but these institutions often lack the capabilities for later-stage drug development. Increased collaboration between academia, CROs, and industry could help address this "valley of death". The document advocates for improved data sharing through public-private partnerships and databases to facilitate collaboration and reduce duplication of efforts. It also suggests that "swarm intelligence" approaches using social media could help identify the best collaborators for research.
This document discusses ways to help advance research for rare and orphan diseases through open collaboration and data sharing. It provides examples of how collaborative platforms and tools can help organizations with limited resources pool data and expertise. Specifically, it describes a secure web-based platform that allows private and public data to be simultaneously searched to facilitate partnerships across various sectors.
Data Science in Biomedicine - Where Are We Headed?Philip Bourne
The document discusses the future of data science in biomedicine. It notes that we are currently at a point of deception, where digitization is occurring but disruption has not yet fully happened. It outlines implications such as open collaborative science becoming more important, and data/analytics increasing in scholarly value. Initiatives like the Precision Medicine Initiative and Big Data to Knowledge are aiming to improve research efficiency and enable precision medicine through large datasets and new methodologies. The future will require cooperation across funders and changes to training to address new skills needed.
Quorum Review's July 2013 Institution Bulletin includes Letter from CEO, Cami Gearhart, JD, addressing support for effective human research protection programs, as well as Quorum's insights on two important topics: The first addresses exculpatory language in consent forms; the second provides insight on considerations when planning eConsent implementation and questions to ask the IRB.
Similar to Exchanging Biology Information with CROs (20)
Fairification experience clarifying the semantics of data matricesPistoia Alliance
This webinar presents the Statistics Ontology, STATO which is a semantic framework to support the creation of standardized analysis reports to help with review of results in the form of data matrices. STATO includes a hierarchy of classes and a vocabulary for annotating statistical methods used in life, natural and biomedical sciences investigations, text mining and statistical analyses.
This webinar discusses driving adoption of microphysiological systems (MPS) in drug R&D. The webinar agenda includes presentations on multi-organ chips for safety and efficacy assessment from TissUse, current applications and future perspectives of organ-on-chips in pharmaceutical industry from AstraZeneca, and driving adoption of MPS from ToxRox Consulting. A panel discussion will be moderated by Mary Ellen Cosenza. The presentations will cover benefits of MPS for reducing drug failures and animal testing, applications across drug discovery and development, challenges for adoption, and perspectives from industry.
Federated Learning (FL) is a learning paradigm that enables collaborative learning without centralizing datasets. In this webinar, NVIDIA present the concept of FL and discuss how it can help overcome some of the barriers seen in the development of AI-based solutions for pharma, genomics and healthcare. Following the presentation, the panel debate on other elements that could drive the adoption of digital approaches more widely and help answer currently intractable science and business questions.
It seems that AI is also becoming a buzzword, like design thinking. Everyone is talking about AI or wants to have AI, and sees all the ideas and benefits – that’s fine, but how do you get started? But what’s different now? Three innovations have finally put AI on the fast track: Big Data, with the internet and sensors everywhere; massive computing power, especially through the Cloud; and the development of breakthrough algorithms, so computers can be trained to accomplish more sophisticated tasks on their own with deep learning. If you use new technology, you need to explore and know what’s possible. With design thinking, it aids to outline the steps and define the ways in which you’re going to create the solution. Starting with mapping the customer journey, defining who will be using that service enhanced with intelligent technology, or who will benefit and gain value from it. We discuss how these two worlds are coming together, and how you get started to transform your venture with Artificial Intelligence using Design Thinking.
Speaker: Claudio Mirti, Principal Solution Specialist – Data & AI, Microsoft
Themes and objectives:
To position FAIR as a key enabler to automate and accelerate R&D process workflows
FAIR Implementation within the context of a use case
Grounded in precise outcomes (e.g. faster and bigger science / more reuse of data to enhance value / increased ability to share data for collaboration and partnership)
To make data actionable through FAIR interoperability
Speakers:
Mathew Woodwark,Head of Data Infrastructure and Tools, Data Science & AI, AstraZeneca
Erik Schultes, International Science Coordinator, GO-FAIR
Georges Heiter, Founder & CEO, Databiology
Knowledge graphs ilaria maresi the hyve 23apr2020Pistoia Alliance
Data for drug discovery and healthcare is often trapped in silos which hampers effective interpretation and reuse. To remedy this, such data needs to be linked both internally and to external sources to make a FAIR data landscape which can power semantic models and knowledge graphs.
2020.04.07 automated molecular design and the bradshaw platform webinarPistoia Alliance
This presentation described how data-driven chemoinformatics methods may automate much of what has historically been done by a medicinal chemist. It explored what is reasonable to expect “AI” approaches might achieve, and what is best left with a human expert. The implications of automation for the human-machine interface were explored and illustrated with examples from Bradshaw, GSK’s experimental automated design environment.
This presentation reviewed the challenges in identifying, acquiring and utilizing research data in relation to an evolving data market. Strategic solutions were examined in which the FAIR principles play a key role in the future of data management.
Dr. Dennis Wang discusses possible ways to enable ML methods to be more powerful for discovery and to reduce ambiguity within translational medicine, allowing data-informed decision-making to deliver the next generation of diagnostics and therapeutics to patients quicker, at lowered costs, and at scale.
The talk by Dr. Dennis Wang was followed by a panel discussion with Mr. Albert Wang, M. Eng., Head, IT Business Partner, Translational Research & Technologies, Bristol-Myers Squibb.
With the explosion of interest in both enhanced knowledge management and open science, the past few years have seen considerable discussion about making scientific data “FAIR” — findable, accessible, interoperable, and reusable. The problem is that most scientific datasets are not FAIR. When left to their own devices, scientists do an absolutely terrible job creating the metadata that describe the experimental datasets that make their way in online repositories. The lack of standardization makes it extremely difficult for other investigators to locate relevant datasets, to re-analyse them, and to integrate those datasets with other data. The Center for Expanded Data Annotation and Retrieval (CEDAR) has the goal of enhancing the authoring of experimental metadata to make online datasets more useful to the scientific community. The CEDAR work bench for metadata management will be presented in this webinar. CEDAR illustrates the importance of semantic technology to driving open science. It also demonstrates a means for simplifying access to scientific data sets and enhancing the reuse of the data to drive new discoveries.
Open interoperability standards, tools and services at EMBL-EBIPistoia Alliance
In this webinar Dr Henriette Harmse from EMBL-EBI presents how they are using their ontology services at EMBL-EBI to scale up the annotation of data and deliver added value through ontologies and semantics to their users.
Fair webinar, Ted slater: progress towards commercial fair data products and ...Pistoia Alliance
Elsevier is a global information analytics business that helps institutions and professional’s
advance healthcare and open science to improve performance for the benefit of humanity.
In this webinar, we discuss how Elsevier is increasingly leveraging the FAIR Guiding Principles to improve its products and services to better serve the scientific community.
Application of recently developed FAIR metrics to the ELIXIR Core Data ResourcesPistoia Alliance
The FAIR (Findable, Accessible, Interoperable and Reusable) principles aim to maximize the discovery and reuse of digital resources. Using recently developed software and metrics to assess FAIRness and supported through an ELIXIR Implementation Study, Michel worked with a subset of ELIXIR Core Data Resources to apply these technologies. In this webinar, he will discuss their approach, findings, and lessons learned towards the understanding and promotion of the FAIR principles.
Implementing Blockchain applications in healthcarePistoia Alliance
Blockchain technology can revolutionise the way information is exchanged between parties by bringing an unprecedented level of security and trust to these transactions. The technology is finding its way into multiple use cases but we are yet to see full adoption and real-world business implementation in the Healthcare industry.
In this webinar we will explore the main challenges and considerations for the implementation of Blockchain technology in Healthcare use cases. This is the third webinar in our Blockchain Education series.
Building trust and accountability - the role User Experience design can play ...Pistoia Alliance
In this webinar our panel of UX specialists give a brief introduction to User Experience before presenting the design opportunities UX can bring to AI. We all know that AI has great potential but has some significant hurdles to overcome not least so the human aspect of trust and ethical considerations when designing in the life sciences.
This document summarizes a webinar on using machine learning and data mining techniques to predict drug repurposing opportunities for chronic pancreatitis. Specifically:
1. Ensemble learning techniques like kernel-based models were used to analyze drug and disease target interaction data from multiple sources to identify potential drug candidates for repurposing.
2. The top 5 repurposing candidates identified through this process were being evaluated further by the partner organization Mission-Cure with the goal of beginning patient trials by January 2020.
3. Additional techniques discussed included using compressed sensing to analyze drug-disease networks and predict side effects to help evaluate candidate drugs identified for repurposing opportunities.
PA webinar on benefits & costs of FAIR implementation in life sciences Pistoia Alliance
The slides from the Pistoia Alliance Debates Webinar where a panel of experts from technology support providers and the biopharma industry, who have been invited to share their views on the "Benefits and costs of FAIR Implementation for life science industry".
Creating novel drugs is an extraordinarily hard and complex problem.
One of the many challenges in drug design is the sheer size of the search space for novel chemical compounds. Scientists need to find molecules that are active toward a biological target or pathway and at the same time have acceptable ADMET properties.
There is now considerable research going on using various AI and ML approaches to tackle these challenges.
Our distinguished speakers, Drs. Alex Tropsha and Ola Engkvist, will discuss their recent work in Drug Design involving Deep Reinforcement Learning and Neural Networks, and will answer questions from the audience on the current state of the research in the field.
Speakers:
Prof Alex Tropsha, Professor at University of North Carolina at Chapel Hill, USA
Dr. Ola Engkvist, Associate Director at AstraZeneca R&D, Gothenburg, Sweden
Alexander Tropsha presented on using AI and machine learning for drug design and discovery. He discussed using QSAR models to predict properties and activity of molecules based on their structural descriptors. He also introduced ReLeaSE, a new method using deep reinforcement learning to generate novel drug-like molecules and guide chemical library design through a thought cycle of molecule generation, model building, and iterative improvement. If successful, this approach could disrupt traditional computational drug discovery pipelines.
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
Session 1 - Intro to Robotic Process Automation.pdfUiPathCommunity
👉 Check out our full 'Africa Series - Automation Student Developers (EN)' page to register for the full program:
https://bit.ly/Automation_Student_Kickstart
In this session, we shall introduce you to the world of automation, the UiPath Platform, and guide you on how to install and setup UiPath Studio on your Windows PC.
📕 Detailed agenda:
What is RPA? Benefits of RPA?
RPA Applications
The UiPath End-to-End Automation Platform
UiPath Studio CE Installation and Setup
💻 Extra training through UiPath Academy:
Introduction to Automation
UiPath Business Automation Platform
Explore automation development with UiPath Studio
👉 Register here for our upcoming Session 2 on June 20: Introduction to UiPath Studio Fundamentals: https://community.uipath.com/events/details/uipath-lagos-presents-session-2-introduction-to-uipath-studio-fundamentals/
"Scaling RAG Applications to serve millions of users", Kevin GoedeckeFwdays
How we managed to grow and scale a RAG application from zero to thousands of users in 7 months. Lessons from technical challenges around managing high load for LLMs, RAGs and Vector databases.
zkStudyClub - LatticeFold: A Lattice-based Folding Scheme and its Application...Alex Pruden
Folding is a recent technique for building efficient recursive SNARKs. Several elegant folding protocols have been proposed, such as Nova, Supernova, Hypernova, Protostar, and others. However, all of them rely on an additively homomorphic commitment scheme based on discrete log, and are therefore not post-quantum secure. In this work we present LatticeFold, the first lattice-based folding protocol based on the Module SIS problem. This folding protocol naturally leads to an efficient recursive lattice-based SNARK and an efficient PCD scheme. LatticeFold supports folding low-degree relations, such as R1CS, as well as high-degree relations, such as CCS. The key challenge is to construct a secure folding protocol that works with the Ajtai commitment scheme. The difficulty, is ensuring that extracted witnesses are low norm through many rounds of folding. We present a novel technique using the sumcheck protocol to ensure that extracted witnesses are always low norm no matter how many rounds of folding are used. Our evaluation of the final proof system suggests that it is as performant as Hypernova, while providing post-quantum security.
Paper Link: https://eprint.iacr.org/2024/257
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Fueling AI with Great Data with Airbyte WebinarZilliz
This talk will focus on how to collect data from a variety of sources, leveraging this data for RAG and other GenAI use cases, and finally charting your course to productionalization.
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Northern Engraving | Modern Metal Trim, Nameplates and Appliance PanelsNorthern Engraving
What began over 115 years ago as a supplier of precision gauges to the automotive industry has evolved into being an industry leader in the manufacture of product branding, automotive cockpit trim and decorative appliance trim. Value-added services include in-house Design, Engineering, Program Management, Test Lab and Tool Shops.
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
This talk will cover ScyllaDB Architecture from the cluster-level view and zoom in on data distribution and internal node architecture. In the process, we will learn the secret sauce used to get ScyllaDB's high availability and superior performance. We will also touch on the upcoming changes to ScyllaDB architecture, moving to strongly consistent metadata and tablets.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
Freshworks Rethinks NoSQL for Rapid Scaling & Cost-EfficiencyScyllaDB
Freshworks creates AI-boosted business software that helps employees work more efficiently and effectively. Managing data across multiple RDBMS and NoSQL databases was already a challenge at their current scale. To prepare for 10X growth, they knew it was time to rethink their database strategy. Learn how they architected a solution that would simplify scaling while keeping costs under control.