Classical forms of publishing centered around printed narrative articles do not seem well-suited for publishing scientific datasets, which have become increasingly important for science. In this talk, I will introduce the concept of nanopublications, an approach to enable provenance-aware data publishing. I will present our work to make nanopublications more general by allowing for informal assertions and meta-nanopublications. By relying on Semantic Web technologies, they allow us to organize and structure scientific knowledge and discourse, and thereby to improve the efficiency of scientific processes. Using cryptographic hash values in their identifiers, we can make nanopublications trustworthy and reliable even in an open and decentralized environment. I will present our proposal of Trusty URIs and a preliminary implementation of a decentralized server network to publish, archive, and retrieve nanopublications in a reliable and efficient manner.
Semantic Publishing and NanopublicationsTobias Kuhn
Making available and archiving scientific results is for the most part still considered the task of classical publishing companies, despite the fact that classical forms of publishing centered around printed narrative articles no longer seem well-suited in the digital age. In particular, there exist currently no efficient, reliable, and agreed-upon methods for publishing scientific datasets, which have become increasingly important for science. In this talk, I outline how we can design scientific data publishing as a Web-based bottom-up process, without top-down control of central authorities. I present a protocol and a server network to decentrally store and archive data in the form of nanopublications, a format based on Semantic Web techniques to represent scientific data with formal semantics. Such nanopublications can be made verifiable and immutable by applying cryptographic methods with identifiers called Trusty URIs. I show how this approach allows researchers to produce, publish, retrieve, address, verify, and recombine datasets and their individual nanopublications in a reliable and trustworthy manner, and I discuss how the current small network can grow to handle the large amounts of structured data that modern science is producing and consuming.
Data Publishing and Post-Publication ReviewsTobias Kuhn
This presentation is about the combination of data publishing and post-publication reviews, and it covers some recent work on nanopublications for data publishing and reputation mechanisms for Web-scale quality metrics of scientific contributions.
Introductory seminar @ International Hellenic University (8 May 2014)
Main topics covered:
- data collection from social sources
- indexing using mongoDB and Solr
- mining (basic analytics & topic detection)
Replication and Synchronization Algorithms for Distributed Databases - Lena W...distributed matters
This talk will provide in-depth background on strategies for replication and synchronization as implemented in modern distributed databases. The following topics will be covered: master-slave vs multi-master replication; epidemic protocols; two-phase commit vs Paxos; multiversion concurrency control; read and write quorums. A concise overview of implementations in current NoSQL databases will be presented.
" NoSQL Databases: An Overview" Lena Wiese, Research Group Knowledge Engineer...Dataconomy Media
"NoSQL Databases: An Overview", Lena Wiese, Research Group Knowledge Engineer at Göttingen University
YouTube Link: https://www.youtube.com/watch?v=RDXgBzAaF4g
Watch more from Data Natives 2015 here: http://bit.ly/1OVkK2J
Visit the conference website to learn more: www.datanatives.io
Follow Data Natives:
https://www.facebook.com/DataNatives
https://twitter.com/DataNativesConf
Stay Connected to Data Natives by Email: Subscribe to our newsletter to get the news first about Data Natives 2016: http://bit.ly/1WMJAqS
About the author:
Dr. Lena Wiese is head of the research group Knowledge Engineering and lecturer at the Georg August University Göttingen. She has been teaching advanced courses on data management and database technology for several years at both graduate and undergraduate level. She is author of the book "Advanced Data Management for SQL, NoSQL, Cloud and Distributed Databases" (DeGruyter, 2015)
Metrics that matter: Making the business case that documentation has valuePublishing Smarter
Presented at CMS/DITA North America 2016 to help people tell the story around content as a business asset. We agree there is value in documentation but have been challenged at times to “prove it”. Demo of how to present to groups including sales, support, service, IT, engineering, QA/testing, manufacturing, HR, training, finance, marketing, and every other business unit in your organization. Discussion on how documentation drives sales and generates corporate revenue to managers and executives helping them see how important documentation is to them.
Semantic Publishing and NanopublicationsTobias Kuhn
Making available and archiving scientific results is for the most part still considered the task of classical publishing companies, despite the fact that classical forms of publishing centered around printed narrative articles no longer seem well-suited in the digital age. In particular, there exist currently no efficient, reliable, and agreed-upon methods for publishing scientific datasets, which have become increasingly important for science. In this talk, I outline how we can design scientific data publishing as a Web-based bottom-up process, without top-down control of central authorities. I present a protocol and a server network to decentrally store and archive data in the form of nanopublications, a format based on Semantic Web techniques to represent scientific data with formal semantics. Such nanopublications can be made verifiable and immutable by applying cryptographic methods with identifiers called Trusty URIs. I show how this approach allows researchers to produce, publish, retrieve, address, verify, and recombine datasets and their individual nanopublications in a reliable and trustworthy manner, and I discuss how the current small network can grow to handle the large amounts of structured data that modern science is producing and consuming.
Data Publishing and Post-Publication ReviewsTobias Kuhn
This presentation is about the combination of data publishing and post-publication reviews, and it covers some recent work on nanopublications for data publishing and reputation mechanisms for Web-scale quality metrics of scientific contributions.
Introductory seminar @ International Hellenic University (8 May 2014)
Main topics covered:
- data collection from social sources
- indexing using mongoDB and Solr
- mining (basic analytics & topic detection)
Replication and Synchronization Algorithms for Distributed Databases - Lena W...distributed matters
This talk will provide in-depth background on strategies for replication and synchronization as implemented in modern distributed databases. The following topics will be covered: master-slave vs multi-master replication; epidemic protocols; two-phase commit vs Paxos; multiversion concurrency control; read and write quorums. A concise overview of implementations in current NoSQL databases will be presented.
" NoSQL Databases: An Overview" Lena Wiese, Research Group Knowledge Engineer...Dataconomy Media
"NoSQL Databases: An Overview", Lena Wiese, Research Group Knowledge Engineer at Göttingen University
YouTube Link: https://www.youtube.com/watch?v=RDXgBzAaF4g
Watch more from Data Natives 2015 here: http://bit.ly/1OVkK2J
Visit the conference website to learn more: www.datanatives.io
Follow Data Natives:
https://www.facebook.com/DataNatives
https://twitter.com/DataNativesConf
Stay Connected to Data Natives by Email: Subscribe to our newsletter to get the news first about Data Natives 2016: http://bit.ly/1WMJAqS
About the author:
Dr. Lena Wiese is head of the research group Knowledge Engineering and lecturer at the Georg August University Göttingen. She has been teaching advanced courses on data management and database technology for several years at both graduate and undergraduate level. She is author of the book "Advanced Data Management for SQL, NoSQL, Cloud and Distributed Databases" (DeGruyter, 2015)
Metrics that matter: Making the business case that documentation has valuePublishing Smarter
Presented at CMS/DITA North America 2016 to help people tell the story around content as a business asset. We agree there is value in documentation but have been challenged at times to “prove it”. Demo of how to present to groups including sales, support, service, IT, engineering, QA/testing, manufacturing, HR, training, finance, marketing, and every other business unit in your organization. Discussion on how documentation drives sales and generates corporate revenue to managers and executives helping them see how important documentation is to them.
Distributed Digital Artifacts on the Semantic WebEditor IJCATR
Distributed digital artifacts incorporate cryptographic hash values to URI called trusty URIs in a distributed environment
building good in quality, verifiable and unchangeable web resources to prevent the rising man in the middle attack. The greatest
challenge of a centralized system is that it gives users no possibility to check whether data have been modified and the communication
is limited to a single server. As a solution for this, is the distributed digital artifact system, where resources are distributed among
different domains to enable inter-domain communication. Due to the emerging developments in web, attacks have increased rapidly,
among which man in the middle attack (MIMA) is a serious issue, where user security is at its threat. This work tries to prevent MIMA
to an extent, by providing self reference and trusty URIs even when presented in a distributed environment. Any manipulation to the
data is efficiently identified and any further access to that data is blocked by informing user that the uniform location has been
changed. System uses self-reference to contain trusty URI for each resource, lineage algorithm for generating seed and SHA-512 hash
generation algorithm to ensure security. It is implemented on the semantic web, which is an extension to the world wide web, using
RDF (Resource Description Framework) to identify the resource. Hence the framework was developed to overcome existing
challenges by making the digital artifacts on the semantic web distributed to enable communication between different domains across
the network securely and thereby preventing MIMA.
A theory from anthropologist and psychologist Robin Dunbar states that the brain capacity of humans limits the number of stable social relationships they can maintain to 150. But what does that mean for B2B organizations with countless contacts?
Benchmarking the Accounting & Finance Function: 2014 Summary PresentationRobert Half
Is your finance and accounting team ready to drive your success throughout 2014? Robert Half’s fifth annual Benchmarking the Accounting & Finance Function report provides metrics on staffing, financial systems, outsourcing and more. Find out how your company measures up to its peers.
ASI 07 - How Auditing Radio campaigns helps Improve Planning and Buying Effic...Paola Furlanetto
. building and maintaining pools for radio auditing
. planning and buying radio: most common inefficiencies
. fine tuning radio planning and buying: 5 guidelines
National Centre for Student Equity in Higher Education (NCSEHE) Director Professor Sue Trinidad presents, "Student equity: policy and practice" at the ACER-sponsored Strategies for Student Retention conference held in Melbourne on Tuesday 29 and Wednesday 30 September 2015. Professor Trinidad provides an overview of the NCSEHE's work, including the development of student personas in order to better identify cohorts of students requiring additional support, and strategies with which to assist.
WikiPathways: how open source and open data can make omics technology more us...Chris Evelo
Presentation about collaborative development of open source pathway analysis code and pathways and about usage in analytical software distributed with analytical machines like mass spectrophotometers.
A Decentralized Network for Publishing Linked Data — Nanopublications, Trusty...Tobias Kuhn
Making available and archiving scientific results is for the most part still considered the task of classical publishing companies, despite the fact that classical forms of publishing centered around printed narrative articles no longer seem well-suited in the digital age. In particular, there exist currently no efficient, reliable, and agreed-upon methods for publishing scientific datasets, which have become increasingly important for science. In this talk, I outline how we can design scientific data publishing as a Web-based bottom-up process, without top-down control of central authorities. I present a protocol and a server network to decentrally store and archive data in the form of nanopublications, a format based on Semantic Web techniques to represent scientific data with formal semantics. Such nanopublications can be made verifiable and immutable by applying cryptographic methods with identifiers called Trusty URIs. I show how this approach allows researchers to produce, publish, retrieve, address, verify, and recombine datasets and their individual nanopublications in a reliable and trustworthy manner, and I discuss how the current small network can grow to handle the large amounts of structured data that modern science is producing and consuming.
(CC license does not apply to third-party content)
Distributed Digital Artifacts on the Semantic WebEditor IJCATR
Distributed digital artifacts incorporate cryptographic hash values to URI called trusty URIs in a distributed environment
building good in quality, verifiable and unchangeable web resources to prevent the rising man in the middle attack. The greatest
challenge of a centralized system is that it gives users no possibility to check whether data have been modified and the communication
is limited to a single server. As a solution for this, is the distributed digital artifact system, where resources are distributed among
different domains to enable inter-domain communication. Due to the emerging developments in web, attacks have increased rapidly,
among which man in the middle attack (MIMA) is a serious issue, where user security is at its threat. This work tries to prevent MIMA
to an extent, by providing self reference and trusty URIs even when presented in a distributed environment. Any manipulation to the
data is efficiently identified and any further access to that data is blocked by informing user that the uniform location has been
changed. System uses self-reference to contain trusty URI for each resource, lineage algorithm for generating seed and SHA-512 hash
generation algorithm to ensure security. It is implemented on the semantic web, which is an extension to the world wide web, using
RDF (Resource Description Framework) to identify the resource. Hence the framework was developed to overcome existing
challenges by making the digital artifacts on the semantic web distributed to enable communication between different domains across
the network securely and thereby preventing MIMA.
A theory from anthropologist and psychologist Robin Dunbar states that the brain capacity of humans limits the number of stable social relationships they can maintain to 150. But what does that mean for B2B organizations with countless contacts?
Benchmarking the Accounting & Finance Function: 2014 Summary PresentationRobert Half
Is your finance and accounting team ready to drive your success throughout 2014? Robert Half’s fifth annual Benchmarking the Accounting & Finance Function report provides metrics on staffing, financial systems, outsourcing and more. Find out how your company measures up to its peers.
ASI 07 - How Auditing Radio campaigns helps Improve Planning and Buying Effic...Paola Furlanetto
. building and maintaining pools for radio auditing
. planning and buying radio: most common inefficiencies
. fine tuning radio planning and buying: 5 guidelines
National Centre for Student Equity in Higher Education (NCSEHE) Director Professor Sue Trinidad presents, "Student equity: policy and practice" at the ACER-sponsored Strategies for Student Retention conference held in Melbourne on Tuesday 29 and Wednesday 30 September 2015. Professor Trinidad provides an overview of the NCSEHE's work, including the development of student personas in order to better identify cohorts of students requiring additional support, and strategies with which to assist.
WikiPathways: how open source and open data can make omics technology more us...Chris Evelo
Presentation about collaborative development of open source pathway analysis code and pathways and about usage in analytical software distributed with analytical machines like mass spectrophotometers.
A Decentralized Network for Publishing Linked Data — Nanopublications, Trusty...Tobias Kuhn
Making available and archiving scientific results is for the most part still considered the task of classical publishing companies, despite the fact that classical forms of publishing centered around printed narrative articles no longer seem well-suited in the digital age. In particular, there exist currently no efficient, reliable, and agreed-upon methods for publishing scientific datasets, which have become increasingly important for science. In this talk, I outline how we can design scientific data publishing as a Web-based bottom-up process, without top-down control of central authorities. I present a protocol and a server network to decentrally store and archive data in the form of nanopublications, a format based on Semantic Web techniques to represent scientific data with formal semantics. Such nanopublications can be made verifiable and immutable by applying cryptographic methods with identifiers called Trusty URIs. I show how this approach allows researchers to produce, publish, retrieve, address, verify, and recombine datasets and their individual nanopublications in a reliable and trustworthy manner, and I discuss how the current small network can grow to handle the large amounts of structured data that modern science is producing and consuming.
(CC license does not apply to third-party content)
In this lightning talk, I introduce an approach to publish RDF data in a provenance-aware and reliable manner. This approach is based on the concept of nanopublications, which we can give unique and verifiable identifiers using cryptographic hash values. Based on that, I present ongoing work to establish a decentralized server network for publishing, archiving, and retrieveing Linked Data in a reliable and trustworthy way.
Publishing without Publishers: a Decentralized Approach to Dissemination, Ret...Tobias Kuhn
Making available and archiving scientific results is for the most part still considered the task of classical publishing companies, despite the fact that classical forms of publishing centered around printed narrative articles no longer seem well-suited in the digital age. In particular, there exist currently no efficient, reliable, and agreed-upon methods for publishing scientific datasets, which have become increasingly important for science.
Here we propose to design scientific data publishing as a Web-based bottom-up process, without top-down control of central authorities such as publishing companies. Based on a novel combination of existing concepts and technologies, we present a server network to decentrally store and archive data in the form of nanopublications, an RDF-based format to represent scientific data.
We show how this approach allows researchers to publish, retrieve, verify, and recombine datasets of nanopublications in a reliable and trustworthy manner, and we argue that this architecture could be used for the Semantic Web in general. Evaluation of the current small network shows that this system is efficient and reliable.
Linked Data Publishing with NanopublicationsTobias Kuhn
When we think about scientific publishing, we mainly think about papers published in journals or proceedings. This publishing model, however, doesn't seem to work very well for datasets, which have become increasingly important in many areas of science. Instead of treating datasets like papers, couldn't we put the data first by allowing scientists to directly publish data entries that represent their results? Nanopublications are an approach in this direction that builds upon the recent maturation of Linked Data technologies and proposes an entirely new paradigm for the future of scholarly communication.
nanopub-java: A Java Library for NanopublicationsTobias Kuhn
The concept of nanopublications was first proposed about six years ago, but it lacked openly available implementations. The library presented here is the first one that has become an official implementation of the nanopublication community. Its core features are stable, but it also contains unofficial and experimental extensions: for publishing to a decentralized server network, for defining sets of nanopublications with indexes, for informal assertions, and for digitally signing nanopublications. Most of the features of the library can also be accessed via an online validator interface.
Open Science, Open Data: towards a new transparent and reproducible ecosystemLIBER Europe
Presented at the Preforma Open Source Workshop 8 April 2016
As a library membership organization, LIBER works on addressing Open Science barriers. Standardisation of file formats can really help in overcoming some of these barriers: it enables us to process and preserve data in a controlled way, it helps ensure that outputs are really open and accessible in the long term and it improves interoperability of new tools and services. Making sure data is stored in a controlled way and can be (re) used today and in the future is an important element in Open Science. We see this as not only a technical challenge but also a social one: awareness, trust and community building is needed in order to ensure uptake of these standards. Libraries therefore have a valuable role to play in the development of good research data management throughout all phases of the Open Data lifecycle.
A open science presentation focusing on the benefits to be gained and basic practices to follow. This was given on behalf of FOSTER at the Open Science Boos(t)camp event at KU Leuven on 24th October 2014.
HathiTrust Research Center Secure CommonsBeth Plale
Introduces HTRC secure commons, expanded secure infrastructure and services for text mining of HT digital data. Shows results comparing n-gram discovery using Solr full text index and a framework using mapReduce. Compute time over 1 million digital volumes is 1 day with 1024 cores. Weaknesses of Solr in n-gram identification are explored.
Many scholars have pointed out that the classical way of publishing scientific articles is ill-suited to deal with the rapid growth of both, volume and complexity, of scientific contributions. To overcome these problems, next generation scientific publishing has to respond to the increasing importance of datasets and software, and needs to provide methods to automatically organize and aggregate reported scientific findings. Perhaps the most important shortcoming of the current publication system is that scientific papers do not come with formal semantics that could be processed, aggregated, and interpreted in an automated fashion.
Semantic publishing is a general approach to tackle this problem using the concepts and tools of the Semantic Web and related fields.
OpenAIRE: Open Science as-a-Service - presentation at #DI4R2016OpenAIRE
Presentation at the Digital Infrastructures for Research 2016 Conference (Sept 30). Title: OpenAIRE: Open Science as-a-Service, by Paolo Manghi, CNR-ISTI
The Controlled Natural Language of Randall Munroe’s Thing Explainer Tobias Kuhn
It is rare that texts or entire books written in a Controlled Natural Language (CNL) become very popular, but exactly this has happened with a book that has been published last year. Randall Munroe's Thing Explainer uses only the 1'000 most often used words of the English language together with drawn pictures to explain complicated things such as nuclear reactors, jet engines, the solar system, and dishwashers. This restricted language is a very interesting new case for the CNL community. I describe here its place in the context of existing approaches on Controlled Natural Languages, and I provide a first analysis from a scientific perspective, covering the word production rules and word distributions.
A Multilingual Semantic Wiki Based on Controlled Natural LanguageTobias Kuhn
This presentation introduces AceWiki-GF, a semantic wiki based on controlled natural language that makes its knowledge base viewable and editable in different languages applying high-quality rule-based machine translation.
How to Evaluate Controlled Natural LanguagesTobias Kuhn
(CC Attribution License does not apply to third-party material on slides 5 and 6; see paper for details: http://attempto.ifi.uzh.ch/site/pubs/papers/cnl2009main_kuhn.pdf )
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...University of Maribor
Slides from talk:
Aleš Zamuda: Remote Sensing and Computational, Evolutionary, Supercomputing, and Intelligent Systems.
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Inter-Society Networking Panel GRSS/MTT-S/CIS Panel Session: Promoting Connection and Cooperation
https://www.etran.rs/2024/en/home-english/
This presentation explores a brief idea about the structural and functional attributes of nucleotides, the structure and function of genetic materials along with the impact of UV rays and pH upon them.
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Salas, V. (2024) "John of St. Thomas (Poinsot) on the Science of Sacred Theol...Studia Poinsotiana
I Introduction
II Subalternation and Theology
III Theology and Dogmatic Declarations
IV The Mixed Principles of Theology
V Virtual Revelation: The Unity of Theology
VI Theology as a Natural Science
VII Theology’s Certitude
VIII Conclusion
Notes
Bibliography
All the contents are fully attributable to the author, Doctor Victor Salas. Should you wish to get this text republished, get in touch with the author or the editorial committee of the Studia Poinsotiana. Insofar as possible, we will be happy to broker your contact.
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Semantic Publishing with Nanopublications
1. Semantic Publishing with Nanopublications
Tobias Kuhn
http://www.tkuhn.ch
@txkuhn
ETH Zurich
Research Colloquium
Stanford Center for Biomedical Informatics Research
12 March 2015
2. Information Overload:
>1M new Scientific Articles Per Year
• How can we help
scientists to stay
up-to-date?
• How can we efficiently
retrieve and aggregate
published scientific
results?
• How should we design
the system for the
future of scientific
communication?
Image from: Kuhn et al. Inheritance patterns in citation networks reveal scientific memes. Physical Review X 4. 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 2 / 30
6. Vision: Changing Scholarly Communication
Now
Narrative articles at the center
Future
Nanopublications at the center
Images from Mons et al. The value of data. Nature genetics, 43(4):281–283, 2011
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 6 / 30
7. Open Problems
Three important open problems for nanopublications are:
• Identifiability: How can we ensure the immutability of
nanopublications and provide verifiable identifiers?
• Persistence: How can we ensure that published
nanopublications can be reliably retrieved by others and remain
permanently available?
• Representation: How can we deal with results that have no
straightforward RDF representation?
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 7 / 30
8. Problem:
Replication and Re-Use of Research Results
Exemplary Situation: Sue publishes a script that should allow
everybody to replicate her scientific analysis:
# Download data:
wget http://some-third-party.org/dataset/1.4
# Analyze data:
...
Problems: What if the third party silently changes that version of the
dataset? What if the resource becomes unavailable at this location?
What if the web site later gets hacked and the data manipulated?
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 8 / 30
9. Identifiability Problem
http://some-third-party.org/dataset/1.4
?
Given a URI for a digital artifact, there is no reliable standard
procedure of checking whether a retrieved file really represents the
correct and original state of that artifact.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 9 / 30
10. Trusty URIs
Basic idea: Use of cryptographic hash values calculated on digital
artifacts.
Requirements:
• To allow for the verification of entire reference trees, the hash
should be part of the reference (i.e. the URI)
• To allow for meta-data, digital artifacts should be allowed to
contain self-references (i.e. their own URI)
• Format-independent hash for different kinds of content
• The complete approach should be decentralized and open
• We want to use them right away
Example:
http://example.org/r1.RA5AbXdpz5DcaYXCh9l3eI9ruBosiL5XDU3rxBbBaUO70
Kuhn, Dumontier. Trusty URIs: Verifiable, Immutable, and Permanent Digital Artifacts for Linked Data. ESWC 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 10 / 30
11. Trusty URIs: Range of Verifiability
With the hash as a part of the URI, the “range of verifiability”
extends to referenced artifacts (if they also use trusty URIs):
http://...RAcbjcRI...
http://...RAQozo2w...
http://...RABMq4Wc...
http://...RAcbjcRI...
http://...RAQozo2w...
http://.../resource23
http://.../resource23
...
http://...RAUx3Pqu...
http://.../resource55
http://...RABMq4Wc...
http://.../resource55
http://...RARz0AX-...
...
http://...RAUx3Pqu...
...
http://...RARz0AX...
...
range of
verifiability
Kuhn, Dumontier. Trusty URIs: Verifiable, Immutable, and Permanent Digital Artifacts for Linked Data. ESWC 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 11 / 30
12. Verifiable — Immutable — Permanent
Whether or not a given resource is the one a given trusty URI is
supposed to represent can be verified with perfect confidence.
(assuming that the trusty URI for the required artifact is known, e.g. because
another artifact contains it as a link)
Kuhn, Dumontier. Making Digital Artifacts on the Web Verifiable and Reliable. IEEE TKDE. To appear. / Kuhn,
Dumontier. Trusty URIs: Verifiable, Immutable, and Permanent Digital Artifacts for Linked Data. ESWC 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 12 / 30
13. Verifiable — Immutable — Permanent
A trusty URI artifact is immutable, as any change in the content
also changes its URI, thereby making it a new artifact.
(as soon as your trusty URI has been picked up by third parties, e.g. cached or
linked from other resources, every change will be noticed)
Kuhn, Dumontier. Trusty URIs: Verifiable, Immutable, and Permanent Digital Artifacts for Linked Data. ESWC 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 13 / 30
14. Verifiable — Immutable — Permanent
Trusty URI artifacts are permanent, as they can be retrieved from
the cache of third-party websites if otherwise no longer available.
(if there are search engines and web archives regularly crawling and caching the
artifacts on the web)
Kuhn, Dumontier. Trusty URIs: Verifiable, Immutable, and Permanent Digital Artifacts for Linked Data. ESWC 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 14 / 30
15. Calculating Trusty URIs is Fast and Reliable
Kuhn, Dumontier. Trusty URIs: Verifiable, Immutable, and Permanent Digital Artifacts for Linked Data. ESWC 2014.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 15 / 30
16. Nanopublication Server Network
Nanopublications
with Trusty URIs
Publication
Retrieval
Propagation /
Archiving
http://npmonitor.inn.ac
Kuhn et al. Publishing without Publishers: a Decentralized Approach to Dissemination, Retrieval, and Archiving of
Data. arXiv:1411.2749.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 16 / 30
17. Decentralized — Open — Real-time
• No a central authority: Everybody can set up a server and join
the network
• No restrictions on publication: Everybody can upload
nanopublications
• No delay between submission and publication: Nanopublications
are made public immediately
• No updates: If a nanopublication is modified, that makes it a
new nanopublication (enforced by trusty URIs)
• No queries: Only simple identifier-based lookup
Kuhn et al. Publishing without Publishers: a Decentralized Approach to Dissemination, Retrieval, and Archiving of
Data. arXiv:1411.2749.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 17 / 30
18. High Performance and High Availability
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 18 / 30
19. Defining Datasets with Nanopublication Indexes
(which are themselves Nanopublications)
appends
has sub-index
has
element
(a) (b)
(c) (f)
(d) (e)
Kuhn et al. Publishing without Publishers: a Decentralized Approach to Dissemination, Retrieval, and Archiving of
Data. arXiv:1411.2749.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 19 / 30
20. Using Nanopublication Datasets
Once published in the network, nanopublication indexes can be cited:
[7] Nanopubs converted from OpenBEL’s Small and Large Corpus 20131211.
Nanopublication index, 4 March 2014,
http://np.inn.ac/RAR5dwELYLKGSfrOclnWhjOj-2nGZN_8BW1JjxwFZINHw
Researchers can then fetch and reuse the data in a reliable and
prefectly reproducible manner:
# Download data:
GetNanopub.sh -c RAR5dwELYLKGSfrOclnWhjOj-2nGZN 8BW1JjxwFZINHw
# Analyze data:
...
Existing data can be recombined in new indexes; and researchers can
unambiguously refer to the used datasets for new results:
this: prov:wasDerivedFrom nps:RAR5dwELYLKGSfrOclnWhjOj-2nGZN 8BW1Jjx
Kuhn et al. Publishing without Publishers: a Decentralized Approach to Dissemination, Retrieval, and Archiving of
Data. arXiv:1411.2749.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 20 / 30
21. Real-Time and Replicable Data Mining
authors
users
curators
structured
data sources
unstructured
data sources
ocean of
nanopublications
nanopublication
portals
in silico
experiments
network
analyses
aggregated
views
hypothesis
generation
...bots
1
2
3
4
5
6
Kuhn et al. Broadening the Scope of Nanopublications. ESWC 2013.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 21 / 30
22. How can we Represent Actual Scientific Results
in RDF?
Example:
[...] the risk of developing neurodegenerative disease in idiopathic
REM sleep behavior disorder is substantial, with the majority of
patients developing Parkinson disease and Lewy body dementia.
[PubMed 19109537]
This is difficult to represent in RDF, but it would correspond to two
nanopublication assertions:
• “The risk of developing neurodegenerative disease in idiopathic REM
sleep behavior disorder is substantial.”
• “The majority of patients with idiopathic REM sleep behavior disorder
who develop a neurodegenerative disease develop Parkinson disease
and Lewy body dementia.”
Nanopublication provenance and metadata, on the other hand, would
be relatively easy to produce.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 22 / 30
23. English Sentences as Anchors to Structure
Scientific Discourse
Malaria is transmitted by mosquitoes.
Mosquitoes transmit malaria.
Malaria is transmitted by female mosquitoes.
study A
same meaning
more specific
meaning of
provides
evidence for provides counter-
evidence against
study B
study C
study D
Malaria is transmitted by moscitos.
corrected version of
Kuhn et al. Broadening the Scope of Nanopublications. ESWC 2013.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 23 / 30
24. AIDA Sentences: Criteria
Single English sentences that are:
• Atomic: a sentence describing one thought that cannot be
further broken down in a practical way
• Independent: a sentence that can stand on its own, without
external references like “this effect” or “we”
• Declarative: a complete sentence ending with a full stop that
could in theory be either true or false
• Absolute: a sentence describing the core of a claim ignoring the
uncertainty about its truth and how it was discovered (no
“probably” or “evaluation showed”); typically in present tense
Kuhn et al. Broadening the Scope of Nanopublications. ESWC 2013.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 24 / 30
25. Controlled Natural Language (CNL)
AIDA is a kind of Controlled Natural Language.
More about that in my talk next week (20 March 2014)
in the Prot´eg´e Seminar.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 25 / 30
26. AIDA Nanopublications
Nanopublications can be combined with AIDA sentences to allow for
informal and underspecified assertions:
Malaria is transmitted by mosquitoes.
Malaria is transmitted by mosquitoes.
ns1:mosquito
ns3:transmission ns1:mosquito
Malaria is transmitted by mosquitoes.
ns2:malaria
ns3:transmission
They can be linked regardless of their degree of formality:
ns1:mosquito
Malaria is transmitted by mosquitoes.
ns2:malaria
ns3:transmission
Kuhn et al. Broadening the Scope of Nanopublications. ESWC 2013.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 26 / 30
27. AIDA Sentences can be Efficiently Created
Manual creation of AIDA sentences from existing abstracts:
163total 100%
114perfect 70%
7typo etc. 4%
10inaccurate 6%
32not AIDA 20%
0 20 40 60 80 100 120 140 160
sentences
Automatic extraction of AIDA sentences from existing GeneRIF dataset:
250total 100%
177perfect 71%
8typo etc. 3%
65not AIDA 26%
0 25 50 75 100 125 150 175 200 225 250
sentences
Kuhn et al. Broadening the Scope of Nanopublications. ESWC 2013.
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 27 / 30
28. Examples: AIDA Sentences for Claims by
Famous Computer Scientists
• Computers are more flexible and more powerful if data and program
instructions are both stored side by side in the same memory. [John von
Neumann]
• The problem of determining the provability of arbitrary propositions in
first-order logic is undecidable. [Alonso Church / Alan Turing]
• There is no general algorithm that can decide for all possible programs and
inputs whether the program will terminate or run forever. [Alan Turing]
• Automated reasoning can be suitably implemented as a search tree with the
hypothesis at the root and by applying heuristics to trim branches. [Herbert
Simon and Allen Newell]
• Single-layer perceptrons cannot learn the XOR relation. [Marvin Minsky and
Seymour Papert]
• The use of GOTO statements in programming is harmful to program
structure. [Edsger Dijkstra]
• Hypertext representations are more flexible, more general, and more natural
than classical text structure for computer-supported information systems.
[Ted Nelson]
https://github.com/tkuhn/aida/
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 28 / 30
29. Summary
• With Trusty URIs, digital artifacts on the Web can be made
immutable and verifiable
• With a decentralized network of nanopublication servers, we
can build a reliable and trustworthy infrastructure for data
publishing
• With AIDA sentences, we can formally represent provenance
and metadata of informal or underspecified assertions
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 29 / 30
30. Thank you for your attention!
Questions?
Tobias Kuhn, ETH Zurich Semantic Publishing with Nanopublications 30 / 30