Accounting for uncertainty in species delineation during the analysis of envi...methodsecolevol
Tutorial accompanying the paper of the same name, published in Methods in Ecology and Evolution
Full paper
http://onlinelibrary.wiley.com/doi/10.1111/j.2041-210X.2011.00122.x/abstract
Friday, October 15th, 2021, Sapporo, Hokkaido, Japan.
Hokkaido University ICReDD - Faculty of Medicine Joint Symposium
https://www.icredd.hokudai.ac.jp/event/5993
ICReDD (Institute for Chemical Reaction Design and Discovery)
https://www.icredd.hokudai.ac.jp
Accounting for uncertainty in species delineation during the analysis of envi...methodsecolevol
Tutorial accompanying the paper of the same name, published in Methods in Ecology and Evolution
Full paper
http://onlinelibrary.wiley.com/doi/10.1111/j.2041-210X.2011.00122.x/abstract
Friday, October 15th, 2021, Sapporo, Hokkaido, Japan.
Hokkaido University ICReDD - Faculty of Medicine Joint Symposium
https://www.icredd.hokudai.ac.jp/event/5993
ICReDD (Institute for Chemical Reaction Design and Discovery)
https://www.icredd.hokudai.ac.jp
Jean-Claude Bradley presents on "The implications of Open Notebook Science and other new forms of scientific communication for Nanoinformatics" at the Nanoinformatics 2010 conference on November 3, 2010. The presentation first covers the use of the laboratory knowledge management system SMIRP for nanotechnology applications during the period of 1999-2001 at Drexel University. The exporting of single experiments from SMIRP and publication to the Chemistry Preprint Archive is then described followed by the evolution to Open Notebook Science in 2005. Abstraction of semantic structure from ONS projects in the areas of drug discovery and solubility is then detailed as an efficient mechanism to provide web services and machine readable data feeds.
Talk by Martin Scharm at the COMBINE meeting September 2013 in Paris.
Find more information and download the slides at http://sems.uni-rostock.de/2013/09/sems-at-the-combine-2013/
Abstract: The processing power of computing devices has increased with number of available cores. This paper presents an approach
towards clustering of categorical data on multi-core platform. K-modes algorithm is used for clustering of categorical data which
uses simple dissimilarity measure for distance computation. The multi-core approach aims to achieve speedup in processing. Open
Multi Processing (OpenMP) is used to achieve parallelism in k-modes algorithm. OpenMP is a shared memory API that uses
thread approach using the fork-join model. The dataset used for experiment is Congressional Voting Dataset collected from UCI
repository. The dataset contains votes of members in categorical format provided in CSV format. The experiment is performed for
increased number of clusters and increasing size of dataset.
Abstract:Reactive Power Optimization is a complex combinatorial optimization problem involving non-linear function having multiple local minima, non-linear and discontinuous constrains. This paper presents PS2O, which extends the dynamics of the canonical PSO algorithm by adding a significant ingredient that takes into account the symbiotic co evolution between species, Hybrid Evolutionary-Conventional Algorithm (HECA) that uses the abilities of evolutionary and conventional algorithm and Genetical Swarm Optimization (GSO), combines Genetic Algorithms (GA) and Particle Swarm Optimization (PSO).All the above said algorithms is used to overcome the Problem of premature convergence. PS2O, HECA , GSO is applied to Reactive Power Optimization problem and is evaluated on standard IEEE 57, practical 191 test Bus Systems. The results shows that all the three algorithms perform well in solving the reactive power problem and prevent premature convergence to high degree but still keep a rapid convergence. Of all the three PS2O has the edge in reducing the real power loss when compared to HECA & GSO.
Keywords:PS2O, Hybrid Evolutionary-Conventional Algorithm, Genetical Swarm Optimization, Reactive Power Optimization.
Particle Swarm Optimization based K-Prototype Clustering Algorithm iosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Using AI Planning to Automate the Performance Analysis of SimulatorsRoland Ewald
Analyzing simulation algorithm performance is cumbersome: execute some runs, observe a performance metric, and analyze the results. Often, the results motivate follow-up experiments, which in turn may lead to additional experiments, and so on. This time-consuming and error-prone process can be automated with planning approaches from artificial intelligence, making simulator performance analysis more convenient and rigorous. This paper introduces ALeSiA, a prototypical system for automatic simulator performance analysis. It is independent of any specific simulation system and realizes a hypothesis-driven approach to evaluate performance.
Novel Class Detection Using RBF SVM Kernel from Feature Evolving Data Streamsirjes
In the data mining field the classification of data stream creates many problems. The challenges
faces in the data stream are infinite length, concept drift, concept evaluation and feature evolution. Most of the
existing system focuses on the only first two challenges. We propose a framework in which each classifier is
prepared with the novel class detector for addressing the two challenges concept drift and concept evaluation
and for addressing the feature evolution feature set homogeneous technique is proposed. We improved the
novel class detection module by building it more adaptive to evolving the stream. SVM based feature extraction
for RBF kernel method is also proposed for detecting the novel class from the steaming data. By using the
concept of permutation and combination RBF kernel extracts the features and find out the relation between
them. This improves the novel class detect technique and provide more accuracy for classifying the data
Performance Comparision of Machine Learning AlgorithmsDinusha Dilanka
In this paper Compare the performance of two
classification algorithm. I t is useful to differentiate
algorithms based on computational performance rather
than classification accuracy alone. As although
classification accuracy between the algorithms is similar,
computational performance can differ significantly and it
can affect to the final results. So the objective of this paper
is to perform a comparative analysis of two machine
learning algorithms namely, K Nearest neighbor,
classification and Logistic Regression. In this paper it
was considered a large dataset of 7981 data points and 112
features. Then the performance of the above mentioned
machine learning algorithms are examined. In this paper
the processing time and accuracy of the different machine
learning techniques are being estimated by considering the
collected data set, over a 60% for train and remaining
40% for testing. The paper is organized as follows. In
Section I, introduction and background analysis of the
research is included and in section II, problem statement.
In Section III, our application and data analyze Process,
the testing environment, and the Methodology of our
analysis are being described briefly. Section IV comprises
the results of two algorithms. Finally, the paper concludes
with a discussion of future directions for research by
eliminating the problems existing with the current
research methodology.
Abstract:Reactive Power Optimization is a complex combinatorial optimization problem involving non-linear function having multiple local minima, non-linear and discontinuous constrains. This paper presents PS2O, which extends the dynamics of the canonical PSO algorithm by adding a significant ingredient that takes into account the symbiotic co evolution between species, Hybrid Evolutionary-Conventional Algorithm (HECA) that uses the abilities of evolutionary and conventional algorithm and Genetical Swarm Optimization (GSO), combines Genetic Algorithms (GA) and Particle Swarm Optimization (PSO).All the above said algorithms is used to overcome the Problem of premature convergence. PS2O, HECA , GSO is applied to Reactive Power Optimization problem and is evaluated on standard IEEE 57, practical 191 test Bus Systems. The results shows that all the three algorithms perform well in solving the reactive power problem and prevent premature convergence to high degree but still keep a rapid convergence. Of all the three PS2O has the edge in reducing the real power loss when compared to HECA & GSO.
Key benefits ADF modeling suite
One-stop modeling shop
Excellent software suite for tackling the most challenging problems in materials science and chemistry. Easy set up and analysis with GUI.
Fast computational toolbox
Working with hardware vendors, we optimize our codes for desktop computers and parallel supercomputers. Latest algorithms.
Heavy elements, spectroscopy, organic electronics
High-quality all-electron Slater basis sets for all elements. Accurate relativity. Many spectroscopic properties, from NMR to X-ray.
Unique organic electronics tools: charge transport, phosphorescence.
Understand chemical bonding
Unique insight in chemical bonds with many chemical analysis tools. Balanced charge decomposition schemes and density analysis tools.
Hassle-free installation, free trial
With parallel binaries for all popular platforms, the entire ADF suite installs out of the box. Try out our powerful modeling tools for free: http://www.scm.com/trial
Discuss your science with experts
With decades of experience, our expert support team (PhDs in chemistry & physics) will help you with any queries that may arise.
Urban strategies to promote resilient cities The case of enhancing Historic C...inventionjournals
This research tackles disaster prevention problems in dense urban areas, concentrating on the urban fire challenge in Historic Cairo district, Egypt, through disaster risk management approach. The study area suffers from the strike of several urban fire outbreaks, that resulted in disfiguring historic monuments and destroying unregulated traditional markets. Therefore, the study investigates the significance of hazard management and how can urban strategies improve the city resilient through reducing the impact of natural and man-made threats. The main findings of the research are the determination of the vulnerability factors in Historic Cairo district, either regarding management deficiency or issues related to the existing urban form. It is found that the absence of the mitigation and preparedness phases is the main problem in the risk management cycle in the case study. Additionally, the coping initiatives adopted by local authorities to address risks are random and insufficient. The study concludes with recommendations which invoke incorporating hazard management stages (pre disaster, during disaster and post disaster) into the process of evolving development planning. Finally, solutions are offered to mitigate, prepare, respond and recover from fire disasters in the case study. The solutions include urban policies, land-use planning, urban design outlines, safety regulation and public awareness and training.
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERijnlc
We aim to model an adaptive log file parser. As the content of log files often evolves over time, we established a dynamic statistical model which learns and adapts processing and parsing rules. First, we limit the amount of unstructured text by clustering based on semantics of log file lines. Next, we only take the most relevant cluster into account and focus only on those frequent patterns which lead to the desired output table similar to Vaarandi [10]. Furthermore, we transform the found frequent patterns and the output stating the parsed table into a Hidden Markov Model (HMM). We use this HMM as a specific, however, flexible representation of a pattern for log file parsing to maintain high quality output. After training our model on one system type and applying it to a different system with slightly different log file patterns, we achieve an accuracy over 99.99%
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERkevig
We aim to model an adaptive log file parser. As the content of log files often evolves over time, we
established a dynamic statistical model which learns and adapts processing and parsing rules. First, we
limit the amount of unstructured text by clustering based on semantics of log file lines. Next, we only take
the most relevant cluster into account and focus only on those frequent patterns which lead to the desired
output table similar to Vaarandi [10]. Furthermore, we transform the found frequent patterns and the
output stating the parsed table into a Hidden Markov Model (HMM). We use this HMM as a specific,
however, flexible representation of a pattern for log file parsing to maintain high quality output. After
training our model on one system type and applying it to a different system with slightly different log file
patterns, we achieve an accuracy over 99.99%.
This paper describe a proposed open semantic representation of chemical structure using JSON-LD (JSON for linked data) and an example of the semantic inferencing of a chemical structure concept (chirality).
ChemExtractor: Enhanced Rule-Based Capture and Identification of PDF Based Pr...Stuart Chalk
Paper presented at the 253rd ACS Meeting in San Francisco, CA. Presentation describes an approach to extracting chemical property data from PDF documents using single line regular expression (regex) created from regex snippets (for data elements) and a regex template (pattern of data elements).
Jean-Claude Bradley presents on "The implications of Open Notebook Science and other new forms of scientific communication for Nanoinformatics" at the Nanoinformatics 2010 conference on November 3, 2010. The presentation first covers the use of the laboratory knowledge management system SMIRP for nanotechnology applications during the period of 1999-2001 at Drexel University. The exporting of single experiments from SMIRP and publication to the Chemistry Preprint Archive is then described followed by the evolution to Open Notebook Science in 2005. Abstraction of semantic structure from ONS projects in the areas of drug discovery and solubility is then detailed as an efficient mechanism to provide web services and machine readable data feeds.
Talk by Martin Scharm at the COMBINE meeting September 2013 in Paris.
Find more information and download the slides at http://sems.uni-rostock.de/2013/09/sems-at-the-combine-2013/
Abstract: The processing power of computing devices has increased with number of available cores. This paper presents an approach
towards clustering of categorical data on multi-core platform. K-modes algorithm is used for clustering of categorical data which
uses simple dissimilarity measure for distance computation. The multi-core approach aims to achieve speedup in processing. Open
Multi Processing (OpenMP) is used to achieve parallelism in k-modes algorithm. OpenMP is a shared memory API that uses
thread approach using the fork-join model. The dataset used for experiment is Congressional Voting Dataset collected from UCI
repository. The dataset contains votes of members in categorical format provided in CSV format. The experiment is performed for
increased number of clusters and increasing size of dataset.
Abstract:Reactive Power Optimization is a complex combinatorial optimization problem involving non-linear function having multiple local minima, non-linear and discontinuous constrains. This paper presents PS2O, which extends the dynamics of the canonical PSO algorithm by adding a significant ingredient that takes into account the symbiotic co evolution between species, Hybrid Evolutionary-Conventional Algorithm (HECA) that uses the abilities of evolutionary and conventional algorithm and Genetical Swarm Optimization (GSO), combines Genetic Algorithms (GA) and Particle Swarm Optimization (PSO).All the above said algorithms is used to overcome the Problem of premature convergence. PS2O, HECA , GSO is applied to Reactive Power Optimization problem and is evaluated on standard IEEE 57, practical 191 test Bus Systems. The results shows that all the three algorithms perform well in solving the reactive power problem and prevent premature convergence to high degree but still keep a rapid convergence. Of all the three PS2O has the edge in reducing the real power loss when compared to HECA & GSO.
Keywords:PS2O, Hybrid Evolutionary-Conventional Algorithm, Genetical Swarm Optimization, Reactive Power Optimization.
Particle Swarm Optimization based K-Prototype Clustering Algorithm iosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Using AI Planning to Automate the Performance Analysis of SimulatorsRoland Ewald
Analyzing simulation algorithm performance is cumbersome: execute some runs, observe a performance metric, and analyze the results. Often, the results motivate follow-up experiments, which in turn may lead to additional experiments, and so on. This time-consuming and error-prone process can be automated with planning approaches from artificial intelligence, making simulator performance analysis more convenient and rigorous. This paper introduces ALeSiA, a prototypical system for automatic simulator performance analysis. It is independent of any specific simulation system and realizes a hypothesis-driven approach to evaluate performance.
Novel Class Detection Using RBF SVM Kernel from Feature Evolving Data Streamsirjes
In the data mining field the classification of data stream creates many problems. The challenges
faces in the data stream are infinite length, concept drift, concept evaluation and feature evolution. Most of the
existing system focuses on the only first two challenges. We propose a framework in which each classifier is
prepared with the novel class detector for addressing the two challenges concept drift and concept evaluation
and for addressing the feature evolution feature set homogeneous technique is proposed. We improved the
novel class detection module by building it more adaptive to evolving the stream. SVM based feature extraction
for RBF kernel method is also proposed for detecting the novel class from the steaming data. By using the
concept of permutation and combination RBF kernel extracts the features and find out the relation between
them. This improves the novel class detect technique and provide more accuracy for classifying the data
Performance Comparision of Machine Learning AlgorithmsDinusha Dilanka
In this paper Compare the performance of two
classification algorithm. I t is useful to differentiate
algorithms based on computational performance rather
than classification accuracy alone. As although
classification accuracy between the algorithms is similar,
computational performance can differ significantly and it
can affect to the final results. So the objective of this paper
is to perform a comparative analysis of two machine
learning algorithms namely, K Nearest neighbor,
classification and Logistic Regression. In this paper it
was considered a large dataset of 7981 data points and 112
features. Then the performance of the above mentioned
machine learning algorithms are examined. In this paper
the processing time and accuracy of the different machine
learning techniques are being estimated by considering the
collected data set, over a 60% for train and remaining
40% for testing. The paper is organized as follows. In
Section I, introduction and background analysis of the
research is included and in section II, problem statement.
In Section III, our application and data analyze Process,
the testing environment, and the Methodology of our
analysis are being described briefly. Section IV comprises
the results of two algorithms. Finally, the paper concludes
with a discussion of future directions for research by
eliminating the problems existing with the current
research methodology.
Abstract:Reactive Power Optimization is a complex combinatorial optimization problem involving non-linear function having multiple local minima, non-linear and discontinuous constrains. This paper presents PS2O, which extends the dynamics of the canonical PSO algorithm by adding a significant ingredient that takes into account the symbiotic co evolution between species, Hybrid Evolutionary-Conventional Algorithm (HECA) that uses the abilities of evolutionary and conventional algorithm and Genetical Swarm Optimization (GSO), combines Genetic Algorithms (GA) and Particle Swarm Optimization (PSO).All the above said algorithms is used to overcome the Problem of premature convergence. PS2O, HECA , GSO is applied to Reactive Power Optimization problem and is evaluated on standard IEEE 57, practical 191 test Bus Systems. The results shows that all the three algorithms perform well in solving the reactive power problem and prevent premature convergence to high degree but still keep a rapid convergence. Of all the three PS2O has the edge in reducing the real power loss when compared to HECA & GSO.
Key benefits ADF modeling suite
One-stop modeling shop
Excellent software suite for tackling the most challenging problems in materials science and chemistry. Easy set up and analysis with GUI.
Fast computational toolbox
Working with hardware vendors, we optimize our codes for desktop computers and parallel supercomputers. Latest algorithms.
Heavy elements, spectroscopy, organic electronics
High-quality all-electron Slater basis sets for all elements. Accurate relativity. Many spectroscopic properties, from NMR to X-ray.
Unique organic electronics tools: charge transport, phosphorescence.
Understand chemical bonding
Unique insight in chemical bonds with many chemical analysis tools. Balanced charge decomposition schemes and density analysis tools.
Hassle-free installation, free trial
With parallel binaries for all popular platforms, the entire ADF suite installs out of the box. Try out our powerful modeling tools for free: http://www.scm.com/trial
Discuss your science with experts
With decades of experience, our expert support team (PhDs in chemistry & physics) will help you with any queries that may arise.
Urban strategies to promote resilient cities The case of enhancing Historic C...inventionjournals
This research tackles disaster prevention problems in dense urban areas, concentrating on the urban fire challenge in Historic Cairo district, Egypt, through disaster risk management approach. The study area suffers from the strike of several urban fire outbreaks, that resulted in disfiguring historic monuments and destroying unregulated traditional markets. Therefore, the study investigates the significance of hazard management and how can urban strategies improve the city resilient through reducing the impact of natural and man-made threats. The main findings of the research are the determination of the vulnerability factors in Historic Cairo district, either regarding management deficiency or issues related to the existing urban form. It is found that the absence of the mitigation and preparedness phases is the main problem in the risk management cycle in the case study. Additionally, the coping initiatives adopted by local authorities to address risks are random and insufficient. The study concludes with recommendations which invoke incorporating hazard management stages (pre disaster, during disaster and post disaster) into the process of evolving development planning. Finally, solutions are offered to mitigate, prepare, respond and recover from fire disasters in the case study. The solutions include urban policies, land-use planning, urban design outlines, safety regulation and public awareness and training.
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERijnlc
We aim to model an adaptive log file parser. As the content of log files often evolves over time, we established a dynamic statistical model which learns and adapts processing and parsing rules. First, we limit the amount of unstructured text by clustering based on semantics of log file lines. Next, we only take the most relevant cluster into account and focus only on those frequent patterns which lead to the desired output table similar to Vaarandi [10]. Furthermore, we transform the found frequent patterns and the output stating the parsed table into a Hidden Markov Model (HMM). We use this HMM as a specific, however, flexible representation of a pattern for log file parsing to maintain high quality output. After training our model on one system type and applying it to a different system with slightly different log file patterns, we achieve an accuracy over 99.99%
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERkevig
We aim to model an adaptive log file parser. As the content of log files often evolves over time, we
established a dynamic statistical model which learns and adapts processing and parsing rules. First, we
limit the amount of unstructured text by clustering based on semantics of log file lines. Next, we only take
the most relevant cluster into account and focus only on those frequent patterns which lead to the desired
output table similar to Vaarandi [10]. Furthermore, we transform the found frequent patterns and the
output stating the parsed table into a Hidden Markov Model (HMM). We use this HMM as a specific,
however, flexible representation of a pattern for log file parsing to maintain high quality output. After
training our model on one system type and applying it to a different system with slightly different log file
patterns, we achieve an accuracy over 99.99%.
This paper describe a proposed open semantic representation of chemical structure using JSON-LD (JSON for linked data) and an example of the semantic inferencing of a chemical structure concept (chirality).
ChemExtractor: Enhanced Rule-Based Capture and Identification of PDF Based Pr...Stuart Chalk
Paper presented at the 253rd ACS Meeting in San Francisco, CA. Presentation describes an approach to extracting chemical property data from PDF documents using single line regular expression (regex) created from regex snippets (for data elements) and a regex template (pattern of data elements).
Started in 2004 (under ASTM Committee E13.15) the Analytical Information Markup Language (AnIML) is an XML based standard for capturing, sharing, viewing, and archiving analytical instrument data from any analytical technique.
This paper discusses the AnIML standard in terms of philosophy, structure, usage, and the resources available to work with the standard. Examples will be given for different techniques as well as strategies for migration of legacy data. Finally, the current status of the standard and time frame for promulgation through ASTM will be reported.
A Generic Scientific Data Model and Ontology for Representation of Chemical DataStuart Chalk
The current movement toward openness and sharing of data is likely to have a profound effect on the speed of scientific research and the complexity of questions we can answer. However, a fundamental problem with currently available datasets (and their metadata) is heterogeneity in terms of implementation, organization, and representation.
To address this issue we have developed a generic scientific data model (SDM) to organize and annotate raw and processed data, and the associated metadata. This paper will present the current status of the SDM, implementation of the SDM in JSON-LD, and the associated scientific data model ontology (SDMO). Example usage of the SDM to store data from a variety of sources with be discussed along with future plans for the work.
Scientific Units in the Electronic AgeStuart Chalk
Scientists have standardized on the SI unit system since the late 1700’s. While much work has been done over the years to refine and redefine the system, little has formally done to standardize the representation of the SI units in electronic systems.
This paper will present a summary of current efforts toward electronic representation of scientific units in text, XML, and RDF, an analysis of needs for current computer/network systems, and an outline of future work.
Toward Semantic Representation of Science in Electronic Laboratory Notebooks ...Stuart Chalk
An electronic laboratory Notebook (ELN) can be characterized as a system that allows scientists to capture the data and resources used in performing scientific experiments. This allows users to easily organize and find their data however, little information about the scientific process is recorded.
In this paper we highlight the current status of progress toward semantic representation of science in ELNs.
Rule-based Capture/Storage of Scientific Data from PDF Files and Export using...Stuart Chalk
Recently, the US government has mandated that publicly funded scientific research data be freely made available in a useable form, allowing integration of data in other systems. While this mandate has been articulated, existing publications and new papers (PDF) still do not provide accessible data, meaning that the usefulness is limited without human intervention.
This presentation outlines our efforts to extract scientific data from PDF files, using the PDFToText software and regular expressions (regex), and process it into a form that structures the data and its context (metadata). Extracted data is processed (cleaned, normalized), organized, and inserted into a contextually developed MySQL database. The data and metadata can then be output using a generic JSON-LD based scientific data model (SDM) under development in our laboratory.
Science is rapidly being brought into the electronic realm and electronic laboratory notebooks (ELN) are a big part of this activity. The representation of the scientific process in the context of an ELN is an important component to making the data recorded in ELNs semantically integrated.
This presentation outlined initial developments of an Electronic Notebook Ontology (ENO) that will help tie together the ExptML ontology, HCLS Community Profile data descriptions, and the VIVO-ISF ontology.
Sharing Science Data: Semantically Reimagining the IUPAC Solubility Series DataStuart Chalk
The IUPAC Solubility Data Series published its first volume in 1979. Since then over 100 volumes of high quality peer reviewed solubility data has been published, first in hardcopy and subsequently electronically as part of the Journal of Physical and Chemistry Reference Data.
In February of this year the National Institute of Standards and Technology (NIST) funded a grant to explore taking the 18 currently available online volumes of data and re-purpose them as a REST based website, with documented API, and semantic representation/annotation. In this way the high quality data from these volumes can be shared, both to humans and computers. In addition, the semantic representation of the data allows integration of the data with other semantically enabled data at repositories across the globe.
This presentation will give an overview of the process of schema development for the dataset, implementation in MySQL, website construction in the CakePHP framework, and architecture of the API access points. A report on the ontology development to support the project will also be discussed.
Bringing Flow injection Analysis to the Semantic WebStuart Chalk
As a mechanism to improve the sharing of data in Flow Injection Analysis, the Flow Analysis Database (http://www.fia.unf.edu) has been re-imagined to improve communication of the research on FIA, SIA, and related technologies across the vibrant communities in Europe, Asia, and the Americas.
This talk will present the new version of the Flow Analysis Database by highlighting
- The REST interface for each access to citation, analyte, matrix, technique, and keyword based resources
- Documented API for automated data integration
- Integration of the ChAMP specification
- Ontological support for FA concepts
- Individual user accounts with author bibliography
Future additions will include
- Language translation support using Google Translate
- ORCID integration
- Personal FIA library, and update notification
Reactions to the Open Spectral DatabaseStuart Chalk
In honor of JC Bradley, and the spirit of openness that he inspired, a new online resource called the Open Spectral Database (http://www.osdb.info) was announced in August of this year (aplha version). Built using open source tools, using open code, and open to community input about design and functionality, the OSD is available for anyone to submit spectral data and make it available to the scientific community. This paper will detail the reaction to the website, look at how the site has evolved since August (beta version), and offer a glimpse of what the future may bring for the site.
A Standard Data Format for Computational Chemistry: CSXStuart Chalk
An overview of the Common Standard for eXchange (CSX) a new markup language for the storage of computational chemistry calculation data. CSX stores publication and molecular system metadata along with calculation data and, optionally, raw input and output files associated with a calculation. The computational chemistry community is invited to participate in the development of CSX. For more information see http://www.chemicalsemantics.com.
Overview of the Analytical Information Markup Language (AnIML)Stuart Chalk
Overview and update on the Analytical Information Markup Language (AnIML) a standard for storage of analytical instrument data and associated metadata. http://animl.sourceforge.net
ACS 248th Paper 146 VIVO/ScientistsDB Integration into EurekaStuart Chalk
Development of plugins for access to researchers identified in VIVO on the ScientistsDB website. Also developed a plugin to access Elasticsearch from within Eureka.
ACS 248th Paper 136 JSmol/JSpecView Eureka IntegrationStuart Chalk
Integration of the combined JSmol/JSpecView molecular viewer/spectral viewer software in the Eureka Research Workbench. Can display molecular structures, spectra and the linked version where clicking on a peak shows molecular movement (IR).
Presentation on a project run in my Chemical Information Science course. Valuable referenced chemical data from 'reliable' static webpages was 'scraped', cleaned, and added to a database for search.
Presentation on the Chemical Analysis Metadata Platform (ChAMP) as a new project to characterize and organize metadata about chemical analysis methods. The project will develop an ontology, controlled vocabularies, and design rules
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxMAGOTI ERNEST
Although Artemia has been known to man for centuries, its use as a food for the culture of larval organisms apparently began only in the 1930s, when several investigators found that it made an excellent food for newly hatched fish larvae (Litvinenko et al., 2023). As aquaculture developed in the 1960s and ‘70s, the use of Artemia also became more widespread, due both to its convenience and to its nutritional value for larval organisms (Arenas-Pardo et al., 2024). The fact that Artemia dormant cysts can be stored for long periods in cans, and then used as an off-the-shelf food requiring only 24 h of incubation makes them the most convenient, least labor-intensive, live food available for aquaculture (Sorgeloos & Roubach, 2021). The nutritional value of Artemia, especially for marine organisms, is not constant, but varies both geographically and temporally. During the last decade, however, both the causes of Artemia nutritional variability and methods to improve poorquality Artemia have been identified (Loufi et al., 2024).
Brine shrimp (Artemia spp.) are used in marine aquaculture worldwide. Annually, more than 2,000 metric tons of dry cysts are used for cultivation of fish, crustacean, and shellfish larva. Brine shrimp are important to aquaculture because newly hatched brine shrimp nauplii (larvae) provide a food source for many fish fry (Mozanzadeh et al., 2021). Culture and harvesting of brine shrimp eggs represents another aspect of the aquaculture industry. Nauplii and metanauplii of Artemia, commonly known as brine shrimp, play a crucial role in aquaculture due to their nutritional value and suitability as live feed for many aquatic species, particularly in larval stages (Sorgeloos & Roubach, 2021).
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptxRASHMI M G
Abnormal or anomalous secondary growth in plants. It defines secondary growth as an increase in plant girth due to vascular cambium or cork cambium. Anomalous secondary growth does not follow the normal pattern of a single vascular cambium producing xylem internally and phloem externally.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
The Evolution of Science Education PraxiLabs’ Vision- Presentation (2).pdfmediapraxi
The rise of virtual labs has been a key tool in universities and schools, enhancing active learning and student engagement.
💥 Let’s dive into the future of science and shed light on PraxiLabs’ crucial role in transforming this field!
hematic appreciation test is a psychological assessment tool used to measure an individual's appreciation and understanding of specific themes or topics. This test helps to evaluate an individual's ability to connect different ideas and concepts within a given theme, as well as their overall comprehension and interpretation skills. The results of the test can provide valuable insights into an individual's cognitive abilities, creativity, and critical thinking skills
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
Toxic effects of heavy metals : Lead and Arsenicsanjana502982
Heavy metals are naturally occuring metallic chemical elements that have relatively high density, and are toxic at even low concentrations. All toxic metals are termed as heavy metals irrespective of their atomic mass and density, eg. arsenic, lead, mercury, cadmium, thallium, chromium, etc.
The ability to recreate computational results with minimal effort and actionable metrics provides a solid foundation for scientific research and software development. When people can replicate an analysis at the touch of a button using open-source software, open data, and methods to assess and compare proposals, it significantly eases verification of results, engagement with a diverse range of contributors, and progress. However, we have yet to fully achieve this; there are still many sociotechnical frictions.
Inspired by David Donoho's vision, this talk aims to revisit the three crucial pillars of frictionless reproducibility (data sharing, code sharing, and competitive challenges) with the perspective of deep software variability.
Our observation is that multiple layers — hardware, operating systems, third-party libraries, software versions, input data, compile-time options, and parameters — are subject to variability that exacerbates frictions but is also essential for achieving robust, generalizable results and fostering innovation. I will first review the literature, providing evidence of how the complex variability interactions across these layers affect qualitative and quantitative software properties, thereby complicating the reproduction and replication of scientific studies in various fields.
I will then present some software engineering and AI techniques that can support the strategic exploration of variability spaces. These include the use of abstractions and models (e.g., feature models), sampling strategies (e.g., uniform, random), cost-effective measurements (e.g., incremental build of software configurations), and dimensionality reduction methods (e.g., transfer learning, feature selection, software debloating).
I will finally argue that deep variability is both the problem and solution of frictionless reproducibility, calling the software science community to develop new methods and tools to manage variability and foster reproducibility in software systems.
Exposé invité Journées Nationales du GDR GPL 2024
BREEDING METHODS FOR DISEASE RESISTANCE.pptxRASHMI M G
Plant breeding for disease resistance is a strategy to reduce crop losses caused by disease. Plants have an innate immune system that allows them to recognize pathogens and provide resistance. However, breeding for long-lasting resistance often involves combining multiple resistance genes
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Semantic properties and units
1. Semantic Properties
and Units for Chemistry
Stuart Chalk
Department of Chemistry
University of North Florida
schalk@unf.edu
2. Semantic Chemical Property Data
IUPAC Green Book for Properties & Units for Chemistry
Concepts in Metrology – the VIM
QUDT – Semantic Metrology of the VIM
The IUPAC Gold Book – Now and Future
Conclusions
Overview
3. Semantic –> Resource Description Framework (RDF)
Store data as Subject-Predicate-Object triples
i.e. benzene containsAtomtype carbon (object)
benzene hasMolarMass 78.11 (literal)
molarMass hasUnit g/mol
Semantic Chemical Property Data
4. A generic data model to store scientific data
Can be implemented in any file/database format
For semantic applications
format in JSON-LD (https://www.w3.org/TR/json-ld/)
use the Scientific Data Model Ontology (SDMO)
Model + ontology creates hybrid relational/graph DB
SciData Data Model
12. Quantities
Quantity kinds
System of quantities
Dimensions
Dimension vectors
Units
Unit system
Metrology Concepts
If machines are going to capture and
process chemical property data
machine actionable representation of
these concepts is needed.
This can be encoded by semantic
annotation of property values and
units.
13. Quantities, Units, Dimensions and DataTypes (QUDT)
Defines common units and quantities
Can be used to define any unit or quantity
Include semantic representation of the VIM concepts
QUDT Ontology
https://qudt.org/
17. Counts of entities – 12 books
Radian – L/L
Steradian – L2/L2
Mole fraction – mol/mol
Parts per million – mg/kg or µg/g
Percent - %(w/w) or %(v/v)
Dimensionless Quantities
18. How do we show these are different to a computer?
Create a representation of the dimensionvector that is
unique for each dimensionless quantity
For radian (L/L)
L'0'M'1'T'0'I'0'H'0'N'0'J'0'D'0'_L'0'M'1'T'0'I'0'H'0'N
'0'J'0'D'0’_D1 OR M'1'_M'1’_D1
For steradian (L2/L2)
L'0'M’2'T'0'I'0'H'0'N'0'J'0'D'0'_L'0'M’2'T'0'I'0'H'0'N
'0'J'0'D'0’_D1 OR M’2’_M’2’_D1
Dimensionless Quantities
20. ‘The Compendium of Chemical Terminology’
Contains over 7000 definitions of chemistry concepts
Some terms are out of date
Currently under renovation to make terms machine
accessible
The IUPAC GoldBook
23. Add terms defined in all current IUPAC PAC
recommendations
Add synonyms, acronyms, legacy terms
Improve linking between terms
Create an ontology for Chemistry
Future of the GoldBook
24. Semantic chemical data is important in the move
toward knowledge discovery
Semantic unit representation requires clear
representation of quantity kinds and
dimensionvectors for interoperability
All chemical properties need to be represented
semantically
Conclusions