This document summarizes the results of analyzing definitions of patterns provided in a survey. The analysis sought to identify linguistic and semantic regularities in the definitions. Tools like word frequency analysis, tree diagrams, and dimensionality reduction techniques were used. Key findings included the identification of frequently used terms related to repetition, hierarchy, networks, and different conceptualizations of patterns from perception to application. The analysis revealed ambiguities and dualities in how patterns were defined, and aimed to develop a toolkit to help extract and detect patterns in text.
A Distributional Semantics Approach for Selective Reasoning on Commonsense Gr...Andre Freitas
Tasks such as question answering and semantic search are dependent
on the ability of querying & reasoning over large-scale commonsense knowledge
bases (KBs). However, dealing with commonsense data demands coping with
problems such as the increase in schema complexity, semantic inconsistency, incompleteness
and scalability. This paper proposes a selective graph navigation
mechanism based on a distributional relational semantic model which can be applied
to querying & reasoning over heterogeneous knowledge bases (KBs). The
approach can be used for approximative reasoning, querying and associational
knowledge discovery. In this paper we focus on commonsense reasoning as the
main motivational scenario for the approach. The approach focuses on addressing
the following problems: (i) providing a semantic selection mechanism for facts
which are relevant and meaningful in a specific reasoning & querying context
and (ii) allowing coping with information incompleteness in large KBs. The approach
is evaluated using ConceptNet as a commonsense KB, and achieved high
selectivity, high scalability and high accuracy in the selection of meaningful nav-
igational paths. Distributional semantics is also used as a principled mechanism
to cope with information incompleteness.
AN EMPIRICAL STUDY OF WORD SENSE DISAMBIGUATIONijnlc
Word Sense Disambiguation (WSD) is an important area which has an impact on improving the performance of applications of computational linguistics such as machine translation, information
retrieval, text summarization, question answering systems, etc. We have presented a brief history of WSD,
discussed the Supervised, Unsupervised, and Knowledge-based approaches for WSD. Though many WSD
algorithms exist, we have considered optimal and portable WSD algorithms as most appropriate since they
can be embedded easily in applications of computational linguistics. This paper will also provide an idea of
some of the WSD algorithms and their performances, which compares and assess the need of the word
sense disambiguation.
Conceptual Spaces for Cognitive Architectures: A Lingua Franca for Different ...Antonio Lieto
We claim that Conceptual Spaces offer a lingua franca that allows to unify and generalize many aspects of the symbolic, sub-symbolic and diagrammatic approaches (by overcoming some of their typical problems) and to integrate them on a common ground. In doing so we extend and detail some of the arguments explored by Gardenfors [23] for defending the need of a conceptual, intermediate, representation level between
the symbolic and the sub-symbolic one. Additionally, we argue that Conceptual Spaces could offer a unifying framework for interpreting many kinds of diagrammatic and analogical representations. As a consequence, their adoption could also favor the integration of diagrammatical representation and
reasoning in Cognitive Architectures
The spread and abundance of electronic documents requires automatic techniques for extracting useful information from the text they contain. The availability of conceptual taxonomies can be of great help, but manually building them is a complex and costly task. Building on previous work, we propose a technique to automatically extract conceptual graphs from text and reason with them. Since automated learning of taxonomies needs to be robust with respect to missing or partial knowledge and flexible with respect to noise, this work proposes a way to deal with these problems. The case of poor data/sparse concepts is tackled by finding generalizations among disjoint pieces of knowledge. Noise is
handled by introducing soft relationships among concepts rather than hard ones, and applying a probabilistic inferential setting. In particular, we propose to reason on the extracted graph using different kinds of relationships among concepts, where each arc/relationship is associated to a number that represents its likelihood among all possible worlds, and to face the problem of sparse knowledge by using generalizations among distant concepts as bridges between disjoint portions of knowledge.
A Distributional Semantics Approach for Selective Reasoning on Commonsense Gr...Andre Freitas
Tasks such as question answering and semantic search are dependent
on the ability of querying & reasoning over large-scale commonsense knowledge
bases (KBs). However, dealing with commonsense data demands coping with
problems such as the increase in schema complexity, semantic inconsistency, incompleteness
and scalability. This paper proposes a selective graph navigation
mechanism based on a distributional relational semantic model which can be applied
to querying & reasoning over heterogeneous knowledge bases (KBs). The
approach can be used for approximative reasoning, querying and associational
knowledge discovery. In this paper we focus on commonsense reasoning as the
main motivational scenario for the approach. The approach focuses on addressing
the following problems: (i) providing a semantic selection mechanism for facts
which are relevant and meaningful in a specific reasoning & querying context
and (ii) allowing coping with information incompleteness in large KBs. The approach
is evaluated using ConceptNet as a commonsense KB, and achieved high
selectivity, high scalability and high accuracy in the selection of meaningful nav-
igational paths. Distributional semantics is also used as a principled mechanism
to cope with information incompleteness.
AN EMPIRICAL STUDY OF WORD SENSE DISAMBIGUATIONijnlc
Word Sense Disambiguation (WSD) is an important area which has an impact on improving the performance of applications of computational linguistics such as machine translation, information
retrieval, text summarization, question answering systems, etc. We have presented a brief history of WSD,
discussed the Supervised, Unsupervised, and Knowledge-based approaches for WSD. Though many WSD
algorithms exist, we have considered optimal and portable WSD algorithms as most appropriate since they
can be embedded easily in applications of computational linguistics. This paper will also provide an idea of
some of the WSD algorithms and their performances, which compares and assess the need of the word
sense disambiguation.
Conceptual Spaces for Cognitive Architectures: A Lingua Franca for Different ...Antonio Lieto
We claim that Conceptual Spaces offer a lingua franca that allows to unify and generalize many aspects of the symbolic, sub-symbolic and diagrammatic approaches (by overcoming some of their typical problems) and to integrate them on a common ground. In doing so we extend and detail some of the arguments explored by Gardenfors [23] for defending the need of a conceptual, intermediate, representation level between
the symbolic and the sub-symbolic one. Additionally, we argue that Conceptual Spaces could offer a unifying framework for interpreting many kinds of diagrammatic and analogical representations. As a consequence, their adoption could also favor the integration of diagrammatical representation and
reasoning in Cognitive Architectures
The spread and abundance of electronic documents requires automatic techniques for extracting useful information from the text they contain. The availability of conceptual taxonomies can be of great help, but manually building them is a complex and costly task. Building on previous work, we propose a technique to automatically extract conceptual graphs from text and reason with them. Since automated learning of taxonomies needs to be robust with respect to missing or partial knowledge and flexible with respect to noise, this work proposes a way to deal with these problems. The case of poor data/sparse concepts is tackled by finding generalizations among disjoint pieces of knowledge. Noise is
handled by introducing soft relationships among concepts rather than hard ones, and applying a probabilistic inferential setting. In particular, we propose to reason on the extracted graph using different kinds of relationships among concepts, where each arc/relationship is associated to a number that represents its likelihood among all possible worlds, and to face the problem of sparse knowledge by using generalizations among distant concepts as bridges between disjoint portions of knowledge.
User centered design assumes that a research phase with a representative sample of the final users should be the basis for the definition of the functional and soft requirements of a project. How can we translate the results of the ux research into actionable requirements?
In my talk, I wish to give you some suggestions on how to informally analyse the verbal results of the ux research to identify the schemata, the ontologies, the taxonomies and the functions of your application.
2007. Introduction to the panel 'Pragmatic Interfaces' organised by the authors at the International Pragmatics Conference (IPRA) in Goteborg (Sweden), July 2007. Didier Maillat and Louis de Saussure
On the Semantic Mapping of Schema-agnostic Queries: A Preliminary StudyAndre Freitas
The growing size, heterogeneity and complexity of databases
demand the creation of strategies to facilitate users and systems to consume
data. Ideally, query mechanisms should be schema-agnostic or
vocabulary-independent, i.e. they should be able to match user queries
in their own vocabulary and syntax to the data, abstracting data consumers
from the representation of the data. Despite being a central requirement across natural language interfaces and entity search, there is a lack on the conceptual analysis of schema-agnosticism and on the associated semantic differences between queries and databases. This work aims at providing an initial conceptualization for schema-agnostic queries aiming at providing a fine-grained classification which can support the scoping, evaluation and development of semantic matching approaches for schema-agnostic queries.
Extending the knowledge level of cognitive architectures with Conceptual Spac...Antonio Lieto
Extending the knowledge level of cognitive architectures with Conceptual Spaces (+ a case study with Dual-PECCS: a hybrid knowledge representation system for common sense reasoning). Talk given at Stockholm, September 2016.
Conceptual Interoperability and Biomedical DataJim McCusker
The goals of conceptual interoperability are:
Make similar but distinct data resources available for search, conversion, and inter-mapping in a way that mirrors human understanding of the data being searched.
Make data resources that use cross-cutting models (HL7-RIM, provenance models, etc.) interoperable with domain-specific models without explicit mappings between them.
Compositional distributional models of meaning (CDMs) aim to unify the two prominent semantic paradigms in natural language: The type-logical compositional approach of formal semantics, and the quantitative perspective of vector space models of meaning. This presentation gives an overview of state-of-the-art research on the field. We review three generic classes of CDMs: vector mixtures, tensor-based models, and deep-learning models.
How hard is this Query? Measuring the Semantic Complexity of Schema-agnostic ...Andre Freitas
The growing size, heterogeneity and complexity of databases demand the creation of strategies to facilitate users and systems to consume data. Ideally, query mechanisms should be schema-agnostic, i.e. they should be able to match user queries in their own vocabulary and syntax to the data, abstracting data consumers from the representation of the data. This work provides an informationtheoretical framework to evaluate the semantic complexity involved in the query-database communication, under a schema-agnostic query scenario. Different entropy measures are introduced to quantify the semantic phenomena involved in the user-database communication, including structural complexity, ambiguity, synonymy and vagueness. The entropy measures are validated using natural language queries over Semantic Web databases. The analysis of the semantic complexity is used to improve the understanding of the core semantic dimensions present at the query-data matching process, allowing the improvement of the design of schema-agnostic query mechanisms and defining measures which can be used to assess the semantic uncertainty or difficulty behind a schema-agnostic querying task.
The CLUES database: automated search for linguistic cognatesMark Planigale
Overview of the design of the CLUES database, developed as an aid to the comparative method in historical linguistics. Includes information on the design of the database and the strategies used to detect correlate forms (potential cognates), including metrics used to rate similarity of form and meaning.
A Method for coordinative syntactic disambiguation in SpanishNervo Verdezoto
Implementation of a method for coordinative syntactic disambiguation in Spanish using the python language (Natural Language Toolkit - NLTK)
Spanish is considered a complex language because its variability structure and different rules. Some of these features can produce ambiguity problems and the most common are:
- Impersonal “se” construction.
- Coordinative and Prepositional constructions
On the other hand, the Natural Language Toolkit - NLTK was used because it is a suite of Python modules distributed under open source license (nltk.org). This toolkit includes a large collection of corpora, statistical natural language processing, extensive documentation, graphical demonstrations and sample data, etc.
User centered design assumes that a research phase with a representative sample of the final users should be the basis for the definition of the functional and soft requirements of a project. How can we translate the results of the ux research into actionable requirements?
In my talk, I wish to give you some suggestions on how to informally analyse the verbal results of the ux research to identify the schemata, the ontologies, the taxonomies and the functions of your application.
2007. Introduction to the panel 'Pragmatic Interfaces' organised by the authors at the International Pragmatics Conference (IPRA) in Goteborg (Sweden), July 2007. Didier Maillat and Louis de Saussure
On the Semantic Mapping of Schema-agnostic Queries: A Preliminary StudyAndre Freitas
The growing size, heterogeneity and complexity of databases
demand the creation of strategies to facilitate users and systems to consume
data. Ideally, query mechanisms should be schema-agnostic or
vocabulary-independent, i.e. they should be able to match user queries
in their own vocabulary and syntax to the data, abstracting data consumers
from the representation of the data. Despite being a central requirement across natural language interfaces and entity search, there is a lack on the conceptual analysis of schema-agnosticism and on the associated semantic differences between queries and databases. This work aims at providing an initial conceptualization for schema-agnostic queries aiming at providing a fine-grained classification which can support the scoping, evaluation and development of semantic matching approaches for schema-agnostic queries.
Extending the knowledge level of cognitive architectures with Conceptual Spac...Antonio Lieto
Extending the knowledge level of cognitive architectures with Conceptual Spaces (+ a case study with Dual-PECCS: a hybrid knowledge representation system for common sense reasoning). Talk given at Stockholm, September 2016.
Conceptual Interoperability and Biomedical DataJim McCusker
The goals of conceptual interoperability are:
Make similar but distinct data resources available for search, conversion, and inter-mapping in a way that mirrors human understanding of the data being searched.
Make data resources that use cross-cutting models (HL7-RIM, provenance models, etc.) interoperable with domain-specific models without explicit mappings between them.
Compositional distributional models of meaning (CDMs) aim to unify the two prominent semantic paradigms in natural language: The type-logical compositional approach of formal semantics, and the quantitative perspective of vector space models of meaning. This presentation gives an overview of state-of-the-art research on the field. We review three generic classes of CDMs: vector mixtures, tensor-based models, and deep-learning models.
How hard is this Query? Measuring the Semantic Complexity of Schema-agnostic ...Andre Freitas
The growing size, heterogeneity and complexity of databases demand the creation of strategies to facilitate users and systems to consume data. Ideally, query mechanisms should be schema-agnostic, i.e. they should be able to match user queries in their own vocabulary and syntax to the data, abstracting data consumers from the representation of the data. This work provides an informationtheoretical framework to evaluate the semantic complexity involved in the query-database communication, under a schema-agnostic query scenario. Different entropy measures are introduced to quantify the semantic phenomena involved in the user-database communication, including structural complexity, ambiguity, synonymy and vagueness. The entropy measures are validated using natural language queries over Semantic Web databases. The analysis of the semantic complexity is used to improve the understanding of the core semantic dimensions present at the query-data matching process, allowing the improvement of the design of schema-agnostic query mechanisms and defining measures which can be used to assess the semantic uncertainty or difficulty behind a schema-agnostic querying task.
The CLUES database: automated search for linguistic cognatesMark Planigale
Overview of the design of the CLUES database, developed as an aid to the comparative method in historical linguistics. Includes information on the design of the database and the strategies used to detect correlate forms (potential cognates), including metrics used to rate similarity of form and meaning.
A Method for coordinative syntactic disambiguation in SpanishNervo Verdezoto
Implementation of a method for coordinative syntactic disambiguation in Spanish using the python language (Natural Language Toolkit - NLTK)
Spanish is considered a complex language because its variability structure and different rules. Some of these features can produce ambiguity problems and the most common are:
- Impersonal “se” construction.
- Coordinative and Prepositional constructions
On the other hand, the Natural Language Toolkit - NLTK was used because it is a suite of Python modules distributed under open source license (nltk.org). This toolkit includes a large collection of corpora, statistical natural language processing, extensive documentation, graphical demonstrations and sample data, etc.
Marcelo Funes-Gallanzi - Simplish - Computational intelligence unconferenceDaniel Lewis
At the computational intelligence unconference 2014, Marcelo Funes-Gallanzi presented Simplish, a system for the conversion of text into Simple English. Here are his slides.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Identifying Hot Topic Trends in Streaming Text Data Using News Sequential Evo...Shakas Technologies
Identifying Hot Topic Trends in Streaming Text Data Using News Sequential Evolution Model Based on Distributed Representations.
Shakas Technologies ( Galaxy of Knowledge)
#11/A 2nd East Main Road,
Gandhi Nagar,
Vellore - 632006.
Mobile : +91-9500218218 / 8220150373| land line- 0416- 3552723
Shakas Training & Development | Shakas Sales & Services | Shakas Educational Trust|IEEE projects | Research & Development | Journal Publication |
Email : info@shakastech.com | shakastech@gmail.com |
website: www.shakastech.com
Facebook: https://www.facebook.com/pages/Shakas-Technologies
SEMANTIC NETWORK BASED MECHANISMS FOR KNOWLEDGE ACQUISITIONcscpconf
This article summarizes research work started with the SeiPro2S (Semantically Enhanced Intellectual Property Protection System) system designed to protect resources from the unauthorized use of intellectual property. The system implements semantic network as a structure of knowledge representation and a new idea of semantic compression. As the author proved that semantic compression is viable concept for English, he decided to focus on potential applications. An algorithm is presented that employing semantic network WiSENet for knowledge acquisition with flexible rules that yield high precision results. Developed algorithm is implemented as a Finite State Automaton with advanced methods for triggering desired actions. Detailed discussion is given with description of devised algorithm, usage examples and results of experiments.
In recent years, great advances have been made in the speed, accuracy, and coverage of automatic word
sense disambiguator systems that, given a word appearing in a certain context, can identify the sense of
that word. In this paper we consider the problem of deciding whether same words contained in different
documents are related to the same meaning or are homonyms. Our goal is to improve the estimate of the
similarity of documents in which some words may be used with different meanings. We present three new
strategies for solving this problem, which are used to filter out homonyms from the similarity computation.
Two of them are intrinsically non-semantic, whereas the other one has a semantic flavor and can also be
applied to word sense disambiguation. The three strategies have been embedded in an article document
recommendation system that one of the most important Italian ad-serving companies offers to its customers.
In recent years, great advances have been made in the speed, accuracy, and coverage of automatic word
sense disambiguator systems that, given a word appearing in a certain context, can identify the sense of
that word. In this paper we consider the problem of deciding whether same words contained in different
documents are related to the same meaning or are homonyms. Our goal is to improve the estimate of the
similarity of documents in which some words may be used with different meanings. We present three new
strategies for solving this problem, which are used to filter out homonyms from the similarity computation.
Two of them are intrinsically non-semantic, whereas the other one has a semantic flavor and can also be
applied to word sense disambiguation. The three strategies have been embedded in an article document
recommendation system that one of the most important Italian ad-serving companies offers to its customers
French machine reading for question answeringAli Kabbadj
This paper proposes to unlock the main barrier to machine reading and comprehension French natural language texts. This open the way to machine to find to a question a precise answer buried in the mass of unstructured French texts. Or to create a universal French chatbot. Deep learning has produced extremely promising results for various tasks in natural language understanding particularly topic classification, sentiment analysis, question answering, and language translation. But to be effective Deep Learning methods need very large training da-tasets. Until now these technics cannot be actually used for French texts Question Answering (Q&A) applications since there was not a large Q&A training dataset. We produced a large (100 000+) French training Dataset for Q&A by translating and adapting the English SQuAD v1.1 Dataset, a GloVe French word and character embed-ding vectors from Wikipedia French Dump. We trained and evaluated of three different Q&A neural network ar-chitectures in French and carried out a French Q&A models with F1 score around 70%.
Salas, V. (2024) "John of St. Thomas (Poinsot) on the Science of Sacred Theol...Studia Poinsotiana
I Introduction
II Subalternation and Theology
III Theology and Dogmatic Declarations
IV The Mixed Principles of Theology
V Virtual Revelation: The Unity of Theology
VI Theology as a Natural Science
VII Theology’s Certitude
VIII Conclusion
Notes
Bibliography
All the contents are fully attributable to the author, Doctor Victor Salas. Should you wish to get this text republished, get in touch with the author or the editorial committee of the Studia Poinsotiana. Insofar as possible, we will be happy to broker your contact.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
1. Survey Mapping the Landscape of Patterns across Domains:
LINGUISTIC AND SEMANTIC ANALYSIS OF PATTERN DEFINITIONS
IN ANSWERS TO OPEN QUESTIONS
BCSSS Research Group Systems Science and Pattern Literacy
Working Paper Version 1.0 - April 2019
Maria Lenzi - Helene Finidori
This is the second set of results from the Survey Mapping the Landscape of Patterns
across Domains initiated by the BCSSS Research Group Systems Science and Pattern
Literacy in early 2018.
Our intention here was to extract linguistic and semantic regularities from the survey
answers where respondents defined patterns in a few sentences.
We assumed that we would find frequently used general terms and associations between
terms. We therefore sought to extract couplings between linguistic and semantic
structures which could be generalized and re-used for the identification of patterns and the
development of patterns knowledge. Such „knowledge units“ would support the
development of a formal language and a diagrammatic presentation of patterns.
We applied manual analysis, tree diagrams and software tools from the open source
platform Voyant Tools, such as Cirrus, which displays word frequencies as wordclouds, as
well as tailed-Stochastic Neighbor Embedding (t-SNE) and Principal Component Analysis
(PCA), which help reduce dimensions of complexity and capture the essential structure of
the data.
We drilled down in the depth of the corpus of answers to extract regularities from varied
and ambivalent pattern definitions.
t-SNE and PCA tools supported the identification of general manifestations and attributes
of patterns, as well as associations and consistent combinations of terms. Some
persistent associations revealed mature concepts behind them.
We also observed that the application of simple software tools like Cirrus constrained the
interpretation of definitions provided in different contexts. More sophisticated linguistic and
semantic structures require more sophisticated tools and a multi-dimensional approach.
We will continue to develop a re-usable toolkit for extraction and detection of patterns.
These will include tools for text analysis to complement those presented in this report.
A sample of 140 definitions is not sufficient to make all embracing conclusions about
existing concepts and perspectives of seeing patterns. Yet it can generate some insights
and generalizations to be validated and fine-tuned with further experience and research.
Our interpretation of survey data is not final and we welcome proposals and ideas
concerning objectives, interpretations and application of tools.
Creative Commons Licence BY - NC - SA 1
2. SUMMARIZED OBJECTIVES
● identifying linguistic regularities in definitions of patterns
● identifying semantic regularities coupled with linguistic regularities
● identifying concepts and perspectives behind definitions
● creating „ knowledge units“ for potential re-use
● developing a toolkit for extraction and detection of patterns
APPLIED TOOLS
- Manual analysis
- Cirrus Diagram using Voyant Tools
- Tree Diagrams using Vensim Software
- t-SNE (tailed stochastic neighbor embedding) using Voyant Tools
- PCA (Principal Component Analysis) using Voyant Tools
Creative Commons Licence BY - NC - SA 2
3. IDENTIFICATION OF GENERAL TERMS WITH THE CIRRUS DIAGRAM
We started with eliciting the most frequently used terms in the corpus of definitions of
patterns with the help of wordcloud diagrams (Cirrus from Voyant tools).
https://voyant-tools.org/?corpus=3f5ebfd521bae2cf37bd6da440db7dcd&stopList=keyword
s-1cbc1dabfb6ec16ef24060bdea65747f&whiteList=&view=Cirrus
Cirrus diagrams helps visualize relative frequencies of usage of words, with the size of
each token in the diagram reflecting the frequency of usage.
Creative Commons Licence BY - NC - SA 3
4. We clustered frequently used terms around the following key dimensions of pattern
definitions:
- similarity
- dynamics
- perception
- repetition
- function
- cognition
- variety
- structure
SOME OBSERVATIONS ON CONSTRAINTS IN GENERATING CIRRUS DIAGRAMS
It is not enough just to extract the frequency of usage of terms in a corpus of answers -
there are ambiguity, asymmetries, redundancies, contextuality and multiplicity of concepts
behind definitions.
All this affect the interpretation of Cirrus Diagram results and constrain the application of
such simple tools for the identification of linguistic and semantic regularities in text.
OBSERVATION 1:
asymmetry of identified frequencies of usage
frequently used terms like „problem“, „phenomena“, „something“ and others were
repeatedly used in the same individual answers.
Example 1: „A pattern is a generalized way of describing a problem in a context, and a
solution to the problem. Patterns help people by a) creating a vocabulary to record or
communicate the problem and the solution b) putting the problem in a context, including
relationships to other patterns c) abstract the problem and the solution to help people
think about the problem space in which the pattern lives“
Example 2: A pattern is a semi-formal and mildly abstract description of the kernel of
successful solutions to problems that occur over and over again in similar contexts. They
help us understanding what works to solve whatever problem, in whatever realm. Once
we've understood this, we must adapt this pattern to our very concrete context, i.e. our
very concrete problem. That is: a pattern is a tool but we must learn how to apply it.
Example 3: An aggregate or cluster of phenomena that recurs in observable/
recognizable form, whether physical or psychosocial or spiritual. A product of synthesis of
these observed phenomena. A help to save energy in recurring situations and a
hindrance to innovation and behaviour change.
Example 4: Something in common. Something with potential. Something worth
Creative Commons Licence BY - NC - SA 4
5. exploring. Generation of a new boundary.
OBSERVATION 2:
linguistic regularities are applied in different contexts that change their meaning
Examples of contextuality for different tokens are shown below – it is always important to
pay attention to the context in which the terms are applied.
Paying attention to contextuality supports identification of more sophisticated linguistic
and semantic regularities.
We can see below how context changes the meaning of some terms:
„SETS“ = s.
Patterns are
- (arrangements of) s. of tilings;
- s. of relationships (among relata),
- s. of relationships (that have some form of repetition),
- s. of actions (that can be identified within different contexts),
- s. of images (that can be identified within different context),
- s. of aspects (frequently found together),
- (regular) s. of features,
- s. of elements (with that pattern)
„DIFFERENT“ = d.
Patterns
- (repeat themselves) across d.subjects
- (can be identified) within d.contexts (by virtue of their signature features),
- (meaningful interaction) of d.competences,
- (can show) d.levels of abstraction,
- (noticed repeatedly) at d.times or in d.contexts;
- (can show up) in d.forms;
- (can be realized) through d.mediums,(something)that d.objects(have in common);
- (a regular structure or process) that appears across d.fields or instantiations,
„ARRANGEMENT“= a.
Patterns are
- a. of sets (of tilings),
- a. of whatever kind (which is able to keep its configuration),
- a. (that gives identity and meaning),
- a structural a. of units (abstracted from the particular types),
- a persistent or frequent a. (of interacting elements),
- a. that has some form (of recognized regularity)
Creative Commons Licence BY - NC - SA 5
6. „REPEATING“= r.
Patterns are
- (evolution of information) on a r. path,
- r.abstractions (of features of a system),
- r. in time and space (or both),
- (an observed) r. of a sequence,
- (an emerging) constellation of r.events,
- (sequences of ) r. elements,
- a r. order of elements (that serve some purpose),
- a r. geometry (covering a surface),
- a r. phenomenon (that we can perceive),
- a r. structure or process (that appears across fields or instantiations)
„TIME“= t.
Patterns are/do
- (repeat) over t.,
- (dynamically stable) over t.,
- (somewhat stable for) some period of t.,
- (repeat themselves) across t.,
- (repeating) in t.,
- (takes attention) over t.,
- space and t. (interaction),
- (iterated at least two) times,
- (noticed repeatedly) at different t.,
- (a recognizable) repetition in t. (and/or space),
- (repeated) through t.,
- (recognizable connections) over t.,
- (a regular behavior) over t.
Creative Commons Licence BY - NC - SA 6
7. MANUAL SEMANTIC ANALYSIS WITH THE HELP OF TREE DIAGRAMS
We applied manual analysis combined with bottom up approach and tree diagrams to
identify linguistic and semantic regularities in the corpus of pattern definitions.
With a special interest for a higher order structure in definitions of patterns we payed
attention to the duality of inclusive and lateral relationships and to ambiguity which were
present in answers.
We made following observations:
OBSERVATION 1 : REPETITION
Patterns have to do with regularity and repetition
Relevance of regularity and repetition is also confirmed by semantic regularities extracted
from the t-SNE and PCA analysis.
Regularity and repetition are mentioned in connection with time, space, context and
structure.
Applied terms:
„repetitive“, „regularities“, „repeating“, „recurring“, „repeated“, „repetition“, „repeat“,
„regular“
OBSERVATION 2: HIERARCHY
Patterns can be defined as entities of a higher order
We have identified a logical hierarchy based on inclusive relationships in definitions of
patterns:
- pattern is some kind of an entity (first order)
- pattern is an order of some kinds of entities (second order)
- pattern is an order of orders of some kinds of entities (third order)
The different orders of logical hierarchy in definitions of patterns are illustrated in the
following tree diagrams which show the orders extracted from single answers.
Creative Commons Licence BY - NC - SA 7
8. 1.Examples of the first-order definitions*: „Pattern is some kind of entity“
*we can observe, that definitions of the first-order in many cases are a hidden form of the
higher-order definitions, because such terms as „arrangement“, „relationship“ etc. already
embrace an additional order (arrangement of elements, relationship of entities etc.)
Creative Commons Licence BY - NC - SA 8
9. 2.Examples of second-order definitions: „Pattern is an order of some kinds of entities“
Creative Commons Licence BY - NC - SA 9
10. 3.Examples of third-order definitions: „Pattern is an order of orders of some kinds of
entities“
Creative Commons Licence BY - NC - SA 10
11. OBSERVATION 3: NETWORKS
patterns can be defined as networks of lateral relationships
Such terms as „arrangement“, „configuration“, „network“, „relationship“, „combination“,
„connection“ which were used to demonstrate a higher-order architecture of patterns can
also reflect lateral relationships between entities or parts - without hierarchical logical
subordination.
There are also clear definitions of patterns based on lateral relationships between entities.
Examples of definitions based on lateral relationships between entities:
Pattern(s) is/are:
● an arrangement of (whatever their kind) which is able to keep its configuration
dynamically stable over time
● an orderly dynamic that links things that may either be obviously or covertly
connected.
● something which rhymes with something else
● a perceived relationship among discrete elements, often in dynamic interaction.
● a collection of concepts or things that have some manner of connection to each
other, however vague
● semantic models defining inter-dependencies between related concepts.
OBSERVATION 4: FROM PERCEPTION TO APPLICATION
Definitions of patterns can be segmented into „conceived“ , „perceived“, „intuited“ and
„utilized“
These definitions also reflect different concepts and perspectives of patterns. Some single
answers combine multiple perspectives and allow multiple interpretations.
„Conceived“ definitions have to do with cognition and construction of meaning –
patterns are defined as a product of thinking, abstracting and generalizing:
- a pattern is a generalized way of describing a problem in a context, and a solution
to the problem
- a pattern is a way of abstracting and decomposing proven solutions to problems
- a pattern is an abstracted representation of common structures
- generic formula that be applied in multiple domains
Creative Commons Licence BY - NC - SA 11
12. „Perceived“ definitions present patterns as a result of observation and experience:
- an aggregate or cluster of phenomena that recurs in observable/recognizable form,
whether physical or psychosocial or spiritual
- a pattern is a systemic holistic creation. A pattern on a flower, or universe, or
design, seems always settling when observed, and balanced for all parts/parties
involved
- a pattern is a non-random, replicable set of relationships among relata, which can
be observed as phenomena and/or qualitatively experienced by participants
„Intuited“ definitions reflect intuition, sensation, emotions and creative states without
conscious reasoning - with special cases of application of metaphors:
- a complexity drop. A signal.
- a pattern starts as a simple seed that proliferates, building complexity, often adding
scales with self-similarity.
- patterns are the way in which we speak, communicate, think and create
- it has certain qualities. A flavour of influencing.
- it's linking everything together. Colour to sound to nature.
- serve as scaffolding. Patterns can emerge from creative cognition.
„Utilized“ definitions show patterns as means for practical and purposeful applications.
- a pattern is a semi-formal and mildly abstract description of the kernel of successful
solutions to problems that occur over and over again in similar contexts
- a representation of a generic response to a recurrent situation
- a piece of knowledge conceptually structured (semantic) in such a way that, either
it can be purposefuly re-used in a larger frame ...or automatically mobilised
(including triggered) when a certain condition is sensed ...
OBSERVATION 5: AMBIGUITY AND DUALITIES
Definitions of patterns are ambiguous and contain dualities and contradictions
There is some kind of a semantic ambiguity in definitions of patterns as a „way“, „may“ or
„something“ with „some“ qualities.
There are also dualities in definitions of patterns as a process and/or a structure; as
hierarchical and/or lateral connections; as bottom-up and/or top-down processes; as
natural and/or human and/or social constructs.
Creative Commons Licence BY - NC - SA 12
13. We identified some complementary terms or dualities, that evolve from the corpus of
definitions of patterns and are also contained in single answers:
regular random
structure process
stable dynamic
unique repeating
similar different
form flow
system network
purpose function
problem solution
energy matter
thinking acting
perception cognition
Examples of ambiguous definitions of patterns:
- Things (plural) with some coherence that can be distinguished from a background
of randomness or other things
- Pattern is a conclusion made and sustained by the mind (an observer) stating that
what they are faced with "something" --in some sense tangible and durable-- rather
than just background mess...
- Patterns guide thought and engage consciousness; serve as scaffolding.
- Patterns need not be visualizable, but many are. Patterns can be temporal, e.g.,
music and speech. Patterns are assigned meaning, or meaning is projected on
them. Patterns have no intrinsic meaning. Structuring-Processes and
Processing-Structures are a complementarity
- A pattern is a generalized way of describing a problem in a context, and a solution
to the problem.
- A pattern has two parts: an issue, which is an understanding of living activity,
human and otherwise; and a solution, which guides the physical form of the built
and natural environment.
- Patterns are a unique combination of behaviors, acts, qualities or events that
repeat themselves over space and time
-
EXPERIMENTS WITH DIMENSIONALITY REDUCTION
To complete our manual analysis we experimented with more sophisticated software tools
that support reduction of dimensionality - we applied t-SNE (tailed-Stochastic Neighbor
Embedding) and PCA (Principal Component Analysis) on the corpus of pattern definitions.
We wanted to identify linguistic and semantic regularities in the data structure and to see
whether there was a correspondence with the semantic regularities from the manual
analysis.
Creative Commons Licence BY - NC - SA 13
14. IDENTIFICATION OF LINGUISTIC AND SEMANTIC REGULARITIES WITH t-SNE
t-SNE - t(tailed)-Stochastic Neighbor Embedding - is a software technique that supports
visualization of high-dimensional data in a two or three-dimensional map with respect to
non-linearity.
t-SNE preserves characteristics of a data structure when it is projected from a higher
dimension to a lower dimension and helps to avoid „crowding problem“ in visualization of
proximities.
We analyzed the corpus of answers to the open question of the survey:
„How would you describe or define a pattern (in a couple of sentences)?“
792 terms were identified in the corpus of 140 answers.
Here is a link to a t-SNE setting supported by Voyant Tools:
https://voyant-tools.org/?corpus=3f5ebfd521bae2cf37bd6da440db7dcd&stopList=keyword
s-44c3d8895e669029e19125fcc0735df9&limit=60&view=ScatterPlot
Our readers are welcome to experiment with different settings and to share ideas and
interpretations with us.
SETTING FOR t-SNE-ANALYSIS
number of terms - 60
perplexity – 15
iterations – 5000
dimensions - 2
Number of terms is reduced from 792 to 60 most frequently used terms.
Perplexity is a rough equivalent to the number of nearest terms (neighbors) and we have
set perplexity to the length of a „couple of sentences“ – 15 terms.
Perplexity and number of iterations are tuned to achieve a relative stability of results.
Points above zero represent the growing proximity (attraction) to other terms, and
points below zero - growing repulsion (distance) to other terms.
Sizes of circles correspond to the frequency of use of the terms.
The results of t-SNE iterations are shown below.
Creative Commons Licence BY - NC - SA 14
15. t-SNE setting: number of terms - 60, perplexity – 15, iterations – 5000, dimensions - 2
INTERPRETATION OF RESULTS OF t-SNE ANALYSIS
Being a stochastic tools t -SNE does not deliver an exact repetition of results – trials with
the same setting always show somehow different results – yet it is possible to identify
repeating consistent structures through multiple trials with the same setting.
t-SNE evens out semantic distances and we can follow the relative difference between
distances and the values of repulsion (below zero) and attraction (above zero).
Frequently used terms that have a shorter semantic distance to other terms in the corpus
of answers are those which appear in the attraction area above zero. They can be
characterized as general attributes of patterns.
We assumed that stand alone terms with the highest repulsion reflect general
manifestations of patterns.
Associations of terms in the repulsion area (high below zero) as well as associations of
terms in the attraction area (high above zero) can represent some kind of a bounded
system or a mature concept.
We can also observe consistent combinations of frequently used terms that could play
the role of linguistic and semantic regularities.
Creative Commons Licence BY - NC - SA 15
16. IDENTIFIED STRUCTURES
STRUCTURES IN REPULSION AREA BELOW ZERO
In the repulsion area below zero we can observe frequently used „stand alone“ terms that
represent how patterns are frequently named – we call them „general manifestations“ of
patterns:
something – elements - structure - form - time – set – repetition – phenomena* - way* -
sequence – relationships – dynamic - meaning - may
Creative Commons Licence BY - NC - SA 16
17. *terms “phenomena“ and „way“ in some t-SNE trials have high negative values on both
axes and in some trials - high negative values on one of the axes.
As a matter of fact patterns are frequently named by survey respondents
„something“, some whole of „elements“, „structure“, „sequence“, „form“, „set“, „repetition“,
„phenomena“, „relationships“, „way“ , „dynamic“ and some kind of assigned „meaning“.
The term „may“ is frequently used in definitions and reflects potentialities of patterns that
„may“ or „may not“ be something.
The term „time“ is not used as a name but is playing an essential role in definitions of
patterns in different contexts, as patterns seem to unfold over time.
In the repulsion area we can observe consistent combinations of terms that remained
without change through all the trials with the same setting:
(elements; sequence), (repetition; set), (dynamic; meaning)
Creative Commons Licence BY - NC - SA 17
18. STRUCTURES IN ATTRACTION AREA ABOVE ZERO
We can observe a consistent association of terms („problem“;„context“; „solution“) in
the area of a high attraction. This association represents a special perspective and a
mature concept of patterns as a tool, expressed by respondents who see pattern as a
solution to a problem in a context.
In the attraction area we can observe frequently used terms, that can be identified as
general attributes of patterns as all of them have short semantic distances and they
accompany most of definitions coupled with other terms:
Systems – Space – Similar – Different – Repeating - Common
There is also a consistent association of terms (recurring; connected; observed) that
repeats in all trials with the same setting. This association of three dimensions in
characteristics of patterns can play a role of a semantic regularity - it means, that we
should pay attention to all these aspects simultaneously when we try to detect and
describe something as a pattern.
It was interesting to compare the results of t-SNE analysis with the results of PCA
analysis, that also visualizes high dimensional data but without taking in account
non-linearity.
Creative Commons Licence BY - NC - SA 18
19. IDENTIFICATION OF LINGUISTIC AND SEMANTIC REGULARITIES WITH PCA
Our readers are welcome to experiment with different settings of the PCA (Principal
Components Analysis) by following the link below:
https://voyant-tools.org/?corpus=3f5ebfd521bae2cf37bd6da440db7dcd&stopList=keyword
s-07c7ca5299031eed316e9296dd1fde33&limit=60&view=ScatterPlot
Both t-SNE and PCA reduce the dimensionality of the data, but PCA does it without
taking in account non-linearity.
PCA uses an orthogonal transformation to convert a set of observations of possibly
correlated values into a set of values of linearly uncorrelated variables called principal
components.
In contrast to t-SNE PCA can have a „crowding problem“.
Visualization with PCA can separate terms that nonetheless semantically belong together,
as PCA renewed iterations tend to widen distances that were already detected.
SETTING FOR PCA ANALYSIS
Number of terms – 60; Number of dimensions - 3
X axis represents principal component 1
Y axis represents principal component 2
Color intensity represents principle component 3
Size of the circles represents frequency of use.
Creative Commons Licence BY - NC - SA 19
20. IDENTIFIED STRUCTURES AND INTERPRETATION OF RESULTS
We can observe some clusters:
1 2 3
Structure
Repetition
Objects
Dynamic
Arrangement
Combination
Relationships
Space
Time
Recurring
Repeating
Interaction
Observed
All identified clusters demonstrate a „pattern“ in definitions of patterns:
they associate terms reflecting some kind/form of structure, dynamic and repetition.
We can try to extract some generic combinations of characteristics of a pattern:
- „dynamic repeating structures/objects“,
- „arrangement /combination of relationships in space and/or time“
- „observed repeating/recurring interaction“
Creative Commons Licence BY - NC - SA 20
21. We see that „problem“ and „solution“, visualized by PCA, drift apart – that can be
connected with reinforcing of some distances through iterations and concealing the actual
semantic proximity.
SOME CONCLUSIONS FROM t-SNE and PCA ANALYSIS
The results from both t-SNE and PCA analysis support some propositions, that we have
already made in the first publication of the survey analysis.
We extracted general manifestations and general attributes of patterns on the basis of the
frequency of usage and proximity values, and identified some consistent associations of
terms.
There can be many ways to interpret PCA and t-SNE visualizations - we believe, that
software tools should not guide the analysis, but should support manual analysis,
heuristics and brainstorming in the starting phase of research.
SOME CONCLUSIONS ABOUT SEMANTIC REGULARITIES
Observations we made with the help of manual analysis and application of software tools
lead us to a conclusion that there are semantic regularities in definitions of patterns - most
of these definitions always reflect some kind of a complex structure coupled with dynamics
and repetition in time. There are some basic dimensions, perspectives and general
functions assigned to patterns, that we have listed in a table below.
Identified semantic regularities are coupled with repeated application of some terms and
combinations of terms.
Formalization of these repeating structures can contribute to the development of a formal
language and will be a part a future work.
In the table below we show some findings from manual analysis and t-SNE/PCA
applications.
Semantic regularities on the basis of
manual analysis
Semantic regularities on the basis of
t-SNE / PCA analysis
TIME
● regularity
● repetition
● dynamic
STRUCTURE
● relationships between parts and a
GENERAL MANIFESTATIONS
Pattern is:
- Something
- Structure
- Form
- Time
- Set
- Repetition
Creative Commons Licence BY - NC - SA 21
22. whole (set, collection,
arrangement)
● emergent wholes (systems,
organizations)
● relationships between parts
(networks, interactions, links)
● relationships in time (process,
sequence)
ORDERS OF LOGIC
- pattern is some kind of an entity
(first order)
- pattern is an order of some kinds of
entities (second order)
- pattern is an order of orders of
some kinds of entities (third order)
BASIC PHYSICAL ENTITIES
● matter
● energy
● information
PROCESS
● bottom-up processes
● top-down processes
● flat processes
COMPLEXITY/VARIETY
● ambiguity of definitions
● duality of definitions
● variety of definitions
BASIC DIMENSIONS
● dimension of space
● dimension of time
● social dimension
● structural dimension
● context
● domain
● meaning
- Elements & Relationships
- Meaning
- Phenomena
- Way
GENERAL ATTRIBUTES
- Systems
- Space
- Similar
- Different
- Repeating
- Common
- Observed
- Connected
- Recurring
ASSOCIATIONS OF TERMS
- (problem; context; solution)
- (recurring; connected; observed)
COMBINATIONS OF TERMS
- (elements; sequence)
- (repetition; set)
- (features; way)
CLUSTERS OF TERMS (PCA analysis )
- Structure, Repetition, Objects,
Dynamic
- Arrangement, Combination,
Relationships, Space, Time
- Recurring, Repeating, Interaction,
Observed
Creative Commons Licence BY - NC - SA 22
23. PERSPECTIVES
● intuited
● perceived
● conceived
● utilized
FUNCTIONS (CYBERNETIC
PERSPECTIVE)
● function of meaning making
● function of linking things together
● function of coordination
● function of predicting
● function of explanation and
understanding
● function of problem solution
SOME INSPIRATIONS
Our pattern research is research-in-progress and interpretations we make are dynamic
and ready for change.
But already now, after having worked though multiplicity of definitions, we can see that
patterns are a kind of a homonym, which is used to define multiple things.
Pulling multiplicity of perspectives and dimensions together, patterns seem to possess a
sufficient variety to absorb complexity, embrace dualities, bridge the gaps and transcend
boundaries.
If we want to detect something as complex as a pattern, then we also need something that
embraces all the regularities we have identified, and remains dynamic and flexible –
something what we can call a „ pattern of knowledge“ about patterns.
We will continue our way following the mystery of patterns. The findings from the survey
inspire our curiosity and a future research – we are moving forwards and keeping you
informed.
Creative Commons Licence BY - NC - SA 23