Co-word analyses study the co-occurrence of pairs of items (for example, keywords) that are representative in a document, to identify relations between the ideas presented in the
texts.
This PPT contain details of Z39.50 and useful for Library Science students. This protocol used for information retrieval and in the end list of different types of protocols are given.
This PPT contain details of Z39.50 and useful for Library Science students. This protocol used for information retrieval and in the end list of different types of protocols are given.
DELNET with passage of time and technological advancements not only widened its scope but has crossed the geographical boundaries. Presently, it is the major resource sharing library network in India connecting more than 5,900 libraries in 23 States and Union Territories in India and eight other countries.
The main objectives of DELNET is to promote resource sharing among the member-libraries by collecting, storing and disseminating information and by providing networked services to the researchers and scholars to supplement their research activity
Management of Library and information CentresSundar B N
in this document BLIS Paper 2 Management of Library and information Centres of KSOU 2019 August Question Paper is Solved.
Subscribe to Vision Academy YouTube Channel
https://www.youtube.com/channel/UCjzpit_cXjdnzER_165mIiw
A digital agenda for library automation & networking. e-Granthalaya is a library automation software from National Informatics Centre, Department of Electronics & Information Technology, Ministry of Communications and Information Technology, Government of India. The software has been designed by a team of experts from software as well as Library and Information Science discipline. Using this software the libraries can automate in-house activities as well as user services. The software can be implemented either in stand-alone or in client-server mode where database and WebOPAC are installed on the server PC while the data entry program is installed on client PCs. The software provides LOCAL/LAN/WAN based data entry solutions for a cluster of libraries where a centralized database can be created with Union Catalog output. The software provides Web OPAC interface to publish the library catalog over Internet/Intranet. The software runs on Windows paltform Only, UNICODE Compliant ,thus, supports data entry in local language.
ASSESSING SIMILARITY BETWEEN ONTOLOGIES: THE CASE OF THE CONCEPTUAL SIMILARITYIJwest
In ontology engineering, there are many cases where assessing similarity between ontologies is required, this is the case of the alignment activities, ontology evolutions, ontology similarities, etc. This paper presents a new method for assessing similarity between concepts of ontologies. The method is based on the
set theory, edges and feature similarity. We first determine the set of concepts that is shared by two ontologies and the sets of concepts that are different from them. Then, we evaluate the average value of similarity for each set by using edges-based semantic similarity. Finally, we compute similarity between
ontologies by using average values of each set and by using feature-based similarity measure too.
ASSESSING SIMILARITY BETWEEN ONTOLOGIES: THE CASE OF THE CONCEPTUAL SIMILARITYdannyijwest
In ontology engineering, there are many cases where assessing similarity between ontologies is required, this is the case of the alignment activities, ontology evolutions, ontology similarities, etc. This paper presents a new method for assessing similarity between concepts of ontologies. The method is based on the set theory, edges and feature similarity. We first determine the set of concepts that is shared by two ontologies and the sets of concepts that are different from them. Then, we evaluate the average value of similarity for each set by using edges-based semantic similarity. Finally, we compute similarity between ontologies by using average values of each set and by using feature-based similarity measure too.
DELNET with passage of time and technological advancements not only widened its scope but has crossed the geographical boundaries. Presently, it is the major resource sharing library network in India connecting more than 5,900 libraries in 23 States and Union Territories in India and eight other countries.
The main objectives of DELNET is to promote resource sharing among the member-libraries by collecting, storing and disseminating information and by providing networked services to the researchers and scholars to supplement their research activity
Management of Library and information CentresSundar B N
in this document BLIS Paper 2 Management of Library and information Centres of KSOU 2019 August Question Paper is Solved.
Subscribe to Vision Academy YouTube Channel
https://www.youtube.com/channel/UCjzpit_cXjdnzER_165mIiw
A digital agenda for library automation & networking. e-Granthalaya is a library automation software from National Informatics Centre, Department of Electronics & Information Technology, Ministry of Communications and Information Technology, Government of India. The software has been designed by a team of experts from software as well as Library and Information Science discipline. Using this software the libraries can automate in-house activities as well as user services. The software can be implemented either in stand-alone or in client-server mode where database and WebOPAC are installed on the server PC while the data entry program is installed on client PCs. The software provides LOCAL/LAN/WAN based data entry solutions for a cluster of libraries where a centralized database can be created with Union Catalog output. The software provides Web OPAC interface to publish the library catalog over Internet/Intranet. The software runs on Windows paltform Only, UNICODE Compliant ,thus, supports data entry in local language.
ASSESSING SIMILARITY BETWEEN ONTOLOGIES: THE CASE OF THE CONCEPTUAL SIMILARITYIJwest
In ontology engineering, there are many cases where assessing similarity between ontologies is required, this is the case of the alignment activities, ontology evolutions, ontology similarities, etc. This paper presents a new method for assessing similarity between concepts of ontologies. The method is based on the
set theory, edges and feature similarity. We first determine the set of concepts that is shared by two ontologies and the sets of concepts that are different from them. Then, we evaluate the average value of similarity for each set by using edges-based semantic similarity. Finally, we compute similarity between
ontologies by using average values of each set and by using feature-based similarity measure too.
ASSESSING SIMILARITY BETWEEN ONTOLOGIES: THE CASE OF THE CONCEPTUAL SIMILARITYdannyijwest
In ontology engineering, there are many cases where assessing similarity between ontologies is required, this is the case of the alignment activities, ontology evolutions, ontology similarities, etc. This paper presents a new method for assessing similarity between concepts of ontologies. The method is based on the set theory, edges and feature similarity. We first determine the set of concepts that is shared by two ontologies and the sets of concepts that are different from them. Then, we evaluate the average value of similarity for each set by using edges-based semantic similarity. Finally, we compute similarity between ontologies by using average values of each set and by using feature-based similarity measure too.
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityVladimir Kulyukin
Automated information retrieval relies heavily on statistical regularities that emerge as terms are deposited to produce text. This paper examines statistical patterns expected of a pair of
terms that are semantically related to each other. Guided by a conceptualization of the text generation process, we derive measures of how tightly two terms are semantically associated.
Our main objective is to probe whether such measures yields reasonable results. Specifically, we examine how the tendency of a content bearing term to clump, as quantified by previously
developed measures of term clumping, is influenced by the presence of other terms. This
approach allows us to present a toolkit from which a range of measures can be constructed.
As an illustration, one of several suggested measures is evaluated on a large text corpus built from an on-line encyclopedia.
ONTOLOGICAL MODEL FOR CHARACTER RECOGNITION BASED ON SPATIAL RELATIONSsipij
In this paper, we present a set of spatial relations between concepts describing an ontological model for a
new process of character recognition. Our main idea is based on the construction of the domain ontology
modelling the Latin script. This ontology is composed by a set of concepts and a set of relations. The
concepts represent the graphemes extracted by segmenting the manipulated document and the relations are
of two types, is-a relations and spatial relations. In this paper we are interested by description of second
type of relations and their implementation by java code.
Interpreting the Semantics of Anomalies Based on Mutual Information in Link M...ijdms
This paper aims to show how mutual information can help provide a semantic interpretation of anomalies in data, characterize the anomalies, and how mutual information can help measure the information that object item X shares with another object item Y. Whilst most link mining approaches focus on predicting link type, link based object classification or object identification, this research focused on using link mining to detect anomalies and discovering links/objects among anomalies. This paper attempts to demonstrate the contribution of mutual information to interpret anomalies using a case study.
Supporting scientific discovery through linkages of literature and dataDon Pellegrino
Poster on "Supporting scientific discovery through linkages of literature and data" by Don Pellegrino, Jean-Claude Bradley and Chaomei Chen. To be presented on May 3, 2011 as part of an event for the Center for Visual and Decision Informatics. More this NSF funded center can be found online at [http://nsf-cvdi.louisiana.edu/]. More on Don Pellegrino's research activities can be found online at [http://www.donpellegrino.com].
International Journal of Engineering Research and Development (IJERD)IJERD Editor
International Journal of Engineering Research and Development is an international premier peer reviewed open access engineering and technology journal promoting the discovery, innovation, advancement and dissemination of basic and transitional knowledge in engineering, technology and related disciplines.
Towards a Relational Paradigm in Sustainability Research, Practice, and Educa...Zack Walsh
Relational thinking has recently gained increasing prominence across academic disciplines in an attempt to understand complex phenomena in terms of constitutive processes and relations. Interdisciplinary fields of study, such as science and technology studies (STS), the environmental humanities, and the posthumanities, for example, have started to reformulate academic understanding of nature-cultures based on relational thinking. Although the sustainability crisis serves as a contemporary backdrop and in fact calls for such innovative forms of interdisciplinary scholarship, the field of sustainability research has not yet tapped into the rich possibilities offered by relational thinking. Against this background, the purpose of this paper is to identify relational approaches to ontology, epistemology, and ethics which are relevant to sustainability research. More specifically, we analyze how relational approaches have been understood and conceptualized across a broad range of disciplines and contexts relevant to sustainability to identify and harness connections and contributions for future sustainability-related work. Our results highlight common themes and patterns across relational approaches, helping to identify and characterize a relational paradigm within sustainability research. On this basis, we conclude with a call to action for sustainability researchers to co-develop a research agenda for advancing this relational paradigm within sustainability research, practice, and education.
SEMANTIC INTEGRATION FOR AUTOMATIC ONTOLOGY MAPPING cscpconf
In the last decade, ontologies have played a key technology role for information sharing and agents interoperability in different application domains. In semantic web domain, ontologies are efficiently used toface the great challenge of representing the semantics of data, in order to bring the actual web to its full
power and hence, achieve its objective. However, using ontologies as common and shared vocabularies requires a certain degree of interoperability between them. To confront this requirement, mapping ontologies is a solution that is not to be avoided. In deed, ontology mapping build a meta layer that allows different applications and information systems to access and share their informations, of course, after resolving the different forms of syntactic, semantic and lexical mismatches. In the contribution presented in this paper, we have integrated the semantic aspect based on an external lexical resource, wordNet, to design a new algorithm for fully automatic ontology mapping. This fully automatic character features the
main difference of our contribution with regards to the most of the existing semi-automatic algorithms of ontology mapping, such as Chimaera, Prompt, Onion, Glue, etc. To better enhance the performances of our algorithm, the mapping discovery stage is based on the combination of two sub-modules. The former
analysis the concept’s names and the later analysis their properties. Each one of these two sub-modules is
it self based on the combination of lexical and semantic similarity measures.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
1. NAME:- DEBOLINA BISWAS
R O L L N O : - 0 0 2 0 0 0 8 0 2 0 1 0
S E S S I O N : 2 0 2 0 - 2 0 2 1
PAPER: I N F O R M E T R I C S . ( M L - 1 1 )
CO-WORD ANALYSIS
2. Introduction
Since World War II, the scope and volume of scientific research have increased dramatically. This is well
reflected in the growth of the literature. Many decades later, along with developments in information
technology, especially in the area of data storage, the amount of information in the world is overwhelming.
In formulating and executing broad spectrum research policy, it is important to understand how
research thrusts have interrelated and evolved over time, how they are projected to evolve, and
how different types of interventions from sponsors and policymakers can affect the evolution and
impact of research.
The traditional way to map the relationships among concepts, ideas, and problems in science is to
seek the views of a relatively small number of experts. But identification of the connectivity of a
broad range of areas is well beyond the expertise of any panel of experts.
Much recent effort has been focused on development of more objective quantitative approaches for
analyzing and integrating written and survey information to supplement analysts or groups of
peers in understanding research trends ; they include co-citation analysis, co-nomination analysis,
and co-word analysis.
So, we can see that modern quantitative techniques utilise computer technologies extensively. One
of techniques exploits the use of co-occurrence phenomena, and it is the method of co-word
analysis.
3. What is Co-Word Analysis?
Co-word analyses study the co-occurrence of pairs of items (for example, keywords) that
are representative in a document, to identify relations between the ideas presented in the
texts (He, 1999).
Co-word analysis is related to co-citation analysis (Small, 1973; Small & Griffith, 1974).
Co-citation analysis provides a method of mapping the structure of a research field
through pairs of documents jointly cited. But co-word analysis deals directly with sets of
terms shared by documents instead of shared citations. Therefore, it maps the pertinent
literature directly from the interaction of key terms instead of the interaction of citations.
So, in co-word analysis, it is assumed that keywords extracted from papers could
represent a specific research direction, research topic or subject of a field. If two
keywords co-occur within one paper, the two research topics they represent are related.
Higher co-word frequency means stronger correlation in keywords pairs, which can
further suggest that two keywords are related to a specific research topic (Cambrosio et
al. 1993).
Co-word analysis has the potential of effectively revealing patterns and trends in a
specific discipline (Ding et al. 2001). Co-word analysis has been utilized to reveal
research advances in many fields.
4. Origins of Co-word Analysis
The origins of co-word phenomena can be traced back at least five decades to the pioneering
work in: 1) lexicography of Homby (1942) to account for co-occurrence knowledge, and 2)
linguistics of De Saussure (1949) to describe how affinity of two language units correlates with
their appearance in the language. (Kostoff, n.d)
But the technique of co-word analysis was first developed in collaboration between the Centre
de Sociologie de I’Innovation of the Ecole Nationale Superieure des Mines of Paris and the
CNRS (Centre National de la Re-cherche Scientifique) of France during the 198Os, and their
system was called “LEXIMAPPE.”
This method, originally dedicated to scientific fields, measures the association strength
between terms in documents to reveal and visualize the evolution of scientific fields through
the construction of clusters and strategic diagram.
For about twenty years, this technique has been employed to map the dynamic development of
several research fields. One of the early studies was carried out by Serge Bauin (1986) to map
the dynamics of aquaculture from 1979 to 1981.
5. Theoretical Foundation of Co-word
Analysis-Actor Network
The co-word analysis technique was first proposed to map the dynamics of science. The most
feasible way to understand the dynamics of science is to take the force of science in present-
day societies into account.
“Actor network is the theoretical foundation for co-word analysis to map the dynamics of
science”. (Callon, 1986)
Even though science cannot be reduced to texts only, texts are still a prime source for studies
on how worlds are created and transformed in the laboratory. Therefore, instead of
following the actors to see how they change the world, following the texts is another way to
map the dynamics of science.
Based on the co-occurrence of pairs of words, co-word analysis seeks to extract the themes of
science and detect the linkages among these themes directly from the subject content of
texts. (Callon et al., 1986b).
Overall, co-word analysis considers the dynamics of science as a result of actor strategies.
(Callon et al., 1991).
6. Co-word analysis method
Co-word analysis is a content analysis technique that uses patterns of co-occurrence of pairs
of items (i.e., words or noun phrases) in a corpus of texts to identify the relationships
between ideas within the subject areas as presented in these texts.
Indexes based on the co-occurrence frequency of items, such as an inclusion index and a
proximity index, are used to measure the strength of relationships between items. Based on
these indexes, items are clustered into groups and displayed in network maps.
For example, an inclusion map is used to highlight the central themes in a domain, and a
proximity map is used to reveal the connections between minor areas hidden behind the
central ones. Some other indexes, such as those based on density and centrality, are
employed to evaluate the shape of each map, showing the degree to which each area is
centrally structured and the extent to which each area is central to the others. By comparing
the network maps for different time periods, the scientific dynamic can be detected.
7. Metrics for co-word analysis
Metrics for co-word analysis have been studied extensively (Grivel & François, 1995). Two
terms, i and j, co-occur if they are used together in a single document. Take a corpus consisting
of N documents. Each document is indexed by a set of unique terms that can occur in multiple
documents. Let Ck be the number of occurrences of term k; i.e., the number of times k is used
for indexing documents in the corpus. Let Cij be the number of co-occurrences of terms i and j
(the number of documents indexed by both terms).
In later co-word studies a index is employed to calculate the association values between word
pairs. This coefficient is called the equivalence index (e-coefficient) (Callon et al., 1991) or
strength (Coulter et al., 1998). The basic metric used for this study is the Association Indice
Eij. The strength of association between terms i and j is given by
the expression:
Two terms that appear many times in isolation but only a few times together will yield a lower E
value than two terms that appear relatively less often alone but have a higher ratio of co-
occurrences. Terms with relatively high E values form the networks’ links. A term network
consists of nodes (terms) connected by links. Each node must be linked to at least one other
node in a network.
8. Continued
This metric provides an intuitive measure of the strength of association between terms,
and only indicates that there is some semantic relationship. This metric is easier to
understand and utilize in the production and interpretation of term association maps. It
allows associations of both major and minor terms and is symmetrical in their
relationships (Callon, Courtial, & Turner, 1991). E can be used as the basis for devising
several complementary measures of term interactions and term networks in a unified
manner.
The co-word algorithm then proceeds in two steps to produce the paired connection of
terms. The first step builds networks that can identify areas of strong focus. The second
step can identify terms that associate in more than one network and thereby indicate
overlapping issues.
Without some minimum constraints, terms that appear infrequently but almost always
together could dominate clusters; hence a minimum co-occurrence Cij value is required to
generate a link. At the same time, some maps can become cluttered due to an excessive
number of legitimate links (generally of decreasing E values); hence, restrictions on
numbers of nodes and sometimes links are required to help discover the major partitions
of concepts.
9. Inclusion Index
In the early studies of co- word analysis, two indexes were introduced: 1. Inclusion
Index and 2. Proximity Index.
The hierarchies of subject areas in a research problem can be detected by calculating an
index, called the inclusion index. The Inclusion Index offers access to a tangible metric
that can help identify gaps in our works and understand what needs to be addressed.
Inclusion Index (Iij):
where,
Cij the number of documents in which the keyword pair (M, and M,) appears;
Ci is the occurrence frequency of keyword Mi in the set of articles;
Cj is the occurrence frequency of keyword M. in the set of articles;
min (Ci,Cj) is the minimum of the two freqdencies Ci,and Cj.
Iij has a value between 0 and 1, and it can be interpreted as a conditional probability.
10. Proximity Index
To detect the minor but potentially growing areas we need the proximity index.
If a situation occurs that implies that there are some mediator keywords, which
have a relatively low occurrence frequency but still have significant
relationships with some of the peripheral key-words. To bring out such
patterns,
a proximity index Pij is defined:
where,
Ci, Cj, and Cij have the same meaning as in formula of inclusion index.
N is the number of articles in the collection.
The mediator and peripheral keywords pulled out by Pij represent minor but
potentially growing areas.
11. Inclusion Map
After the inclusion and proximity indexes are calculated, inclusion and
proximity maps are created.
The inclusion maps are designed to discover the central themes in a domain and
depict their relationship to keywords that occur less frequently. The inclusion map
involves getting more information about a certain theme.
To create inclusion maps, the link that has the highest inclusion index value is
selected first. These linked nodes become the starting points for the first inclusion
map (subnetwork).
Other links and their corresponding nodes are then added into the map in the
decreasing order of their inclusion index until the threshold I0 is reached.
Other links and their corresponding nodes are then added into the map in the
decreasing order of their inclusion index until the threshold I, is reached.
Keywords that appear on the top level of inclusion maps are called “central poles” of
the domain of research. Keywords that are included in the central poles, and
themselves include some other words at lower levels, are called “mediator words”
(Callon et al., 1986).
12. Proximity Map
The proximity maps are designed to discover connections between minor
ideas hidden behind the central themes. The proximity maps concerns the
analysis of the links between themes.
The process to create proximity maps is similar to that for inclusion maps. The
difference is that the proximity index is used instead of the inclusion index.
If the threshold P, is lowered enough, more proximity connections between
keywords will appear in the map and, eventually, the mediators and central
poles found in inclusion maps will reappear. In this way, the relationship
between minor issues and central poles can be studied (Callon et al., 1986).
13. Density and Centrality
Density: Density is used to measure the strength of the links that tie together the
words making up the cluster; that is the internal strength of a cluster. It provides a good
representation of the cluster’s capacity to maintain itself and to develop over the course
of time in the field under consideration (Callon et al., 1991).
The value of the density of a given cluster can be measured in several ways. Generally, the
index value for links between each word pair is calculated first. Then, the density value
can be the average value (mean) of internal links. the median value of internal links or
the sum of the squares of the value of internal links.
Centrality: Centrality is used to measure the strength of a subject area’s interaction
with other subject areas. Ranking subject areas (clusters) with respect to their centrality
shows the extent to which each area is central within a global research network. The
greater the number and strength of a subject area’s connections with other subject areas,
the more central this subject area will be in the research network (Bauin et al., 1991).
For a given cluster (area), its centrality can be the sum of all external link values or the
square root of the sum of the squares of all external link values (e.g., Coulter et al., 1998).
An external link is a link that goes from a word belonging to a cluster to a word external
to the cluster.
14. Strategic diagram
A strategic diagram that offers a global representation of the structure of
any field or subfield can be created by plotting centrality and density into a
two-dimensional diagram (Law et al., 1988).
Typically, the horizontal axis represents centrality, the vertical axis
represents density, and the origin of the graph is at the median of the
respective axis values. This map situates each subject area within a two-
dimensional space divided into four quadrants.
The strategic diagram is used in many co-word analysis studies and the
analysis based on it is similar among these studies.
15. Strategic diagram
Quadrant 3 : Peripheral
and developed Words
Quadrant 1: Central and
developed Words
Quadrant 4: Peripheral
and undeveloped Words
Quadrant 2: Central and
undeveloped Words
Density
Centrality
16. Thematic Network
In a theme, the keywords and their interconnections draw a network graph, called a
thematic network. Each thematic network is labeled using the name of the most
significant keyword in the associated theme (usually identified by the most central
keyword of the theme).
In this figure, several keywords are
interconnected, where the volume of the
spheres is proportional to the number of
Documents corresponding to each keyword,
the thickness of the link between two
spheres i and j is proportional to the
equivalence index Iij .
Network is built, based on the documents linked to each thematic network. A
document is linked to a theme if it has at least two keywords presented in the
thematic network.
17. Dynamics of Networks
A striking feature of some strategic diagrams is the radical change in
the configuration of the network at two periods. This reflects the
dynamics of science. Based on the strategic diagram, we can analyze the
stability of the networks and foresee their changes in the future. This
issue is addressed in many studies, and the methods used in these
studies fall into two categories: The study of strategic diagrams, and the
ratio of centrality to density.
The comparison of two networks, N1 and N2, which might be two net-
works at different times or two distinct networks at the same time, can
be done by the method network comparison.
18. Database Tomography
Database Tomography is a patented system for analyzing large amounts of textual
computerized material (Kostoff et al., 1995).
It can be considered as another generation of co-word analysis. Algorithms for extracting
multi-word phrase frequencies and performing phrase proximity analysis are included in
this system. Phrase frequency analysis can be used to discover the pervasive themes of a
database while the phrase proximity analysis can be used to detect the relationships
among these themes and between these themes and subthemes (Kostoff et al., 1997a).
Similar to co-word analysis, Database Tomography can identify the main intellectual
thrust areas and the relationship among these thrust areas. It provides a comprehensive
overview of a research network and allows specific starting points to be chosen rationally
for more detailed investigations into a topic of interest.
Based on the term co-occurrence information, Database Tomography can also be used to
expand the initial query in information retrieval (IR) systems and, in turn, allow the
retrieval of relevant documents that would not have been retrieved with the initial query
(Kostoff et al., 1997b).
19. Advantages of Co-Word Analysis
Quantitative Over Qualitative: The advantages of co-word analysis over
qualitative analysis were recognized by researchers. Using the quantitative in pursuit of
the qualitative, researchers are also able to highlight features of scientific fields that have
not always been recognized. The heterogeneity of scientific world-building is preserved in
co-word analysis, where experimental findings, research methods, concepts, social
problems, material artifacts and locations may appear together on the maps.
Flexibility: Compared with other methods of analysis that focus on texts, co-word
analysis is much more flexible in that it shows the research net- work with graphs. On the
one hand, these graphs can be simplified to show the overall structure of the network. On
the other hand, one can zoom in on certain areas and trace the co-word patterns in as
much detail as one wishes.
20. Limitations of Co-word Analysis
Indexer effect: Solely making use of keywords and index words was the biggest
problem of early co- word analysis. It was called “indexer effect” and was addressed by
many researchers.
The representations of the results given by co-word analysis are too difficult to read.
(Whittaker, 1989)
The coverage of database is incomplete. Certain types of literature, such as patents arid
the “grey literature,” lie outside of the publication circuit and are not indexed in the
database. This means that the results from co-word analysis cannot reflect the whole
picture of the research field in question.
The delay between the writing of a document and the moment when it is indexed and
entered into a database causes the co-word analysis to fail to detect emerging research
themes at an early stage. (Callon et al., 1986)
21. Some key Issues in Co-word Analysis
The maps obtained by co-word analysis are generally considered very
difficult to understand in isolation. They have to be interpreted with
caution. It is suggested that the interpretation must be active and based
on the comparison of maps (Callon and al. 1986).
The choice of the words (keywords, descriptors) is another issue to be
carefully considered and has led to many discussions (Leydesdorff,
1997, Whittaker, 1989). The "normalization" of words must be
considered as well .
22. Assumptions of Co-word Analysis
The assumptions of co-word analysis are presented by Whittaker (1989) as follows:
“It [co-word analysis] relies upon the argument that (i) authors of scientific articles choose
their technical terms carefully; (ii) when different terms are used in the same articles it is
therefore because the author is either recognizing or postulating some non-trivial
relationship between their referents; and (iii) if enough different authors appear to recognize
the same relationship, then that relation- ship may be assumed to have some significance
within the area of science concerned.”
Later, Law and Whittaker (1992) have restated two of the assumptions. First. co-word
analysis assumes that the keywords used by indexers to index a paper reflect the present
stages of the scientific re- search in question. Then, co-word analysis assumes that
arguments received by other scientists will lead to the publication of further scientific
papers that are indexed by similar sets of keywords.
23. Applications of Co-word analysis
Co-word analysis has been used to study conceptual work in different domains
by many researchers. full text articles included in the journal Scientrometrics to
mapping the intellectual structure of scientrometrics using text-mining and co-
word analysis.
Co-word analysis is the bibliometric technique used to identify the main
themes. All this allows us to quantify and visualize the thematic evolution of
many research fields. It also helps to both experts and novices to understand
the current state of the field and to predict where future research could lead.
24. Conclusion
The appearance of co-word analysis added another choice in the area of
bibliometrics and provided another way to discover knowledge in databases.
Similar to co-citation analysis, co-word analysis is also a kind of relational study
based on the idea that publications should not be considered as discrete units.
Instead, each is built upon others.
Previous studies have shown co-word analysis to be a powerful tool to discover
knowledge in databases. It has been used to detect the themes in a given research
area, the relationship between these themes, the extent to which these themes are
central to the whole area, and the degree to which these themes are internally
structured. In the last twenty years, co-word analysis has been improved in many
aspects. The main progress ran be found in two fields: 1. Source of words and 2.
Measurements.
So lastly we can say that co-word analysis is the one of the most common methods
used for constructing the thematic and strategic map of a field.
25. References
Bauin, S. (1986). Aquaculture: A field by bureaucratic fiat. In M. Callon, J. Law, & A. Rip
(Eds.),Mapping the dynamics of sriencr and technology: Sociology ofscience in the real
world (pp. 124-141). London: The Macmillan Press Ltd.
Chllon, M. (1986). Pinpointing industrial invention: An exploration of quantitative meth-
ods for the analysis of patents. In M. Callon, J. Law, & A. Rip (Eds.), Mapping thr
dynamics of science and techno1 ociology of science in the real world (pp. 163-188). Lon-
don: The Macmillan Press Ltd.
Callon, M.; Courtid1,J-P.; & Ladle, F. (1991). Co-word analysis as a tool for describing the
nrtwork of interactions between basic and technological research: The case of poly- mer
chemistry. Scienlomrtrirs, 22(1).
He, Q., 1999, Knowledge Discovery Through Co-Word Analysis, Library Trends, 48(1).
Whittaker, J. (1989). Creativity and conformity in science: Titles, keywords and co-word
analysis. Social Studies of Science, 19, 473-496.