The document describes the design and implementation of a meetings document management and retrieval system. Key features of the system include:
1. Capturing, storing, indexing, and retrieving meeting documents such as agendas, minutes, and registration forms from a database.
2. Implementing a search facility to allow users to quickly locate topics of interest within documents.
3. Incorporating hyperlinks to enable navigation between related documents and sections.
4. Developing the system as a web application using ASP.NET to allow remote access by authorized users.
The system was designed using object-oriented principles and includes security features to protect unauthorized access to documents. It aims to improve organization and access of meeting
The document discusses the components and design of information storage and retrieval systems (ISRS). It describes ISRS as having three main components: the user interface, knowledge base, and search agent. The user interface allows users to input queries and view results, and should be intuitive. The knowledge base stores the information to be retrieved in a database. And the search agent acts to translate user queries and match them to the knowledge base to retrieve relevant information. The document provides details on each of these components and discusses best practices for designing an effective ISRS.
Classification-based Retrieval Methods to Enhance Information Discovery on th...IJMIT JOURNAL
The widespread adoption of the World-Wide Web (the Web) has created challenges both for society as a whole and for the technology used to build and maintain the Web. The ongoing struggle of information retrieval systems is to wade through this vast pile of data and satisfy users by presenting them with information that most adequately it’s their needs. On a societal level, the Web is expanding faster than we can comprehend its implications or develop rules for its use. The ubiquitous use of the Web has raised important social concerns in the areas of privacy, censorship, and access to information. On a technical level, the novelty of the Web and the pace of its growth have created challenges not only in the development of new applications that realize the power of the Web, but also in the technology needed to scale applications to accommodate the resulting large data sets and heavy loads. This thesis presents searching algorithms and hierarchical classification techniques for increasing a search service's understanding of web queries. Existing search services rely solely on a query's occurrence in the document collection to locate relevant documents. They typically do not perform any task or topic-based analysis of queries using other available resources, and do not leverage changes in user query patterns over time. Provided within are a set of techniques and metrics for performing temporal analysis on query logs. Our log analyses are shown to be reasonable and informative, and can be used to detect changing trends and patterns in the query stream, thus providing valuable data to a search service.
Intelligent Semantic Web Search Engines: A Brief Survey dannyijwest
The World Wide Web (WWW) allows the people to share the information (data) from the large database repositories globally. The amount of information grows billions of databases. We need to search the information will specialize tools known generically search engine. There are many of search engines available today, retrieving meaningful information is difficult. However to overcome this problem in search engines to retrieve meaningful information intelligently, semantic web technologies are playing a major role. In this paper we present survey on the search engine generations and the role of search engines in intelligent web and semantic search technologies.
NATURE: A TOOL RESULTING FROM THE UNION OF ARTIFICIAL INTELLIGENCE AND NATURA...ijaia
This paper presents the final results of the research project that aimed for the construction of a tool which
is aided by Artificial Intelligence through an Ontology with a model trained with Machine Learning, and is
aided by Natural Language Processing to support the semantic search of research projects of the Research
System of the University of Nariño. For the construction of NATURE, as this tool is called, a methodology
was used that includes the following stages: appropriation of knowledge, installation and configuration of
tools, libraries and technologies, collection, extraction and preparation of research projects, design and
development of the tool. The main results of the work were three: a) the complete construction of the
Ontology with classes, object properties (predicates), data properties (attributes) and individuals
(instances) in Protegé, SPARQL queries with Apache Jena Fuseki and the respective coding with
Owlready2 using Jupyter Notebook with Python within the virtual environment of anaconda; b) the
successful training of the model for which Machine Learning algorithms were used and specifically
Natural Language Processing algorithms such as: SpaCy, NLTK, Word2vec and Doc2vec, this was also
performed in Jupyter Notebook with Python within the virtual environment of anaconda and with
Elasticsearch; and c) the creation of NATURE by managing and unifying the queries for the Ontology and
for the Machine Learning model. The tests showed that NATURE was successful in all the searches that
were performed as its results were satisfactory
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Context Driven Technique for Document ClassificationIDES Editor
In this paper we present an innovative hybrid Text
Classification (TC) system that bridges the gap between
statistical and context based techniques. Our algorithm
harnesses contextual information at two stages. First it extracts
a cohesive set of keywords for each category by using lexical
references, implicit context as derived from LSA and wordvicinity
driven semantics. And secondly, each document is
represented by a set of context rich features whose values are
derived by considering both lexical cohesion as well as the extent
of coverage of salient concepts via lexical chaining. After
keywords are extracted, a subset of the input documents is
apportioned as training set. Its members are assigned categories
based on their keyword representation. These labeled
documents are used to train binary SVM classifiers, one for
each category. The remaining documents are supplied to the
trained classifiers in the form of their context-enhanced feature
vectors. Each document is finally ascribed its appropriate
category by an SVM classifier.
Comparative Study on Graph-based Information Retrieval: the Case of XML DocumentIJAEMSJORNAL
The processing of massive amounts of data has become indispensable especially with the potential proliferation of big data. The volume of information available nowadays makes it difficult for the user to find relevant information in a vast collection of documents. As a result, the exploitation of vast document collections necessitates the implementation of automated technologies that enable appropriate and effective retrieval. In this paper, we will examine the state of the art of IR in XML documents. We will also discuss some works that have used graphs to represent documents in the context of IR. In the same vein, the relationships between the components of a graph are the center of our attention.
The huge volume of text documents available on the internet has made it difficult to find valuable
information for specific users. In fact, the need for efficient applications to extract interested knowledge
from textual documents is vitally important. This paper addresses the problem of responding to user
queries by fetching the most relevant documents from a clustered set of documents. For this purpose, a
cluster-based information retrieval framework was proposed in this paper, in order to design and develop
a system for analysing and extracting useful patterns from text documents. In this approach, a pre-
processing step is first performed to find frequent and high-utility patterns in the data set. Then a Vector
Space Model (VSM) is performed to represent the dataset. The system was implemented through two main
phases. In phase 1, the clustering analysis process is designed and implemented to group documents into
several clusters, while in phase 2, an information retrieval process was implemented to rank clusters
according to the user queries in order to retrieve the relevant documents from specific clusters deemed
relevant to the query. Then the results are evaluated according to evaluation criteria. Recall and Precision
(P@5, P@10) of the retrieved results. P@5 was 0.660 and P@10 was 0.655.
The document discusses the components and design of information storage and retrieval systems (ISRS). It describes ISRS as having three main components: the user interface, knowledge base, and search agent. The user interface allows users to input queries and view results, and should be intuitive. The knowledge base stores the information to be retrieved in a database. And the search agent acts to translate user queries and match them to the knowledge base to retrieve relevant information. The document provides details on each of these components and discusses best practices for designing an effective ISRS.
Classification-based Retrieval Methods to Enhance Information Discovery on th...IJMIT JOURNAL
The widespread adoption of the World-Wide Web (the Web) has created challenges both for society as a whole and for the technology used to build and maintain the Web. The ongoing struggle of information retrieval systems is to wade through this vast pile of data and satisfy users by presenting them with information that most adequately it’s their needs. On a societal level, the Web is expanding faster than we can comprehend its implications or develop rules for its use. The ubiquitous use of the Web has raised important social concerns in the areas of privacy, censorship, and access to information. On a technical level, the novelty of the Web and the pace of its growth have created challenges not only in the development of new applications that realize the power of the Web, but also in the technology needed to scale applications to accommodate the resulting large data sets and heavy loads. This thesis presents searching algorithms and hierarchical classification techniques for increasing a search service's understanding of web queries. Existing search services rely solely on a query's occurrence in the document collection to locate relevant documents. They typically do not perform any task or topic-based analysis of queries using other available resources, and do not leverage changes in user query patterns over time. Provided within are a set of techniques and metrics for performing temporal analysis on query logs. Our log analyses are shown to be reasonable and informative, and can be used to detect changing trends and patterns in the query stream, thus providing valuable data to a search service.
Intelligent Semantic Web Search Engines: A Brief Survey dannyijwest
The World Wide Web (WWW) allows the people to share the information (data) from the large database repositories globally. The amount of information grows billions of databases. We need to search the information will specialize tools known generically search engine. There are many of search engines available today, retrieving meaningful information is difficult. However to overcome this problem in search engines to retrieve meaningful information intelligently, semantic web technologies are playing a major role. In this paper we present survey on the search engine generations and the role of search engines in intelligent web and semantic search technologies.
NATURE: A TOOL RESULTING FROM THE UNION OF ARTIFICIAL INTELLIGENCE AND NATURA...ijaia
This paper presents the final results of the research project that aimed for the construction of a tool which
is aided by Artificial Intelligence through an Ontology with a model trained with Machine Learning, and is
aided by Natural Language Processing to support the semantic search of research projects of the Research
System of the University of Nariño. For the construction of NATURE, as this tool is called, a methodology
was used that includes the following stages: appropriation of knowledge, installation and configuration of
tools, libraries and technologies, collection, extraction and preparation of research projects, design and
development of the tool. The main results of the work were three: a) the complete construction of the
Ontology with classes, object properties (predicates), data properties (attributes) and individuals
(instances) in Protegé, SPARQL queries with Apache Jena Fuseki and the respective coding with
Owlready2 using Jupyter Notebook with Python within the virtual environment of anaconda; b) the
successful training of the model for which Machine Learning algorithms were used and specifically
Natural Language Processing algorithms such as: SpaCy, NLTK, Word2vec and Doc2vec, this was also
performed in Jupyter Notebook with Python within the virtual environment of anaconda and with
Elasticsearch; and c) the creation of NATURE by managing and unifying the queries for the Ontology and
for the Machine Learning model. The tests showed that NATURE was successful in all the searches that
were performed as its results were satisfactory
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Context Driven Technique for Document ClassificationIDES Editor
In this paper we present an innovative hybrid Text
Classification (TC) system that bridges the gap between
statistical and context based techniques. Our algorithm
harnesses contextual information at two stages. First it extracts
a cohesive set of keywords for each category by using lexical
references, implicit context as derived from LSA and wordvicinity
driven semantics. And secondly, each document is
represented by a set of context rich features whose values are
derived by considering both lexical cohesion as well as the extent
of coverage of salient concepts via lexical chaining. After
keywords are extracted, a subset of the input documents is
apportioned as training set. Its members are assigned categories
based on their keyword representation. These labeled
documents are used to train binary SVM classifiers, one for
each category. The remaining documents are supplied to the
trained classifiers in the form of their context-enhanced feature
vectors. Each document is finally ascribed its appropriate
category by an SVM classifier.
Comparative Study on Graph-based Information Retrieval: the Case of XML DocumentIJAEMSJORNAL
The processing of massive amounts of data has become indispensable especially with the potential proliferation of big data. The volume of information available nowadays makes it difficult for the user to find relevant information in a vast collection of documents. As a result, the exploitation of vast document collections necessitates the implementation of automated technologies that enable appropriate and effective retrieval. In this paper, we will examine the state of the art of IR in XML documents. We will also discuss some works that have used graphs to represent documents in the context of IR. In the same vein, the relationships between the components of a graph are the center of our attention.
The huge volume of text documents available on the internet has made it difficult to find valuable
information for specific users. In fact, the need for efficient applications to extract interested knowledge
from textual documents is vitally important. This paper addresses the problem of responding to user
queries by fetching the most relevant documents from a clustered set of documents. For this purpose, a
cluster-based information retrieval framework was proposed in this paper, in order to design and develop
a system for analysing and extracting useful patterns from text documents. In this approach, a pre-
processing step is first performed to find frequent and high-utility patterns in the data set. Then a Vector
Space Model (VSM) is performed to represent the dataset. The system was implemented through two main
phases. In phase 1, the clustering analysis process is designed and implemented to group documents into
several clusters, while in phase 2, an information retrieval process was implemented to rank clusters
according to the user queries in order to retrieve the relevant documents from specific clusters deemed
relevant to the query. Then the results are evaluated according to evaluation criteria. Recall and Precision
(P@5, P@10) of the retrieved results. P@5 was 0.660 and P@10 was 0.655.
Information retrieval (IR) is the science of searching for documents and information within documents. IR is interdisciplinary and involves computer science, mathematics, psychology and other fields. Information storage and retrieval (ISAR) systems allow users to store, manipulate and analyze data and report results on a regular basis. ISAR is widely used in fields like science, business and healthcare. The information retrieval process begins with a user entering a query, which is then matched to objects stored in the database to retrieve relevant results based on scoring algorithms. Performance is evaluated using precision and recall metrics.
This document provides an overview of an information retrieval system. It defines an information retrieval system as a system capable of storing, retrieving, and maintaining information such as text, images, audio, and video. The objectives of an information retrieval system are to minimize the overhead for a user to locate needed information. The document discusses functions like search, browse, indexing, cataloging, and various capabilities to facilitate querying and retrieving relevant information from the system.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering for recommendations. A web spider collects content indexes from learning management systems and stores data in an LDAP directory. Users can search for related materials, and the system mines log data to associate frequently searched terms and recommend additional resources.
The technology of object oriented databases was introduced to system developers in
the late 1980’s. Object DBMSs add database functionality to object programming languages. A
major benefit of this approach is the unification of the application and database development into
a seamless data model and language environment. As a result, applications require less code, use
more natural data modeling, and code bases are easier to maintain.
Data and Information Integration: Information ExtractionIJMER
Information extraction is generally concerned with the location of different items in any document, may be textual or web document. This paper is concerned with the methodologies and applications of information extraction. The field of information extraction plays a very important role in the natural language processing community. The architecture of information extraction system which acts as the base for all languages and fields is also discussed along with its different components. Information is hidden in the large volume of web pages and thus it is necessary to extract useful information from the web content, called Information Extraction. In information extraction, given a sequence of instances, we identify and pull out a sub-sequence of the input that represents information we are interested in.
Manual data extraction from semi supervised web pages is a difficult task. This paper focuses on study of various data extraction techniques and also some web data extraction techniques. In the past years, there was a rapid expansion of activities in the information extraction area. Many methods have been proposed for automating the process of extraction. We will survey various web data extraction tools. Several real-world applications of information extraction will be introduced. What role information extraction plays in different fields is discussed in these applications. Current challenges being faced by the available information extraction techniques are briefly discussed along with the future work going on using the current researches is discussed.
The Survey of Data Mining Applications And Feature Scope IJCSEIT Journal
In this paper we have focused a variety of techniques, approaches and different areas of the research which
are helpful and marked as the important field of data mining Technologies. As we are aware that many MNC’s
and large organizations are operated in different places of the different countries. Each place of operation
may generate large volumes of data. Corporate decision makers require access from all such sources and
take strategic decisions .The data warehouse is used in the significant business value by improving the
effectiveness of managerial decision-making. In an uncertain and highly competitive business
environment, the value of strategic information systems such as these are easily recognized however in
today’s business environment, efficiency or speed is not the only key for competitiveness. This type of huge
amount of data’s are available in the form of tera- to peta-bytes which has drastically changed in the areas
of science and engineering. To analyze, manage and make a decision of such type of huge amount of data
we need techniques called the data mining which will transforming in many fields. This paper imparts more
number of applications of the data mining and also o focuses scope of the data mining which will helpful in
the further research.
IRJET-Model for semantic processing in information retrieval systemsIRJET Journal
This document proposes a model for semantic information retrieval that improves upon traditional keyword matching approaches. It involves three main components:
1. A crawling and indexing component that identifies websites and pages, extracts metadata, and generates a knowledge graph through semantic annotation.
2. A processing component that analyzes user queries and profiles to understand search intent, calculates semantic similarity between queries and indexed documents, and determines result relevance.
3. A presentation component that displays search results to users through both simple and advanced search interfaces, prioritizing the most relevant information based on the above processing.
The model is intended to address deficiencies in current Cuban web search by better understanding natural language queries and the contextual meaning of information through semantic technologies
A Review: Text Classification on Social Media DataIOSR Journals
This document provides a review of different classifiers used for text classification on social media data. It discusses how social media data is often unstructured and contains users' opinions and sentiments. Various machine learning algorithms can be used to classify this social media text data, extracting meaningful information. The document focuses on describing Naive Bayes classifiers, which are commonly used for text classification tasks. It explains how Naive Bayes classifiers work by calculating the posterior probability that a document belongs to a certain class, based on applying Bayes' theorem with an independence assumption between features.
The document provides an overview of the key components and objectives of an information retrieval system. It discusses how an IR system aims to minimize the time a user spends locating needed information by facilitating search generation, presenting search results in a relevant order, and processing incoming documents through normalization, indexing, and selective dissemination to users. The major measures of an IR system's effectiveness are precision and recall.
Survey of Machine Learning Techniques in Textual Document ClassificationIOSR Journals
Classification of Text Document points towards associating one or more predefined categories based
on the likelihood expressed by the training set of labeled documents. Many machine learning algorithms plays
an important role in training the system with predefined categories. The importance of Machine learning
approach has felt because of which the study has been taken up for text document classification based on the
statistical event models available. The aim of this paper is to present the important techniques and
methodologies that are employed for text documents classification, at the same time making awareness of some
of the interesting challenges that remain to be solved, focused mainly on text representation and machine
learning techniques.
Sentimental classification analysis of polarity multi-view textual data using...IJECEIAES
The data and information available in most community environments is complex in nature. Sentimental data resources may possibly consist of textual data collected from multiple information sources with different representations and usually handled by different analytical models. These types of data resource characteristics can form multi-view polarity textual data. However, knowledge creation from this type of sentimental textual data requires considerable analytical efforts and capabilities. In particular, data mining practices can provide exceptional results in handling textual data formats. Besides, in the case of the textual data exists as multi-view or unstructured data formats, the hybrid and integrated analysis efforts of text data mining algorithms are vital to get helpful results. The objective of this research is to enhance the knowledge discovery from sentimental multi-view textual data which can be considered as unstructured data format to classify the polarity information documents in the form of two different categories or types of useful information. A proposed framework with integrated data mining algorithms has been discussed in this paper, which is achieved through the application of X-means algorithm for clustering and HotSpot algorithm of association rules. The analysis results have shown improved accuracies of classifying the sentimental multi-view textual data into two categories through the application of the proposed framework on online polarity user-reviews dataset upon a given topics.
The document summarizes a research paper that proposes a customized ontological model for representing user profiles to improve web information gathering. The model uses both a global knowledge base and local user repositories to construct personalized ontologies. It introduces a multidimensional ontology mining method to analyze ontology concepts. The local repositories are then used to populate the personalized ontologies with background knowledge. An evaluation compares the proposed model to benchmarks and finds it successfully represents user profiles.
Searching and Analyzing Qualitative Data on Personal ComputerIOSR Journals
This document presents the design and implementation of a desktop search system using Lucene. It describes the key components of indexing, analyzing text, storing indexes, and searching. For indexing, it discusses how documents are preprocessed, tokenized, and stored in an inverted index. For searching, it explains how queries are analyzed and the index is searched to return results. The system allows users to search for files on their personal computer. It includes a user interface to input queries and view results. Lucene provides an open-source toolkit to add full-text search capabilities to applications.
WHAT IS A DBMS? EXPLAIN DIFFERENT MYSQL COMMANDS AND CONSTRAINTS OF THE SAME.`Shweta Bhavsar
This document discusses database management systems (DBMS) and MySQL commands and constraints. It begins by defining a DBMS and describing their components and characteristics, including data models, query languages, and advantages like data integrity and sharing. It then explains common MySQL commands to create and manage databases, tables, and insert values. Constraints are also discussed as ways to define data types and validate values in tables.
A novel method to search information through multi agent search and retrieIAEME Publication
The document proposes a novel method for searching information through multi-agent search and retrieval that uses both content and context-based search. It describes a system that accepts two text inputs, processes them according to whether desktop or internet search is selected, and provides relevant results through indexing and multi-agents, with content search performed on Hadoop for increased performance. The system aims to provide faster and more accurate search results by filtering out irrelevant results.
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...Kato Mivule
This document provides an overview of data privacy challenges in multi-agent learning systems. It discusses literature on privacy issues in multi-agent systems, including problems related to autonomy, trust, and defining privacy. Key privacy issues for multi-agent systems are identified as information collection, processing, and dissemination by agents. Several proposed abstract architectures aim to address privacy in multi-agent design, but challenges remain regarding trade-offs between privacy and data utility.
This document provides an overview of the course "Database Management System" (MBA 758) offered by the National Open University of Nigeria. The course aims to teach students how to design, manipulate, and manage databases. It contains 13 units across 3 modules covering topics such as database models, SQL, security, administration, and more. Students will be assessed through tutor-marked assignments making up 30% of the final grade, and an end of course exam making up the remaining 70%. The course is designed to provide graduates with skills in effectively managing data through the use of database management systems.
An Improved Annotation Based Summary Generation For Unstructured DataMelinda Watson
This document discusses annotation-based summarization of unstructured data. It begins with an introduction to annotation and information retrieval. Current annotation processes cannot maintain modifications due to frequent document updates. The document then reviews literature on automatic text classification, applying annotations to linked open data sets, and using domain ontologies for automatic document annotation. Keywords, sentences and contexts are extracted from documents for annotation. Different annotation models are discussed. The goal is to develop an improved annotation approach for summarizing unstructured data that can handle frequent document changes.
This document provides an overview of information retrieval models. It begins with definitions of information retrieval and how it differs from data retrieval. It then discusses the retrieval process and logical representations of documents. A taxonomy of IR models is presented including classic, structured, and browsing models. Boolean, vector, and probabilistic models are explained as examples of classic models. The document concludes with descriptions of ad-hoc retrieval and filtering tasks and formal characteristics of IR models.
UML MODELING AND SYSTEM ARCHITECTURE FOR AGENT BASED INFORMATION RETRIEVALijcsit
In this current technological era, there is an enormous increase in the information available on web and
also in the online databases. This information abundance increases the complexity of finding relevant
information. To solve such challenges, there is a need for improved and intelligent systems for efficient
search and retrieval. Intelligent Agents can be used for better search and information retrieval in a
document collection. The information required by a user is scattered in a large number of databases. In this
paper, the object oriented modeling for agent based information retrieval system is presented. The paper
also discusses the framework of agent architecture for obtaining the best combination terms that serve as
an input query to the information retrieval system. The communication and cooperation among the agents
are also explained. Each agent has a task to perform in information retrieval.
Algorithm for calculating relevance of documents in information retrieval sys...IRJET Journal
The document proposes an algorithm to calculate the relevance of documents returned in response to user queries in information retrieval systems. It is based on classical similarity formulas like cosine, Jaccard, and dice that calculate similarity between document and query vectors. The algorithm aims to integrate user search preferences as a variable in determining document relevance, as classic models do not account for this. It uses text and web mining techniques to process user query and document metadata.
Information retrieval (IR) is the science of searching for documents and information within documents. IR is interdisciplinary and involves computer science, mathematics, psychology and other fields. Information storage and retrieval (ISAR) systems allow users to store, manipulate and analyze data and report results on a regular basis. ISAR is widely used in fields like science, business and healthcare. The information retrieval process begins with a user entering a query, which is then matched to objects stored in the database to retrieve relevant results based on scoring algorithms. Performance is evaluated using precision and recall metrics.
This document provides an overview of an information retrieval system. It defines an information retrieval system as a system capable of storing, retrieving, and maintaining information such as text, images, audio, and video. The objectives of an information retrieval system are to minimize the overhead for a user to locate needed information. The document discusses functions like search, browse, indexing, cataloging, and various capabilities to facilitate querying and retrieving relevant information from the system.
In tech application-of_data_mining_technology_on_e_learning_material_recommen...Enhmandah Hemeelee
The document describes a recommendation system that applies data mining techniques to recommend e-learning materials. It proposes using LDAP for fast searching of materials across systems, JAXB for parsing content, and association rule mining and collaborative filtering for recommendations. A web spider collects content indexes from learning management systems and stores data in an LDAP directory. Users can search for related materials, and the system mines log data to associate frequently searched terms and recommend additional resources.
The technology of object oriented databases was introduced to system developers in
the late 1980’s. Object DBMSs add database functionality to object programming languages. A
major benefit of this approach is the unification of the application and database development into
a seamless data model and language environment. As a result, applications require less code, use
more natural data modeling, and code bases are easier to maintain.
Data and Information Integration: Information ExtractionIJMER
Information extraction is generally concerned with the location of different items in any document, may be textual or web document. This paper is concerned with the methodologies and applications of information extraction. The field of information extraction plays a very important role in the natural language processing community. The architecture of information extraction system which acts as the base for all languages and fields is also discussed along with its different components. Information is hidden in the large volume of web pages and thus it is necessary to extract useful information from the web content, called Information Extraction. In information extraction, given a sequence of instances, we identify and pull out a sub-sequence of the input that represents information we are interested in.
Manual data extraction from semi supervised web pages is a difficult task. This paper focuses on study of various data extraction techniques and also some web data extraction techniques. In the past years, there was a rapid expansion of activities in the information extraction area. Many methods have been proposed for automating the process of extraction. We will survey various web data extraction tools. Several real-world applications of information extraction will be introduced. What role information extraction plays in different fields is discussed in these applications. Current challenges being faced by the available information extraction techniques are briefly discussed along with the future work going on using the current researches is discussed.
The Survey of Data Mining Applications And Feature Scope IJCSEIT Journal
In this paper we have focused a variety of techniques, approaches and different areas of the research which
are helpful and marked as the important field of data mining Technologies. As we are aware that many MNC’s
and large organizations are operated in different places of the different countries. Each place of operation
may generate large volumes of data. Corporate decision makers require access from all such sources and
take strategic decisions .The data warehouse is used in the significant business value by improving the
effectiveness of managerial decision-making. In an uncertain and highly competitive business
environment, the value of strategic information systems such as these are easily recognized however in
today’s business environment, efficiency or speed is not the only key for competitiveness. This type of huge
amount of data’s are available in the form of tera- to peta-bytes which has drastically changed in the areas
of science and engineering. To analyze, manage and make a decision of such type of huge amount of data
we need techniques called the data mining which will transforming in many fields. This paper imparts more
number of applications of the data mining and also o focuses scope of the data mining which will helpful in
the further research.
IRJET-Model for semantic processing in information retrieval systemsIRJET Journal
This document proposes a model for semantic information retrieval that improves upon traditional keyword matching approaches. It involves three main components:
1. A crawling and indexing component that identifies websites and pages, extracts metadata, and generates a knowledge graph through semantic annotation.
2. A processing component that analyzes user queries and profiles to understand search intent, calculates semantic similarity between queries and indexed documents, and determines result relevance.
3. A presentation component that displays search results to users through both simple and advanced search interfaces, prioritizing the most relevant information based on the above processing.
The model is intended to address deficiencies in current Cuban web search by better understanding natural language queries and the contextual meaning of information through semantic technologies
A Review: Text Classification on Social Media DataIOSR Journals
This document provides a review of different classifiers used for text classification on social media data. It discusses how social media data is often unstructured and contains users' opinions and sentiments. Various machine learning algorithms can be used to classify this social media text data, extracting meaningful information. The document focuses on describing Naive Bayes classifiers, which are commonly used for text classification tasks. It explains how Naive Bayes classifiers work by calculating the posterior probability that a document belongs to a certain class, based on applying Bayes' theorem with an independence assumption between features.
The document provides an overview of the key components and objectives of an information retrieval system. It discusses how an IR system aims to minimize the time a user spends locating needed information by facilitating search generation, presenting search results in a relevant order, and processing incoming documents through normalization, indexing, and selective dissemination to users. The major measures of an IR system's effectiveness are precision and recall.
Survey of Machine Learning Techniques in Textual Document ClassificationIOSR Journals
Classification of Text Document points towards associating one or more predefined categories based
on the likelihood expressed by the training set of labeled documents. Many machine learning algorithms plays
an important role in training the system with predefined categories. The importance of Machine learning
approach has felt because of which the study has been taken up for text document classification based on the
statistical event models available. The aim of this paper is to present the important techniques and
methodologies that are employed for text documents classification, at the same time making awareness of some
of the interesting challenges that remain to be solved, focused mainly on text representation and machine
learning techniques.
Sentimental classification analysis of polarity multi-view textual data using...IJECEIAES
The data and information available in most community environments is complex in nature. Sentimental data resources may possibly consist of textual data collected from multiple information sources with different representations and usually handled by different analytical models. These types of data resource characteristics can form multi-view polarity textual data. However, knowledge creation from this type of sentimental textual data requires considerable analytical efforts and capabilities. In particular, data mining practices can provide exceptional results in handling textual data formats. Besides, in the case of the textual data exists as multi-view or unstructured data formats, the hybrid and integrated analysis efforts of text data mining algorithms are vital to get helpful results. The objective of this research is to enhance the knowledge discovery from sentimental multi-view textual data which can be considered as unstructured data format to classify the polarity information documents in the form of two different categories or types of useful information. A proposed framework with integrated data mining algorithms has been discussed in this paper, which is achieved through the application of X-means algorithm for clustering and HotSpot algorithm of association rules. The analysis results have shown improved accuracies of classifying the sentimental multi-view textual data into two categories through the application of the proposed framework on online polarity user-reviews dataset upon a given topics.
The document summarizes a research paper that proposes a customized ontological model for representing user profiles to improve web information gathering. The model uses both a global knowledge base and local user repositories to construct personalized ontologies. It introduces a multidimensional ontology mining method to analyze ontology concepts. The local repositories are then used to populate the personalized ontologies with background knowledge. An evaluation compares the proposed model to benchmarks and finds it successfully represents user profiles.
Searching and Analyzing Qualitative Data on Personal ComputerIOSR Journals
This document presents the design and implementation of a desktop search system using Lucene. It describes the key components of indexing, analyzing text, storing indexes, and searching. For indexing, it discusses how documents are preprocessed, tokenized, and stored in an inverted index. For searching, it explains how queries are analyzed and the index is searched to return results. The system allows users to search for files on their personal computer. It includes a user interface to input queries and view results. Lucene provides an open-source toolkit to add full-text search capabilities to applications.
WHAT IS A DBMS? EXPLAIN DIFFERENT MYSQL COMMANDS AND CONSTRAINTS OF THE SAME.`Shweta Bhavsar
This document discusses database management systems (DBMS) and MySQL commands and constraints. It begins by defining a DBMS and describing their components and characteristics, including data models, query languages, and advantages like data integrity and sharing. It then explains common MySQL commands to create and manage databases, tables, and insert values. Constraints are also discussed as ways to define data types and validate values in tables.
A novel method to search information through multi agent search and retrieIAEME Publication
The document proposes a novel method for searching information through multi-agent search and retrieval that uses both content and context-based search. It describes a system that accepts two text inputs, processes them according to whether desktop or internet search is selected, and provides relevant results through indexing and multi-agents, with content search performed on Hadoop for increased performance. The system aims to provide faster and more accurate search results by filtering out irrelevant results.
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...Kato Mivule
This document provides an overview of data privacy challenges in multi-agent learning systems. It discusses literature on privacy issues in multi-agent systems, including problems related to autonomy, trust, and defining privacy. Key privacy issues for multi-agent systems are identified as information collection, processing, and dissemination by agents. Several proposed abstract architectures aim to address privacy in multi-agent design, but challenges remain regarding trade-offs between privacy and data utility.
This document provides an overview of the course "Database Management System" (MBA 758) offered by the National Open University of Nigeria. The course aims to teach students how to design, manipulate, and manage databases. It contains 13 units across 3 modules covering topics such as database models, SQL, security, administration, and more. Students will be assessed through tutor-marked assignments making up 30% of the final grade, and an end of course exam making up the remaining 70%. The course is designed to provide graduates with skills in effectively managing data through the use of database management systems.
An Improved Annotation Based Summary Generation For Unstructured DataMelinda Watson
This document discusses annotation-based summarization of unstructured data. It begins with an introduction to annotation and information retrieval. Current annotation processes cannot maintain modifications due to frequent document updates. The document then reviews literature on automatic text classification, applying annotations to linked open data sets, and using domain ontologies for automatic document annotation. Keywords, sentences and contexts are extracted from documents for annotation. Different annotation models are discussed. The goal is to develop an improved annotation approach for summarizing unstructured data that can handle frequent document changes.
This document provides an overview of information retrieval models. It begins with definitions of information retrieval and how it differs from data retrieval. It then discusses the retrieval process and logical representations of documents. A taxonomy of IR models is presented including classic, structured, and browsing models. Boolean, vector, and probabilistic models are explained as examples of classic models. The document concludes with descriptions of ad-hoc retrieval and filtering tasks and formal characteristics of IR models.
UML MODELING AND SYSTEM ARCHITECTURE FOR AGENT BASED INFORMATION RETRIEVALijcsit
In this current technological era, there is an enormous increase in the information available on web and
also in the online databases. This information abundance increases the complexity of finding relevant
information. To solve such challenges, there is a need for improved and intelligent systems for efficient
search and retrieval. Intelligent Agents can be used for better search and information retrieval in a
document collection. The information required by a user is scattered in a large number of databases. In this
paper, the object oriented modeling for agent based information retrieval system is presented. The paper
also discusses the framework of agent architecture for obtaining the best combination terms that serve as
an input query to the information retrieval system. The communication and cooperation among the agents
are also explained. Each agent has a task to perform in information retrieval.
Algorithm for calculating relevance of documents in information retrieval sys...IRJET Journal
The document proposes an algorithm to calculate the relevance of documents returned in response to user queries in information retrieval systems. It is based on classical similarity formulas like cosine, Jaccard, and dice that calculate similarity between document and query vectors. The algorithm aims to integrate user search preferences as a variable in determining document relevance, as classic models do not account for this. It uses text and web mining techniques to process user query and document metadata.
An effective pre processing algorithm for information retrieval systemsijdms
The Internet is probably the most successful distributed computing system ever. However, our capabilities
for data querying and manipulation on the internet are primordial at best. The user expectations are
enhancing over the period of time along with increased amount of operational data past few decades. The
data-user expects more deep, exact, and detailed results. Result retrieval for the user query is always
relative o the pattern of data storage and index. In Information retrieval systems, tokenization is an
integrals part whose prime objective is to identifying the token and their count. In this paper, we have
proposed an effective tokenization approach which is based on training vector and result shows that
efficiency/ effectiveness of proposed algorithm. Tokenization on documents helps to satisfy user’s
information need more precisely and reduced search sharply, is believed to be a part of information
retrieval. Pre-processing of input document is an integral part of Tokenization, which involves preprocessing
of documents and generates its respective tokens which is the basis of these tokens probabilistic
IR generate its scoring and gives reduced search space. The comparative analysis is based on the two
parameters; Number of Token generated, Pre-processing time.
Web based of electronic document management systemsjournal ijrtem
ABSTRACT : An electronic document management system is a software program that manages the creation, storage and control of document electronically. The primary function of that system is to manage electronic information within an organization’s workflow. In this paper the aim is to develop web based application for document management systems that is able to deliver access to any one authorized anytime, anyplace on any device. Keywords: Document management systems, electronic document, web application, digital document, document scanning
IRJET-Computational model for the processing of documents and support to the ...IRJET Journal
This document proposes a computational model for processing documents and supporting decision making in information retrieval systems. The model includes five main components: 1) a tracking and indexing component to crawl the web and store document metadata, 2) an information processing component to categorize documents and define user profiles, 3) a decision support component to analyze stored information and generate statistical reports, 4) a display component to provide search interfaces and visualization tools, and 5) specialized roles to administer the system. The goal of the model is to provide a framework for developing large-scale search engines.
Here are the key points about using content-based filtering techniques:
- Content-based filtering relies on analyzing the content or description of items to recommend items similar to what the user has liked in the past. It looks for patterns and regularities in item attributes/descriptions to distinguish highly rated items.
- The item content/descriptions are analyzed automatically by extracting information from sources like web pages, or entered manually from product databases.
- It focuses on objective attributes about items that can be extracted algorithmically, like text analysis of documents.
- However, personal preferences and what makes an item appealing are often subjective qualities not easily extracted algorithmically, like writing style or taste.
- So while content-based filtering can
A Systems Approach To Qualitative Data Management And AnalysisMichele Thomas
This article proposes a systematic approach to qualitative data management centered around a database with four main elements: (1) characteristics of data sources, (2) primary data collected from sources, (3) secondary data generated to assist interpretation, and (4) characteristics of coders. The approach tracks the analysis process from framing a research question to developing an empirical answer. It emphasizes distinguishing different types of data, tracking relationships between elements, and supporting reliability assessments to facilitate efficient and valid analysis.
Characterizing and Processing of Big Data Using Data Mining TechniquesIJTET Journal
The document discusses big data and techniques for processing it, including data mining. It begins by defining big data and its key characteristics of volume, variety, and velocity. It then discusses various data mining techniques that can be used to process big data, including clustering, classification, and prediction. It introduces the HACE theorem for characterizing big data based on its huge size, heterogeneous and diverse sources, decentralized control, and complex relationships within the data. The document proposes a big data processing model involving data set aggregation, pre-processing, connectivity-based clustering, and subset selection to efficiently retrieve relevant data. It evaluates the performance of subset selection versus deterministic search methods.
1. The essay describes the proper technique for holding a deck of cards in one's hand in preparation for various card manipulation moves.
2. It outlines the starting position for holding the deck, with the fingers curled around and supporting the deck in a tilted position resting on the forefinger.
3. Different spreads are explained, such as the regular spread where the left thumb pushes cards to the right and the right hand collects them in a straight line.
Leveraging Open Source Technologies to Enable Scientific Archiving and Discovery; Steve Hughes, NASA; Data Publication Repositories
The 2nd Research Data Access and Preservation (RDAP) Summit
An ASIS&T Summit
March 31-April 1, 2011 Denver, CO
In cooperation with the Coalition for Networked Information
http://asist.org/Conferences/RDAP11/index.html
Clustering of Deep WebPages: A Comparative Studyijcsit
The internethas massive amount of information. This information is stored in the form of zillions of
webpages. The information that can be retrieved by search engines is huge, and this information constitutes
the ‘surface web’.But the remaining information, which is not indexed by search engines – the ‘deep web’,
is much bigger in size than the ‘surface web’, and remains unexploited yet.
Several machine learning techniques have been commonly employed to access deep web content. Under
machine learning, topic models provide a simple way to analyze large volumes of unlabeled text. A ‘topic’is
a cluster of words that frequently occur together and topic models can connect words with similar
meanings and distinguish between words with multiple meanings. In this paper, we cluster deep web
databases employing several methods, and then perform a comparative study. In the first method, we apply
Latent Semantic Analysis (LSA) over the dataset. In the second method, we use a generative probabilistic
model called Latent Dirichlet Allocation(LDA) for modeling content representative of deep web
databases.Both these techniques are implemented after preprocessing the set of web pages to extract page
contents and form contents.Further, we propose another version of Latent Dirichlet Allocation (LDA) to the
dataset. Experimental results show that the proposed method outperforms the existing clustering methods.
1. The paper proposes techniques to extract hidden databases when a user query returns many valid tuples but only some are displayed, with the others hidden.
2. It focuses on interfaces called "TOP-k-COUNT" interfaces that display some tuples and provide the count of other matching tuples.
3. The COUNT-DECISION-TREE algorithm samples the hidden database using a decision tree to generalize the attribute order, allowing different attributes at each level.
1. The document proposes techniques to improve search performance by matching schemas between structured and unstructured data sources.
2. It involves constructing schema mappings using named entities and schema structures. It also uses strategies to narrow the search space to relevant documents.
3. The techniques were shown to improve search accuracy and reduce time/space complexity compared to existing methods.
DRMS cu uuvhg. HbbhyfufucufMiniproject.pptxshubhamrkokare
The document outlines a student project to develop a Document Repository Management System called DocVault. It will address challenges with managing documents for events in the college's computer department. A team of 4 students led by Ms. Rohini Bhosale will create DocVault, which aims to provide centralized storage, simple organization and search, and secure access to documents. The project will involve researching existing systems, designing interfaces and features, development, and testing to ensure DocVault streamlines document management for the department.
IRJET- Concept Extraction from Ambiguous Text Document using K-MeansIRJET Journal
This document discusses using a K-means clustering algorithm to extract concepts from ambiguous text documents. It involves preprocessing the text by tokenizing, removing stop words, and stemming words. The words are then represented as vectors and dimensionality reduction using PCA is applied. Finally, K-means clustering is used to group similar words into clusters to identify the overall concepts in the document without reading the entire text. The aim is to help users understand the key topics in a document in a time-efficient manner without having to read the full text.
This document discusses concepts related to data processing and databases. It begins with definitions of data processing and how it converts raw data into usable information. It then discusses the differences between data and information. The rest of the document covers database concepts like file processing, database management systems, database design principles like normalization, and different database models.
2016 BE Final year Projects in chennai - 1 Crore Projects 1crore projects
IEEE PROJECTS 2016 - 2017
1 crore projects is a leading Guide for ieee Projects and real time projects Works Provider.
It has been provided Lot of Guidance for Thousands of Students & made them more beneficial in all Technology Training.
Project Domain list 2016
1. IEEE based on datamining and knowledge engineering,
2. IEEE based on mobile computing,
3. IEEE based on networking,
4. IEEE based on Image processing,
5. IEEE based on Multimedia,
6. IEEE based on Network security,
7. IEEE based on parallel and distributed systems
Project Domain list 2016
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
ECE IEEE Projects 2016
1. Matlab project
2. Ns2 project
3. Embedded project
4. Robotics project
5. IOT Projects
Eligibility
Final Year students of
1. BSc (C.S)
2. BCA/B.E(C.S)
3. B.Tech IT
4. BE (C.S)
5. MSc (C.S)
6. MSc (IT)
7. MCA
8. MS (IT)
9. ME(ALL)
10. BE(ECE)(EEE)(E&I)
TECHNOLOGY USED AND FOR TRAINING IN
1. DOT NET
2. C sharp
3. ASP
4. VB
5. SQL SERVER
6. JAVA
7. J2EE
8. STRINGS
9. ORACLE
10. VB dotNET
11. EMBEDDED
12. MAT LAB
13. LAB VIEW
14. Multi Sim
CONTACT US:-
1 CRORE PROJECTS
Door No: 66 ,Ground Floor,
No. 172, Raahat Plaza, (Shopping Mall) ,Arcot Road, Vadapalani, Chennai,
Tamin Nadu, INDIA - 600 026
Email id: 1croreprojects@gmail.com
website:1croreprojects.com
Phone : +91 97518 00789 / +91 7708150152
The document discusses an electronic document management system (EDMS) called "University Electronic Management System" that was developed for Dnipropetrovs'k National University. The system provides a web-based platform for storing, searching, and retrieving electronic documents. It includes modules for administration, employees, curricula, the university structure, and students. The system aims to simplify documentation processes and provide timely information to support operations and strategic planning.
Similar to Design and Implementation of Meetings Document Management and Retrieval System (20)
LAND USE LAND COVER AND NDVI OF MIRZAPUR DISTRICT, UPRAHUL
This Dissertation explores the particular circumstances of Mirzapur, a region located in the
core of India. Mirzapur, with its varied terrains and abundant biodiversity, offers an optimal
environment for investigating the changes in vegetation cover dynamics. Our study utilizes
advanced technologies such as GIS (Geographic Information Systems) and Remote sensing to
analyze the transformations that have taken place over the course of a decade.
The complex relationship between human activities and the environment has been the focus
of extensive research and worry. As the global community grapples with swift urbanization,
population expansion, and economic progress, the effects on natural ecosystems are becoming
more evident. A crucial element of this impact is the alteration of vegetation cover, which plays a
significant role in maintaining the ecological equilibrium of our planet.Land serves as the foundation for all human activities and provides the necessary materials for
these activities. As the most crucial natural resource, its utilization by humans results in different
'Land uses,' which are determined by both human activities and the physical characteristics of the
land.
The utilization of land is impacted by human needs and environmental factors. In countries
like India, rapid population growth and the emphasis on extensive resource exploitation can lead
to significant land degradation, adversely affecting the region's land cover.
Therefore, human intervention has significantly influenced land use patterns over many
centuries, evolving its structure over time and space. In the present era, these changes have
accelerated due to factors such as agriculture and urbanization. Information regarding land use and
cover is essential for various planning and management tasks related to the Earth's surface,
providing crucial environmental data for scientific, resource management, policy purposes, and
diverse human activities.
Accurate understanding of land use and cover is imperative for the development planning
of any area. Consequently, a wide range of professionals, including earth system scientists, land
and water managers, and urban planners, are interested in obtaining data on land use and cover
changes, conversion trends, and other related patterns. The spatial dimensions of land use and
cover support policymakers and scientists in making well-informed decisions, as alterations in
these patterns indicate shifts in economic and social conditions. Monitoring such changes with the
help of Advanced technologies like Remote Sensing and Geographic Information Systems is
crucial for coordinated efforts across different administrative levels. Advanced technologies like
Remote Sensing and Geographic Information Systems
9
Changes in vegetation cover refer to variations in the distribution, composition, and overall
structure of plant communities across different temporal and spatial scales. These changes can
occur natural.
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
Executive Directors Chat Leveraging AI for Diversity, Equity, and InclusionTechSoup
Let’s explore the intersection of technology and equity in the final session of our DEI series. Discover how AI tools, like ChatGPT, can be used to support and enhance your nonprofit's DEI initiatives. Participants will gain insights into practical AI applications and get tips for leveraging technology to advance their DEI goals.
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
How to Manage Your Lost Opportunities in Odoo 17 CRMCeline George
Odoo 17 CRM allows us to track why we lose sales opportunities with "Lost Reasons." This helps analyze our sales process and identify areas for improvement. Here's how to configure lost reasons in Odoo 17 CRM
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
বাংলাদেশের অর্থনৈতিক সমীক্ষা ২০২৪ [Bangladesh Economic Review 2024 Bangla.pdf] কম্পিউটার , ট্যাব ও স্মার্ট ফোন ভার্সন সহ সম্পূর্ণ বাংলা ই-বুক বা pdf বই " সুচিপত্র ...বুকমার্ক মেনু 🔖 ও হাইপার লিংক মেনু 📝👆 যুক্ত ..
আমাদের সবার জন্য খুব খুব গুরুত্বপূর্ণ একটি বই ..বিসিএস, ব্যাংক, ইউনিভার্সিটি ভর্তি ও যে কোন প্রতিযোগিতা মূলক পরীক্ষার জন্য এর খুব ইম্পরট্যান্ট একটি বিষয় ...তাছাড়া বাংলাদেশের সাম্প্রতিক যে কোন ডাটা বা তথ্য এই বইতে পাবেন ...
তাই একজন নাগরিক হিসাবে এই তথ্য গুলো আপনার জানা প্রয়োজন ...।
বিসিএস ও ব্যাংক এর লিখিত পরীক্ষা ...+এছাড়া মাধ্যমিক ও উচ্চমাধ্যমিকের স্টুডেন্টদের জন্য অনেক কাজে আসবে ...
A review of the growth of the Israel Genealogy Research Association Database Collection for the last 12 months. Our collection is now passed the 3 million mark and still growing. See which archives have contributed the most. See the different types of records we have, and which years have had records added. You can also see what we have for the future.
Design and Implementation of Meetings Document Management and Retrieval System
1. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 376
Design and Implementation of Meetings Document Management
and Retrieval System
Obagbuwa, I. C. ibidunobagbuwa@yahoo.com
Department of Computer Sciences,
Lagos State University,
Lagos, +234, Nigeria.
Oyelade. O. J. ola.oyelade@covenantuniversity.edu.ng
Department of Computer and Information Sciences,
Covenant University,
Ota, +234, Nigeria.
Oladipupo. O. O. frajooje@yahoo.com
Department of Computer and Information Sciences,
Covenant University,
Ota, +234, Nigeria.
Ewejobi I. M. aitee4real2000@yahoo.com
Department of Computer and Information Sciences,
Covenant University,
Ota, +234, Nigeria.
Aborisade D. O. adoj_olan@hotmail.com
Department of Computer Science,
College of Natural Sciences,
Federal University of Agriculture,
Abeokuta, Ogun State, +234, Nigeria.
Abstract
Meetings management system has components to capture, storing/archiving, retrieve, browse,
and distribute documents from the system and Security to protect documents from unauthorized
access. Lack of proper organization, storage and easy access of meeting documents, bottleneck
of keeping paper documents, slow distribution, and misplacement of documents necessitated the
need for this work. Document management software that can be used to organize and maintain
the records of meetings has been developed. The system, developed as a web application, is
based on the use of objects and Web technologies. A search facility is included to support rapid
location of topics of interest, and navigation is enabled by the employment of hyperlinks. The
system was implemented using asp.net. This document management system can enable users
to follow the development of any topic through several meetings of a particular body or
committee, Members of the body should be able to have instant and full access to what has been
discussed and decided about the given issue no matter how long that had been.
Keywords: Document Management, Capture, Storage, Retrieval, Security, Hypertext and
Hyperlinks.
1. INTRODUCTION
Document management systems are software applications that capture paper and electronic
documents and provide the storage, retrieval, security and archival of those documents. These
electronic documents they manage can include any kind of digital object -- bitmap images, HTML
files, SGML, PDF, graphics, spreadsheets, and word-processed documents. The document
management process begins with the conversion of paper documents and records to electronic
2. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 377
files. Document management systems expedite business processes by allowing instant access to
information; greater collaboration within and among departments and offices; enhanced security
for files and records [6].
Some document management systems are already in existence, such as paperless office [12],
Document imaging system[14], content management [15], Business document management[13]
etc, but none has been designed to manage meeting documents ; meeting agendas, members
registration, the minutes of the meetings, for proper storage and easy retrieval. The focus of this
work is managing meetings documents.
In this research work, a document management system has components for: capturing and
bringing documents into the system; methods for storing and archiving documents; indexing and
retrieval tools to locate documents; and for distribution of documents. It is possible with this
system to manage millions of documents and retrieve and document of choice. The major tool for
achieving this are the indexing methods which are used to categorize and link documents,
indexing allows users to quickly sort large volumes of data to find the right document.
The three primary ways of indexing files in a document management system are: Full-text
indexing, or indexing every word contained within a document; Index fields, or indexing through
keyword categories of documents; and Folder/file structure, or indexing by associated document
groups.
Retrieval is where the quality of the indexing system is most evident. Some document
management systems let users search only by indexed keywords, which requires a person to
know how the document was categorized and what index fields were assigned to it [6,12,14]. A
powerful indexing system will make it possible for users to find any document based on what they
know, even if that amounts to no more than a word or phrase within the document [6].
Since a document can be viewed as a complex object, utilizing object-oriented approach in the
design and implementation of the document management system is likely to enhance it. For
example, a meeting could be viewed as consisting of a schedule, an agenda, a minute collection,
and so on. An object-oriented database management system (OODBMS) is a database
management system that supports the modeling and creation of data as objects [8,10]. This
includes some kind of support for classes of objects and the inheritance of class properties and
methods by subclasses and their objects.
Object-orientation is yet another step in the quest for expressing solutions to problems in a more
natural, easier-to-understand way [1,8], "The fundamental characteristic of this level of system
description being that it is closer to the human conceptualization of a problem domain.
Descriptions at this level can enhance communication between system designers, domain
experts and, ultimately, system end-users."
A data model consists of static properties such as objects; attributes and relationships; and
integrity rules over objects and operations and dynamic properties such as operations or rules
defining new database states based on applied state changes [7,8]. Object-oriented databases
have the ability to model all of these components directly within the database supporting a
complete problem/solution modeling capability. Prior to object-oriented databases, databases
relied on applications for defining the dynamic properties of the model. The disadvantage of
delegating the dynamic properties to applications is that these dynamic properties could not be
applied uniformly in all database usage scenarios since they were defined outside of the
database in autonomous applications [1,7,8]. Object-oriented databases provide a unifying
paradigm that allows the integration of all three aspects of data modeling and to apply them
uniformly to all users of the database [7].
3. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 378
The use of hypertext in documents could also enhance the document browsing experience.
Hypertext has to do with interrelating information elements (linking pieces of information) and
using these links to access them. Thus, a hypertext is a collection or web of interrelated or linked
nodes, these nodes or information element ranging from single ideas or chunks to entire
documents [3, 4]. The system allows an author to create the nodes and the links among them,
and allows a reader to traverse these links, i.e., to navigate from one node to another using these
links. The nodes contain the content and attributes of information elements.
Typically hypertext systems mark link access points or link anchors in some manner within a
node when displaying it on a computer screen (e.g., underlined text displayed within documents
on World Wide Web browsers). When the user selects the link marker, e.g., by clicking on it with
a mouse cursor, the hypertext system traverses to and displays the node at the other end of the
link. Hypertext systems include navigation, annotation and structural features, which take
advantage of the node and link structure to support authors and readers. Hypertext enables
people to read, author and comprehend information more effectively than traditional documents
[3,9]. People typically read documents from start to end, i.e., in a linear, sequential manner.
Hypertext frees readers and authors from this linear, sequential form of expression. Authors can
structure information as a web of information chunks and interrelating links. Presenting
information as a web enables readers to access information in the order most appropriate to their
purposes [2].
2. ANALYSIS AND DESIGN
The meeting management system is expected to have a database for storing minutes of
meetings and other information pertaining to meetings; Capture documents into the system;
storing and archiving documents; retrieve documents; Distribution of documents from the system;
and Security to protect documents from unauthorized access.
Object oriented analysis for any system typically begins with the identification of the major actors
within the system. For the system under consideration the actors are the Users of the system.
The Users can be classified into two – Administrator and Members. The Use Case model for the
system is depicted in Figure 1.0 below.
Security
Dispatch
e-mail/letters
Manage
members.
Manage meetings
Enquiries /Report
Registration
Enquiries
Hypertext
document
Minutes
(hypertext)
FIGURE 2.1: Administrator and Member
Members
Administrator
4. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 379
Use Case model for the system. The use cases are as described below:
CASE 1 - ADMINISTRATOR
In this system, administrator performs four main functions:
(a) Security: Ensure that the system is secure; ensure strict access denial to non members;
Create the users of the system; and Create users authorization and levels of
authorization.
(b) Manage members: Register members; and Edit member’s accounts.
(c) Manage meetings: Create meetings; and Edit meeting accounts.
(d) Manage hyperlinks: Create links among related information, meetings etc.
(e) Dispatch notice /letter, minutes of meeting to members e-mail.
CASE 2 MEMBERS
Register for a meeting; Make enquiries, the enquiries is displayed in hypertext i.e links of
nodes; View hypertext document.
2.1 IDENTIFICATION OF CLASSES
Two main classes can be identified in the system, viz., User and Meeting. User has two
subclasses: Member and Administrator. The Meeting class has three subclasses: Schedule,
Minute and Registration. This is shown in figure 2.0.
User
Administrator Member
Meeting
Schedule
Minutes
Registration
FIGURE 2.2: Objects and Classes
5. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 380
2.2 MEETING SYSTEM DIAGRAM
The design of the system is shown in Figure 2.3 below.
FIGURE 2.3: Meeting Management System Diagram.
The system was broken down into six sub-systems: Administrative manager, Schedule manager,
Minutes manager, Registration manager, Dispatch manager, Enquiry/report manager
Administrative manager: manage members; creates members’ data such as: Names,
Addresses, Email, Phone number, Password, and so on. It also manages security, creating user
authorization and authorization levels.
Schedule manager: Creates a record for a proposed meeting containing elements such as:
Agenda, Chairman, Special guest, Date, Time, and Venue.
Dispatch manager: Dispatches both scheduled meeting notices and minutes of meetings held to
members’ e –mails.
Registration manager: Creates forms for members to register for meetings.
Minute manager: Used after the meeting has been held to: Enter the minutes of the meeting, the
attendance, and any other remarks.
Enquiry/report manager: Supports enquiries and report generation. Enquiries could be about
the meeting schedules, minutes, agenda and attendance. Management reports could also be
generated on these matters.
Meeting management system
Administrative
Manager
Schedule
Manager
Dispatch
Manager
Minutes
Manager
Enquiries/report
Manager
Registration
Manager
securityMember
Management
6. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 381
2.3 DATA STORES
Meeting management database has six stores as described below:
Member (Member ID, Surname, Firstname, Address1, Address2, Address3, e- mail,
phone, password): Stores members’ data and passwords.
Meeting (Meeting ID, Schedule ID, Chairman, Agenda): Stores the meeting schedule chairman
and agenda
Schedule (Schedule ID, Meeting ID, Date, Time, Venue): Stores meeting date, time and
venue
Attendance (Meeting ID, Member ID): Stores members attendance at meetings.
Guest speaker (Meeting ID, Speaker, Topic): Stores speaker and topic
Minutes (Meeting ID, Details): Store meeting details
2.4 DOCUMENT RETRIEVAL
One important component of this system is the document retrieval and browsing component. This
component is implemented as part of the Enquiry /Report Manager. Our main motivation for this
work is to create a system that can enable users to follow the development of any topic through
several meetings of a particular body or committee. Members of the body should be able to have
instant and full access to what has been discussed and decided about the given issue no matter
how long that had been. This is important because we often have bodies and committees
discussing the same issues without the benefit of what had been earlier done, either by the same
or a different membership. This is also an important consideration for management in the process
of decision making and evaluation.
2.5 SYSTEM ARCHITECTURE
All web clients communicate with ASP.NET applications through IIS(internet information service) .
IIS finds the requested source (meeting management system application) and if the client is
authorized, returns the appropriate resources.
7. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 382
2.6 SECURITY CHECK
In this application, the ASP.NET form authentication is used.
FIGURE 2.5: A process flow for the forms authentication process [5].
ASP.NET
application
.NET
framework
Windows
NT/2000/XP
IIS
Web client
Meeting
mgt
system
Business
rules
Database
FIGURE 2.4: System Architecture.
8. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 383
3.0 SYSTEM IMPLEMENTATION
Web Forms pages are built with ASP.NET technology. ASP.NET is a unified Web platform that
provides all the services necessary to build enterprise-class applications.
ASP.NET, in turn, is built on the .NET Framework, so the entire framework is available to any
ASP.NET application. Applications can be authored in any language compatible with the common
language runtime, including Microsoft Visual Basic, Visual C#, and JScript .NET.
FIGURE 3.1: Login Page.
When the user is authenticated, the default page is displayed. If the password does not tally with
the existing password in the database, a access denial page is displayed. If a password is found,
registration page that has members account will be displayed as shown below in figure 3.2.
FIGURE 3.2: Members Account Page.
9. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 384
Members can register their accounts as shown in the figure 3.3 below
FIGURE 3.3: Viewing Page.
Members can make enquiries about meetings such as meetings in 2004
FIGURE 3.4: Enquiry Page.
11. Obagbuwa I. C.,Oyelade O. J., Oladipupo O. O., Ewejobi I. M. and Aborisade D. O.
International Journal of Computer Science and Security (IJCSS), Volume (6) : Issue (6) : 2012 386
14. Paperless office – Imaging solution ;1998 advanced information technologies.
15. Drupal content management system. http:www.org.accessed 28 march 2006.
16. Anas Tawile “ free and open source document management system” 2002