This document discusses various architectures, models, and methodologies used in information retrieval. It describes query models, ranking models, and feedback models used by researchers. It also highlights the importance of using context-based queries to better understand a user's search intent. The document provides an extensive survey of different approaches used in information retrieval systems and how adding context can help improve search results.
This paper proposes a framework called CIME (Competitive Intelligence Made Easy) that uses natural language processing techniques to automate the process of gathering competitive intelligence from public online sources such as news articles, blogs, and websites. CIME scrapes text from websites using keyword searches, applies natural language processing including tokenization and stemming to extract relevant information, and presents it in a structured format like an Excel template. The paper describes the architecture of CIME and provides an example use case where it was able to automatically generate a battlecard for cloud services competitors in under an hour, with 65% precision compared to manual analysis.
Algorithm for calculating relevance of documents in information retrieval sys...IRJET Journal
The document proposes an algorithm to calculate the relevance of documents returned in response to user queries in information retrieval systems. It is based on classical similarity formulas like cosine, Jaccard, and dice that calculate similarity between document and query vectors. The algorithm aims to integrate user search preferences as a variable in determining document relevance, as classic models do not account for this. It uses text and web mining techniques to process user query and document metadata.
Information Architecture Primer - Integrating search,tagging, taxonomy and us...Dan Keldsen
This document discusses the importance of taxonomy and classification within an information architecture. It defines key terms like taxonomy, thesaurus, ontology, and classification. It explains that taxonomy and classification help address the eternal problems of effectively cataloging and retrieving unstructured information. The document also discusses challenges like ambiguity, multiple meanings of words, and the importance of browsing versus searching in navigating large amounts of information.
AN ELABORATION OF TEXT CATEGORIZATION AND AUTOMATIC TEXT CLASSIFICATION THROU...cseij
This document discusses text categorization and automatic text classification. It begins by explaining how digitization has increased the need to organize and classify text for more efficient information retrieval. It then provides mathematical and graphical models for text mining, text categorization, and automatic text classification. These models are intended to improve understanding of these techniques and shorten response times for text retrieval. The document defines key terms and concepts related to text categorization and mining, and explains knowledge engineering and machine learning approaches to automatic text classification.
1. The document proposes techniques to improve search performance by matching schemas between structured and unstructured data sources.
2. It involves constructing schema mappings using named entities and schema structures. It also uses strategies to narrow the search space to relevant documents.
3. The techniques were shown to improve search accuracy and reduce time/space complexity compared to existing methods.
This document provides an overview of information retrieval models. It begins with definitions of information retrieval and how it differs from data retrieval. It then discusses the retrieval process and logical representations of documents. A taxonomy of IR models is presented including classic, structured, and browsing models. Boolean, vector, and probabilistic models are explained as examples of classic models. The document concludes with descriptions of ad-hoc retrieval and filtering tasks and formal characteristics of IR models.
Information Retrieval on Text using Concept Similarityrahulmonikasharma
This document summarizes a research paper on concept-based information retrieval using semantic analysis and WordNet. It discusses some of the challenges with keyword-based retrieval, such as synonymy and polysemy problems. Concept-based retrieval aims to address these issues by mapping documents and queries to semantic concepts rather than keywords. The paper proposes extracting concepts from text documents using WordNet to identify synonyms, hypernyms and hyponyms. It involves calculating term frequencies to determine a hierarchy of important concepts. The methodology is implemented using Java and WordNet to extract concepts from sample input documents.
Perception Determined Constructing Algorithm for Document ClusteringIRJET Journal
This document discusses an approach to document clustering called "Semantic Lingo" which aims to identify key concepts in documents and automatically generate an ontology based on these concepts to better conceptualize the documents. It provides background on challenges with traditional document clustering techniques and search engines. The proposed approach uses semantic information from domain ontologies to improve web search clustering quality by addressing issues like synonyms, polysemy and high dimensionality. It also discusses using text segments within documents that focus on one or more topics to aid multi-topic document clustering.
This paper proposes a framework called CIME (Competitive Intelligence Made Easy) that uses natural language processing techniques to automate the process of gathering competitive intelligence from public online sources such as news articles, blogs, and websites. CIME scrapes text from websites using keyword searches, applies natural language processing including tokenization and stemming to extract relevant information, and presents it in a structured format like an Excel template. The paper describes the architecture of CIME and provides an example use case where it was able to automatically generate a battlecard for cloud services competitors in under an hour, with 65% precision compared to manual analysis.
Algorithm for calculating relevance of documents in information retrieval sys...IRJET Journal
The document proposes an algorithm to calculate the relevance of documents returned in response to user queries in information retrieval systems. It is based on classical similarity formulas like cosine, Jaccard, and dice that calculate similarity between document and query vectors. The algorithm aims to integrate user search preferences as a variable in determining document relevance, as classic models do not account for this. It uses text and web mining techniques to process user query and document metadata.
Information Architecture Primer - Integrating search,tagging, taxonomy and us...Dan Keldsen
This document discusses the importance of taxonomy and classification within an information architecture. It defines key terms like taxonomy, thesaurus, ontology, and classification. It explains that taxonomy and classification help address the eternal problems of effectively cataloging and retrieving unstructured information. The document also discusses challenges like ambiguity, multiple meanings of words, and the importance of browsing versus searching in navigating large amounts of information.
AN ELABORATION OF TEXT CATEGORIZATION AND AUTOMATIC TEXT CLASSIFICATION THROU...cseij
This document discusses text categorization and automatic text classification. It begins by explaining how digitization has increased the need to organize and classify text for more efficient information retrieval. It then provides mathematical and graphical models for text mining, text categorization, and automatic text classification. These models are intended to improve understanding of these techniques and shorten response times for text retrieval. The document defines key terms and concepts related to text categorization and mining, and explains knowledge engineering and machine learning approaches to automatic text classification.
1. The document proposes techniques to improve search performance by matching schemas between structured and unstructured data sources.
2. It involves constructing schema mappings using named entities and schema structures. It also uses strategies to narrow the search space to relevant documents.
3. The techniques were shown to improve search accuracy and reduce time/space complexity compared to existing methods.
This document provides an overview of information retrieval models. It begins with definitions of information retrieval and how it differs from data retrieval. It then discusses the retrieval process and logical representations of documents. A taxonomy of IR models is presented including classic, structured, and browsing models. Boolean, vector, and probabilistic models are explained as examples of classic models. The document concludes with descriptions of ad-hoc retrieval and filtering tasks and formal characteristics of IR models.
Information Retrieval on Text using Concept Similarityrahulmonikasharma
This document summarizes a research paper on concept-based information retrieval using semantic analysis and WordNet. It discusses some of the challenges with keyword-based retrieval, such as synonymy and polysemy problems. Concept-based retrieval aims to address these issues by mapping documents and queries to semantic concepts rather than keywords. The paper proposes extracting concepts from text documents using WordNet to identify synonyms, hypernyms and hyponyms. It involves calculating term frequencies to determine a hierarchy of important concepts. The methodology is implemented using Java and WordNet to extract concepts from sample input documents.
Perception Determined Constructing Algorithm for Document ClusteringIRJET Journal
This document discusses an approach to document clustering called "Semantic Lingo" which aims to identify key concepts in documents and automatically generate an ontology based on these concepts to better conceptualize the documents. It provides background on challenges with traditional document clustering techniques and search engines. The proposed approach uses semantic information from domain ontologies to improve web search clustering quality by addressing issues like synonyms, polysemy and high dimensionality. It also discusses using text segments within documents that focus on one or more topics to aid multi-topic document clustering.
Context Driven Technique for Document ClassificationIDES Editor
In this paper we present an innovative hybrid Text
Classification (TC) system that bridges the gap between
statistical and context based techniques. Our algorithm
harnesses contextual information at two stages. First it extracts
a cohesive set of keywords for each category by using lexical
references, implicit context as derived from LSA and wordvicinity
driven semantics. And secondly, each document is
represented by a set of context rich features whose values are
derived by considering both lexical cohesion as well as the extent
of coverage of salient concepts via lexical chaining. After
keywords are extracted, a subset of the input documents is
apportioned as training set. Its members are assigned categories
based on their keyword representation. These labeled
documents are used to train binary SVM classifiers, one for
each category. The remaining documents are supplied to the
trained classifiers in the form of their context-enhanced feature
vectors. Each document is finally ascribed its appropriate
category by an SVM classifier.
Enhanced Web Usage Mining Using Fuzzy Clustering and Collaborative Filtering ...inventionjournals
This document discusses an enhanced web usage mining system using fuzzy clustering and collaborative filtering recommendation algorithms. It aims to address challenges with existing recommender systems like producing low quality recommendations for large datasets. The system architecture uses fuzzy clustering to predict future user access based on browsing behavior. Collaborative filtering is then used to produce expected results by combining fuzzy clustering outputs with a web database. This approach aims to provide users with more relevant recommendations in a shorter time compared to other systems.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
IRJET-Model for semantic processing in information retrieval systemsIRJET Journal
This document proposes a model for semantic information retrieval that improves upon traditional keyword matching approaches. It involves three main components:
1. A crawling and indexing component that identifies websites and pages, extracts metadata, and generates a knowledge graph through semantic annotation.
2. A processing component that analyzes user queries and profiles to understand search intent, calculates semantic similarity between queries and indexed documents, and determines result relevance.
3. A presentation component that displays search results to users through both simple and advanced search interfaces, prioritizing the most relevant information based on the above processing.
The model is intended to address deficiencies in current Cuban web search by better understanding natural language queries and the contextual meaning of information through semantic technologies
Comparison of Semantic and Syntactic Information Retrieval System on the basi...Waqas Tariq
In this paper information retrieval system for local databases are discussed. The approach is to search the web both semantically and syntactically. The proposal handles the search queries related to the user who is interested in the focused results regarding a product with some specific characteristics. The objective of the work will be to find and retrieve the accurate information from the available information warehouse which contains related data having common keywords. This information retrieval system can eventually be used for accessing the internet also. Accuracy in information retrieval that is achieving both high precision and recall is difficult. So both semantic and syntactic search engine are compared for information retrieval using two parameters i.e. precision and recall.
Projection Multi Scale Hashing Keyword Search in Multidimensional DatasetsIRJET Journal
The document discusses a novel method called ProMiSH (Projection and Multi Scale Hashing) for keyword search in multi-dimensional datasets. ProMiSH uses random projection and hash-based index structures to achieve high scalability and speedup of more than four orders over state-of-the-art tree-based techniques. Empirical studies on real and synthetic datasets of sizes up to 10 million objects and 100 dimensions show ProMiSH scales linearly with dataset size, dimension, query size, and result size. The method groups objects embedded in a vector space that are tagged with keywords matching a given query.
Metadata: Towards Machine-Enabled Intelligencedannyijwest
World Wide Web has revolutionized the means of data availability, but with its current structure model , it is becoming increasingly difficult to retrieve relevant information, with reasonable precision and recall, using the major search engines. However, with use of metadata, combined with the use of improved searching techniques, helps to enhance relevant information retrieval .The design of structured, descriptions of Web resources enables greater search precision and a more accurate relevance ranking of retrieved information .One such efforts towards standardization is , Dublin Core standard, which has been developed as Metadata Standard and also other standards which enhances retrieval of a wide range of information resources. This paper discuses the importance of metadata, various metadata schemas and elements, and the need of standardization of Metadata. This paper further discusses how the metadata can be generated using various tools which assist intelligent agents for efficient retrieval
Information retrieval (IR) is the science of searching for documents and information within documents. IR is interdisciplinary and involves computer science, mathematics, psychology and other fields. Information storage and retrieval (ISAR) systems allow users to store, manipulate and analyze data and report results on a regular basis. ISAR is widely used in fields like science, business and healthcare. The information retrieval process begins with a user entering a query, which is then matched to objects stored in the database to retrieve relevant results based on scoring algorithms. Performance is evaluated using precision and recall metrics.
Intelligent Semantic Web Search Engines: A Brief Survey dannyijwest
The World Wide Web (WWW) allows the people to share the information (data) from the large database repositories globally. The amount of information grows billions of databases. We need to search the information will specialize tools known generically search engine. There are many of search engines available today, retrieving meaningful information is difficult. However to overcome this problem in search engines to retrieve meaningful information intelligently, semantic web technologies are playing a major role. In this paper we present survey on the search engine generations and the role of search engines in intelligent web and semantic search technologies.
The document discusses the history and evolution of intelligent information retrieval and web search engines. It covers early concepts from the 1970s of intelligent information retrieval that focused on using intelligent agents. However, this concept did not work well and missed the key aspects of intelligence in information retrieval systems. True intelligence in information retrieval requires appropriate allocation of responsibilities among all actors in the system and effective support of people in their interactions with information. The document also discusses the evolution of web search engines from early keyword-based systems to modern approaches that focus more on user experience, engagement, and quality over quantity.
This document provides an overview of an information retrieval system (IRS). It defines IRS as obtaining relevant information from a collection to meet a user's need. The IRS has three main components: a document subsystem for acquiring, representing and organizing data; a user subsystem for representing queries; and a search/retrieval subsystem for matching queries to documents. It describes the basic concepts like how a user enters a query that is scored and ranked to return relevant results, which can be iterated. The objectives are to highlight probabilistic models and establish relationships between popular techniques. The functions are to analyze information sources and queries to match and retrieve relevant items. [/SUMMARY]
Semantics in Financial Services -David NewmanPeter Berger
David Newman serves as a Senior Architect in the Enterprise Architecture group at Wells Fargo Bank. He has been following semantic technology for the last 3 years; and has developed several business ontologies. He has been instrumental in thought leadership at Wells Fargo on the application of Semantic Technology and is a representative of the Financial Services Technology Consortium (FSTC)on the W3C SPARQL Working Group.
Comparative Study on Graph-based Information Retrieval: the Case of XML DocumentIJAEMSJORNAL
The processing of massive amounts of data has become indispensable especially with the potential proliferation of big data. The volume of information available nowadays makes it difficult for the user to find relevant information in a vast collection of documents. As a result, the exploitation of vast document collections necessitates the implementation of automated technologies that enable appropriate and effective retrieval. In this paper, we will examine the state of the art of IR in XML documents. We will also discuss some works that have used graphs to represent documents in the context of IR. In the same vein, the relationships between the components of a graph are the center of our attention.
IRJET - Re-Ranking of Google Search ResultsIRJET Journal
This document summarizes a research paper that proposes a hybrid personalized re-ranking approach to search results. It models a user's search interests using a conceptual user profile containing categories and concepts extracted from clicked results and a concept hierarchy. The user profile contains two types of documents - taxonomy documents representing general interests and viewed documents representing specific interests. A hybrid re-ranking process then semantically integrates the user's general and specific interests from their profile with search engine rankings to improve result relevance.
The document discusses a framework for web information retrieval using automatic multi-document summarization. It proposes using multi-level document summarization to enhance the effectiveness of web information retrieval by supporting indexing and ranking of retrieved documents with an intelligent decision making system based on fuzzy inference rules. The paper tests the approach on CACM test data and finds that information retrieval results can be improved after performing a multi-document summarization process.
IRJET-Computational model for the processing of documents and support to the ...IRJET Journal
This document proposes a computational model for processing documents and supporting decision making in information retrieval systems. The model includes five main components: 1) a tracking and indexing component to crawl the web and store document metadata, 2) an information processing component to categorize documents and define user profiles, 3) a decision support component to analyze stored information and generate statistical reports, 4) a display component to provide search interfaces and visualization tools, and 5) specialized roles to administer the system. The goal of the model is to provide a framework for developing large-scale search engines.
This document provides an overview of text mining and web mining. It defines data mining and describes the common data mining tasks of classification, clustering, association rule mining and sequential pattern mining. It then discusses text mining, defining it as the process of analyzing unstructured text data to extract meaningful information and structure. The document outlines the seven practice areas of text mining as search/information retrieval, document clustering, document classification, web mining, information extraction, natural language processing, and concept extraction. It provides brief descriptions of the problems addressed within each practice area.
An improved technique for ranking semantic associationst07IJwest
The primary focus of the search techniques in the first generation of the Web is accessing relevant
documents from the Web. Though it satisfies user requirements, but it is insufficient as the user sometimes
wishes to access actionable information involvin
g complex relationships between two given entities.
Finding such complex relationships (also known as semantic associations) is especially useful in
applications
such as
National Security, Pharmacy, Business Intelligence etc. Therefore the next frontier is
discovering relevant semantic associations between two entities present in large semantic metadata
repositories. Given two entities, there exist a huge number of semantic associations between two entities.
Hence ranking of these associations is required i
n order to find more relevant associations. For this
Aleman Meza et al. proposed a method involving six metrics viz. context, subsumption, rarity, popularity,
association length and trust. To compute the overall rank of the associations this method compute
s
context, subsumption, rarity and popularity values for each component of the association and for all the
associations. However it is obvious that, many components appears repeatedly in many associations
therefore it is not necessary to compute context, s
ubsumption, rarity
,
popularity
,
and
trust
values of the
components every time for each association rather the previously computed values may be used while
computing the overall rank of the associations. Thi
s paper proposes a method to re
use the previously
computed values using a hash data structure thus reduce the execution time. To demonstrate the
effectiveness of the proposed method, experiments were conducted on SWETO ontology. Results show
that the proposed method is more efficient than the other existi
ng methods
.
Solving np hard problem using artificial bee colony algorithmIAEME Publication
The document presents an artificial bee colony (ABC) algorithm to solve the NP-hard shortest common supersequence problem. The ABC algorithm is inspired by the foraging behavior of honey bees. It represents solutions as food sources and uses employed, onlooker, and scout bees to explore the search space. The algorithm calculates character frequencies in input strings to guide random supersequence generation. Fitness is evaluated by comparing sequences using a modified merge algorithm. Results show the ABC approach finds near-optimal solutions compared to other algorithms for solving shortest common supersequences.
An empirical study on customer engagement practices in tourism industryIAEME Publication
This document summarizes a research study on customer engagement practices in the tourism industry. The study examines how user-generated content (UGC) and online travel reviews (OTR) impact travelers' decisions and influence the rapid growth of the online travel industry. Through surveys, the study aims to determine the most engaging travel-related activities online for travelers and understand what factors travelers consider most important when planning vacations. The results can help online travel companies better understand how to utilize customer engagement and UGC.
Context Driven Technique for Document ClassificationIDES Editor
In this paper we present an innovative hybrid Text
Classification (TC) system that bridges the gap between
statistical and context based techniques. Our algorithm
harnesses contextual information at two stages. First it extracts
a cohesive set of keywords for each category by using lexical
references, implicit context as derived from LSA and wordvicinity
driven semantics. And secondly, each document is
represented by a set of context rich features whose values are
derived by considering both lexical cohesion as well as the extent
of coverage of salient concepts via lexical chaining. After
keywords are extracted, a subset of the input documents is
apportioned as training set. Its members are assigned categories
based on their keyword representation. These labeled
documents are used to train binary SVM classifiers, one for
each category. The remaining documents are supplied to the
trained classifiers in the form of their context-enhanced feature
vectors. Each document is finally ascribed its appropriate
category by an SVM classifier.
Enhanced Web Usage Mining Using Fuzzy Clustering and Collaborative Filtering ...inventionjournals
This document discusses an enhanced web usage mining system using fuzzy clustering and collaborative filtering recommendation algorithms. It aims to address challenges with existing recommender systems like producing low quality recommendations for large datasets. The system architecture uses fuzzy clustering to predict future user access based on browsing behavior. Collaborative filtering is then used to produce expected results by combining fuzzy clustering outputs with a web database. This approach aims to provide users with more relevant recommendations in a shorter time compared to other systems.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
IRJET-Model for semantic processing in information retrieval systemsIRJET Journal
This document proposes a model for semantic information retrieval that improves upon traditional keyword matching approaches. It involves three main components:
1. A crawling and indexing component that identifies websites and pages, extracts metadata, and generates a knowledge graph through semantic annotation.
2. A processing component that analyzes user queries and profiles to understand search intent, calculates semantic similarity between queries and indexed documents, and determines result relevance.
3. A presentation component that displays search results to users through both simple and advanced search interfaces, prioritizing the most relevant information based on the above processing.
The model is intended to address deficiencies in current Cuban web search by better understanding natural language queries and the contextual meaning of information through semantic technologies
Comparison of Semantic and Syntactic Information Retrieval System on the basi...Waqas Tariq
In this paper information retrieval system for local databases are discussed. The approach is to search the web both semantically and syntactically. The proposal handles the search queries related to the user who is interested in the focused results regarding a product with some specific characteristics. The objective of the work will be to find and retrieve the accurate information from the available information warehouse which contains related data having common keywords. This information retrieval system can eventually be used for accessing the internet also. Accuracy in information retrieval that is achieving both high precision and recall is difficult. So both semantic and syntactic search engine are compared for information retrieval using two parameters i.e. precision and recall.
Projection Multi Scale Hashing Keyword Search in Multidimensional DatasetsIRJET Journal
The document discusses a novel method called ProMiSH (Projection and Multi Scale Hashing) for keyword search in multi-dimensional datasets. ProMiSH uses random projection and hash-based index structures to achieve high scalability and speedup of more than four orders over state-of-the-art tree-based techniques. Empirical studies on real and synthetic datasets of sizes up to 10 million objects and 100 dimensions show ProMiSH scales linearly with dataset size, dimension, query size, and result size. The method groups objects embedded in a vector space that are tagged with keywords matching a given query.
Metadata: Towards Machine-Enabled Intelligencedannyijwest
World Wide Web has revolutionized the means of data availability, but with its current structure model , it is becoming increasingly difficult to retrieve relevant information, with reasonable precision and recall, using the major search engines. However, with use of metadata, combined with the use of improved searching techniques, helps to enhance relevant information retrieval .The design of structured, descriptions of Web resources enables greater search precision and a more accurate relevance ranking of retrieved information .One such efforts towards standardization is , Dublin Core standard, which has been developed as Metadata Standard and also other standards which enhances retrieval of a wide range of information resources. This paper discuses the importance of metadata, various metadata schemas and elements, and the need of standardization of Metadata. This paper further discusses how the metadata can be generated using various tools which assist intelligent agents for efficient retrieval
Information retrieval (IR) is the science of searching for documents and information within documents. IR is interdisciplinary and involves computer science, mathematics, psychology and other fields. Information storage and retrieval (ISAR) systems allow users to store, manipulate and analyze data and report results on a regular basis. ISAR is widely used in fields like science, business and healthcare. The information retrieval process begins with a user entering a query, which is then matched to objects stored in the database to retrieve relevant results based on scoring algorithms. Performance is evaluated using precision and recall metrics.
Intelligent Semantic Web Search Engines: A Brief Survey dannyijwest
The World Wide Web (WWW) allows the people to share the information (data) from the large database repositories globally. The amount of information grows billions of databases. We need to search the information will specialize tools known generically search engine. There are many of search engines available today, retrieving meaningful information is difficult. However to overcome this problem in search engines to retrieve meaningful information intelligently, semantic web technologies are playing a major role. In this paper we present survey on the search engine generations and the role of search engines in intelligent web and semantic search technologies.
The document discusses the history and evolution of intelligent information retrieval and web search engines. It covers early concepts from the 1970s of intelligent information retrieval that focused on using intelligent agents. However, this concept did not work well and missed the key aspects of intelligence in information retrieval systems. True intelligence in information retrieval requires appropriate allocation of responsibilities among all actors in the system and effective support of people in their interactions with information. The document also discusses the evolution of web search engines from early keyword-based systems to modern approaches that focus more on user experience, engagement, and quality over quantity.
This document provides an overview of an information retrieval system (IRS). It defines IRS as obtaining relevant information from a collection to meet a user's need. The IRS has three main components: a document subsystem for acquiring, representing and organizing data; a user subsystem for representing queries; and a search/retrieval subsystem for matching queries to documents. It describes the basic concepts like how a user enters a query that is scored and ranked to return relevant results, which can be iterated. The objectives are to highlight probabilistic models and establish relationships between popular techniques. The functions are to analyze information sources and queries to match and retrieve relevant items. [/SUMMARY]
Semantics in Financial Services -David NewmanPeter Berger
David Newman serves as a Senior Architect in the Enterprise Architecture group at Wells Fargo Bank. He has been following semantic technology for the last 3 years; and has developed several business ontologies. He has been instrumental in thought leadership at Wells Fargo on the application of Semantic Technology and is a representative of the Financial Services Technology Consortium (FSTC)on the W3C SPARQL Working Group.
Comparative Study on Graph-based Information Retrieval: the Case of XML DocumentIJAEMSJORNAL
The processing of massive amounts of data has become indispensable especially with the potential proliferation of big data. The volume of information available nowadays makes it difficult for the user to find relevant information in a vast collection of documents. As a result, the exploitation of vast document collections necessitates the implementation of automated technologies that enable appropriate and effective retrieval. In this paper, we will examine the state of the art of IR in XML documents. We will also discuss some works that have used graphs to represent documents in the context of IR. In the same vein, the relationships between the components of a graph are the center of our attention.
IRJET - Re-Ranking of Google Search ResultsIRJET Journal
This document summarizes a research paper that proposes a hybrid personalized re-ranking approach to search results. It models a user's search interests using a conceptual user profile containing categories and concepts extracted from clicked results and a concept hierarchy. The user profile contains two types of documents - taxonomy documents representing general interests and viewed documents representing specific interests. A hybrid re-ranking process then semantically integrates the user's general and specific interests from their profile with search engine rankings to improve result relevance.
The document discusses a framework for web information retrieval using automatic multi-document summarization. It proposes using multi-level document summarization to enhance the effectiveness of web information retrieval by supporting indexing and ranking of retrieved documents with an intelligent decision making system based on fuzzy inference rules. The paper tests the approach on CACM test data and finds that information retrieval results can be improved after performing a multi-document summarization process.
IRJET-Computational model for the processing of documents and support to the ...IRJET Journal
This document proposes a computational model for processing documents and supporting decision making in information retrieval systems. The model includes five main components: 1) a tracking and indexing component to crawl the web and store document metadata, 2) an information processing component to categorize documents and define user profiles, 3) a decision support component to analyze stored information and generate statistical reports, 4) a display component to provide search interfaces and visualization tools, and 5) specialized roles to administer the system. The goal of the model is to provide a framework for developing large-scale search engines.
This document provides an overview of text mining and web mining. It defines data mining and describes the common data mining tasks of classification, clustering, association rule mining and sequential pattern mining. It then discusses text mining, defining it as the process of analyzing unstructured text data to extract meaningful information and structure. The document outlines the seven practice areas of text mining as search/information retrieval, document clustering, document classification, web mining, information extraction, natural language processing, and concept extraction. It provides brief descriptions of the problems addressed within each practice area.
An improved technique for ranking semantic associationst07IJwest
The primary focus of the search techniques in the first generation of the Web is accessing relevant
documents from the Web. Though it satisfies user requirements, but it is insufficient as the user sometimes
wishes to access actionable information involvin
g complex relationships between two given entities.
Finding such complex relationships (also known as semantic associations) is especially useful in
applications
such as
National Security, Pharmacy, Business Intelligence etc. Therefore the next frontier is
discovering relevant semantic associations between two entities present in large semantic metadata
repositories. Given two entities, there exist a huge number of semantic associations between two entities.
Hence ranking of these associations is required i
n order to find more relevant associations. For this
Aleman Meza et al. proposed a method involving six metrics viz. context, subsumption, rarity, popularity,
association length and trust. To compute the overall rank of the associations this method compute
s
context, subsumption, rarity and popularity values for each component of the association and for all the
associations. However it is obvious that, many components appears repeatedly in many associations
therefore it is not necessary to compute context, s
ubsumption, rarity
,
popularity
,
and
trust
values of the
components every time for each association rather the previously computed values may be used while
computing the overall rank of the associations. Thi
s paper proposes a method to re
use the previously
computed values using a hash data structure thus reduce the execution time. To demonstrate the
effectiveness of the proposed method, experiments were conducted on SWETO ontology. Results show
that the proposed method is more efficient than the other existi
ng methods
.
Solving np hard problem using artificial bee colony algorithmIAEME Publication
The document presents an artificial bee colony (ABC) algorithm to solve the NP-hard shortest common supersequence problem. The ABC algorithm is inspired by the foraging behavior of honey bees. It represents solutions as food sources and uses employed, onlooker, and scout bees to explore the search space. The algorithm calculates character frequencies in input strings to guide random supersequence generation. Fitness is evaluated by comparing sequences using a modified merge algorithm. Results show the ABC approach finds near-optimal solutions compared to other algorithms for solving shortest common supersequences.
An empirical study on customer engagement practices in tourism industryIAEME Publication
This document summarizes a research study on customer engagement practices in the tourism industry. The study examines how user-generated content (UGC) and online travel reviews (OTR) impact travelers' decisions and influence the rapid growth of the online travel industry. Through surveys, the study aims to determine the most engaging travel-related activities online for travelers and understand what factors travelers consider most important when planning vacations. The results can help online travel companies better understand how to utilize customer engagement and UGC.
Numerical modeling of reinforced soil segmental wall under surcharge loadingIAEME Publication
This document summarizes a numerical study of a reinforced soil segmental wall under surcharge loading. The study uses finite element analysis to model the wall's performance considering different parameters like backfill and foundation soil properties, reinforcement, and surcharge loading. Parametric analyses were conducted to understand the effects of soil stiffness, Poisson's ratio, and surcharge on the wall's deformation. Design charts were developed relating backfill and foundation stiffness to deformation limits to aid wall designers. The finite element software developed and validated was used to conduct the numerical analyses.
This document summarizes the seismic evaluation and retrofit assessment of an old hospital building in Srinagar, India. The hospital was constructed in 1889 and has experienced structural deterioration over time. The summary evaluates the building, finds it unsafe under seismic shear stresses, and recommends retrofit measures like adding seismic belts, grouting cracks, concrete jacketing of walls, and foundation strengthening to improve the building's seismic performance.
Dropping based contention resolution for service differentiation to provide qo sIAEME Publication
This document discusses an approach for providing quality of service (QoS) in optical burst switching (OBS) networks through service differentiation. It proposes a dropping-based contention resolution algorithm that aims to reduce the loss of high priority data, specifically most high priority (MHP) data bursts that have tight delay constraints. The algorithm allows MHP data bursts to preempt channels allocated to lower priority data if resource thresholds are reached. It is evaluated through simulations to study the dropping probability behavior of different priority data bursts under increasing load. The performance is compared to an existing algorithm. The proposed approach does not require independent wavelength assignment or clock synchronization across nodes like some other approaches.
This document summarizes a research study on the relationship between working capital management and profitability among Indian manufacturing firms. The study uses financial data from 1,198 manufacturing firms over a 5-year period. The study aims to analyze how variables affecting working capital management influence firm profitability. Specifically, it examines the impact of days of debtors, days of inventory, days of creditors, and cash conversion cycle on profitability. The results of the study provide insights for Indian manufacturing firms on managing working capital to improve profitability.
1) The document describes a regenerative loading system that uses a converter to emulate various active and reactive power loads and regenerate consumed power back to the grid.
2) The system uses vector control to independently control active and reactive power by controlling the quadrature and direct axis current components.
3) A phase locked loop and transformations between stationary and synchronously rotating reference frames are used to generate control signals for the converter.
This document summarizes and reviews different techniques for video inpainting. It begins by defining video inpainting and distinguishing it from image inpainting due to additional temporal factors that must be considered. It then categorizes and reviews three main approaches: PDE-based methods, texture synthesis methods, and patch-based methods. For each approach, one or two influential works are described, focusing on the techniques and limitations. The review concludes by noting that while progress has been made, developing a video inpainting technique that can ensure both spatial and temporal consistency remains a challenge.
Durability studies on high strength high performance concrete 2IAEME Publication
The document discusses durability studies on high strength high performance concrete. Three key findings are:
1) Initial surface absorption values for the mixes studied were generally low, below 0.17 ml/m2/s after 10 minutes, indicating dense microstructure limiting water ingress.
2) For a given microsilica content, there exists an optimum superplasticizer dosage that yields best durability results, and this dosage increases with increasing microsilica.
3) At the same powder content, mixes with higher cement content exhibited lower absorption and permeability, emphasizing the complex relationship between cement, microsilica and superplasticizer quantities.
This document summarizes a research paper that proposes a framework for personalized web search using query log and clickthrough data. The framework implements a re-ranking approach that combines user search context and browsing behavior to generate personalized search results with high relevance. The framework consists of five components: a request handler, query processor, result handler, event handler, and response handler. The result handler applies a re-ranking approach using query log and clickthrough data to personalize search results before returning them to the user. An evaluation found the framework and re-ranking approach to be effective for personalized web search and information retrieval.
This document summarizes a research paper that proposes a novel framework for personalized web search using query log and clickthrough data. The framework implements a re-ranking approach to generate personalized search results with high relevance. It derives an extended set of user preferences and concepts based on extracted data from query logs and clickthrough information. An evaluation found the framework and re-ranking approach to be highly effective for personalized search and information retrieval.
The document proposes a text mining template-based algorithm to improve business intelligence by categorizing text. It begins with an introduction to data mining and text mining. It then discusses related work on text mining algorithms. The document proposes a methodology using a configuration file to identify fields in documents based on regular expressions. The algorithm reads documents line by line, matches lines to conditions in the configuration file, and stores the identified fields in an array. The array is then used to populate a structured table for analysis to improve decision making. The methodology is experimentally tested on 100 resumes to select candidates, demonstrating its ability to extract structured data from unstructured documents.
A Survey on Automatically Mining Facets for Queries from their Search ResultsIRJET Journal
This document summarizes research on automatically mining query facets from search results. Query facets provide useful summaries of a query by grouping related terms and phrases. The document reviews existing methods for query recommendation and facet extraction. It also proposes an unsupervised technique to mine query facets from top search results without additional domain knowledge. The technique aims to help users better understand queries and explore information through faceted search.
Classification-based Retrieval Methods to Enhance Information Discovery on th...IJMIT JOURNAL
The widespread adoption of the World-Wide Web (the Web) has created challenges both for society as a whole and for the technology used to build and maintain the Web. The ongoing struggle of information retrieval systems is to wade through this vast pile of data and satisfy users by presenting them with information that most adequately it’s their needs. On a societal level, the Web is expanding faster than we can comprehend its implications or develop rules for its use. The ubiquitous use of the Web has raised important social concerns in the areas of privacy, censorship, and access to information. On a technical level, the novelty of the Web and the pace of its growth have created challenges not only in the development of new applications that realize the power of the Web, but also in the technology needed to scale applications to accommodate the resulting large data sets and heavy loads. This thesis presents searching algorithms and hierarchical classification techniques for increasing a search service's understanding of web queries. Existing search services rely solely on a query's occurrence in the document collection to locate relevant documents. They typically do not perform any task or topic-based analysis of queries using other available resources, and do not leverage changes in user query patterns over time. Provided within are a set of techniques and metrics for performing temporal analysis on query logs. Our log analyses are shown to be reasonable and informative, and can be used to detect changing trends and patterns in the query stream, thus providing valuable data to a search service.
Structured and Unstructured Information Extraction Using Text Mining and Natu...rahulmonikasharma
Information on web is increasing at infinitum. Thus, web has become an unstructured global area where information even if available, cannot be directly used for desired applications. One is often faced with an information overload and demands for some automated help. Information extraction (IE) is the task of automatically extracting structured information from unstructured and/or semi-structured machine-readable documents by means of Text Mining and Natural Language Processing (NLP) techniques. Extracted structured information can be used for variety of enterprise or personal level task of varying complexity. The Information Extraction (IE) in also a set of knowledge in order to answer to user consultations using natural language. The system is based on a Fuzzy Logic engine, which takes advantage of its flexibility for managing sets of accumulated knowledge. These sets may be built in hierarchic levels by a tree structure. Information extraction is structured data or knowledge from unstructured text by identifying references to named entities as well as stated relationships between such entities. Data mining research assumes that the information to be “mined” is already in the form of a relational database. IE can serve an important technology for text mining. The knowledge discovered is expressed directly in the documents to be mined, then IE alone can serve as an effective approach to text mining. However, if the documents contain concrete data in unstructured form rather than abstract knowledge, it may be useful to first use IE to transform the unstructured data in the document corpus into a structured database, and then use traditional data mining tools to identify abstract patterns in this extracted data. We propose a novel method for text mining with natural language processing techniques to extract the information from data base with efficient way, where the extraction time and accuracy is measured and plotted with simulation. Where the attributes of entities and relationship entities from structured and semi structured information .Results are compared with conventional methods.
`A Survey on approaches of Web Mining in Varied Areasinventionjournals
There has been lot of research in recent years for efficient web searching. Several papers have proposed algorithm for user feedback sessions, to evaluate the performance of inferring user search goals. When the information is retrieved, user clicks on a particular URL. Based on the click rate, ranking will be done automatically, clustering the feedback sessions. Web search engines have made enormous contributions to the web and society. They make finding information on the web quick and easy. However, they are far from optimal. A major deficiency of generic search engines is that they follow the ‘‘one size fits all’’ model and are not adaptable to individual users.
A novel method to search information through multi agent search and retrieIAEME Publication
The document proposes a novel method for searching information through multi-agent search and retrieval that uses both content and context-based search. It describes a system that accepts two text inputs, processes them according to whether desktop or internet search is selected, and provides relevant results through indexing and multi-agents, with content search performed on Hadoop for increased performance. The system aims to provide faster and more accurate search results by filtering out irrelevant results.
Decision Support for E-Governance: A Text Mining ApproachIJMIT JOURNAL
Information and communication technology has the capability to improve the process by which governments involve citizens in formulating public policy and public projects. Even though much of government regulations may now be in digital form (and often available online), due to their complexity and diversity, identifying the ones relevant to a particular context is a non-trivial task. Similarly, with the advent of a number of electronic online forums, social networking sites and blogs, the opportunity of gathering citizens’ petitions and stakeholders’ views on government policy and proposals has increased greatly, but the volume and the complexity of analyzing unstructured data makes this difficult. On the other hand, text mining has come a long way from simple keyword search, and matured into a discipline capable of dealing with much more complex tasks. In this paper we discuss how text-mining techniques can help in retrieval of information and relationships from textual data sources, thereby assisting policy makers in discovering associations between policies and citizens’ opinions expressed in electronic public forums and blogs etc. We also present here, an integrated text mining based architecture for e-governance decision support along with a discussion on the Indian scenario.
A Simple Information Retrieval Techniqueidescitation
The document presents a simple information retrieval technique that involves removing stop words and punctuation from documents, calculating term frequency and inverse document frequency, constructing a master document matrix, and ranking documents based on similarity to user queries. The technique is demonstrated on a sample collection of 5 documents. For a query on "information retrieval system", the documents are ranked from most similar to least similar as document 5, document 1, document 2, document 3, document 4. The technique provides an easy way to search and retrieve relevant documents from a collection.
The document describes a proposed framework called UPS for privacy-preserving personalized web search. The UPS framework aims to generalize user profiles for each query according to user-specified privacy requirements, while balancing privacy risk and personalization utility. Two key contributions of the proposed system are: 1) Supporting runtime profiling to dynamically generalize user profiles on a per-query basis; and 2) Allowing for customization of privacy requirements by users to designate sensitive topics in their profiles. Algorithms are proposed to generalize profiles to optimize these metrics during the personalization process.
An effective pre processing algorithm for information retrieval systemsijdms
The Internet is probably the most successful distributed computing system ever. However, our capabilities
for data querying and manipulation on the internet are primordial at best. The user expectations are
enhancing over the period of time along with increased amount of operational data past few decades. The
data-user expects more deep, exact, and detailed results. Result retrieval for the user query is always
relative o the pattern of data storage and index. In Information retrieval systems, tokenization is an
integrals part whose prime objective is to identifying the token and their count. In this paper, we have
proposed an effective tokenization approach which is based on training vector and result shows that
efficiency/ effectiveness of proposed algorithm. Tokenization on documents helps to satisfy user’s
information need more precisely and reduced search sharply, is believed to be a part of information
retrieval. Pre-processing of input document is an integral part of Tokenization, which involves preprocessing
of documents and generates its respective tokens which is the basis of these tokens probabilistic
IR generate its scoring and gives reduced search space. The comparative analysis is based on the two
parameters; Number of Token generated, Pre-processing time.
The document provides an introduction to information retrieval, including its history, key concepts, and challenges. It discusses how information retrieval aims to retrieve relevant documents from a collection to satisfy a user's information need. The main challenge in information retrieval is determining relevance, as relevance depends on personal assessment, task, context, time, location, and device. Three main issues in information retrieval are determining relevance, representing documents and queries, and developing effective retrieval models and algorithms.
The document provides an introduction to information retrieval, including its history, key concepts, and challenges. It discusses how information retrieval aims to retrieve relevant documents from a collection to satisfy a user's information need. The main challenge in information retrieval is determining relevance, as relevance depends on personal assessment and can change based on context, time, location, and device. The document outlines the major issues and developments in the field over time from the 1950s to present day.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
This document summarizes a research paper on developing user profiles from search engine queries to enable personalized search results. It discusses how current search engines generally return the same results regardless of individual user interests. The paper proposes methods to construct user profiles capturing both positive and negative preferences from search histories and click-through data. Experimental results showed profiles including both preferences performed best by improving query clustering and separating similar vs. dissimilar queries. Future work aims to use profiles for collaborative filtering and predicting new query intents.
UML MODELING AND SYSTEM ARCHITECTURE FOR AGENT BASED INFORMATION RETRIEVALijcsit
In this current technological era, there is an enormous increase in the information available on web and
also in the online databases. This information abundance increases the complexity of finding relevant
information. To solve such challenges, there is a need for improved and intelligent systems for efficient
search and retrieval. Intelligent Agents can be used for better search and information retrieval in a
document collection. The information required by a user is scattered in a large number of databases. In this
paper, the object oriented modeling for agent based information retrieval system is presented. The paper
also discusses the framework of agent architecture for obtaining the best combination terms that serve as
an input query to the information retrieval system. The communication and cooperation among the agents
are also explained. Each agent has a task to perform in information retrieval.
An Improved Annotation Based Summary Generation For Unstructured DataMelinda Watson
This document discusses annotation-based summarization of unstructured data. It begins with an introduction to annotation and information retrieval. Current annotation processes cannot maintain modifications due to frequent document updates. The document then reviews literature on automatic text classification, applying annotations to linked open data sets, and using domain ontologies for automatic document annotation. Keywords, sentences and contexts are extracted from documents for annotation. Different annotation models are discussed. The goal is to develop an improved annotation approach for summarizing unstructured data that can handle frequent document changes.
Enhancing the Privacy Protection of the User Personalized Web Search Using RDFIJTET Journal
Abstract— Personalized searches refers to search experiences that are tailored specifically to an individual's interest by incorporating information about the individual beyond specific query provided. User may not aware of some privacy issues in search results where personalized and wonder why things that are interested in have become so relevant. Such irrelevance is largely due to the enormous variety of user’s contexts and backgrounds, as well as the ambiguity of texts. In contrast, Profile-based methods can be potentially effective for almost all sorts of queries, but are reported to be unstable under some circumstances. The amount of structured data available on the web has been increasing rapidly, especially RDF data. This proliferation of RDF data can also be attributed to the generality of the underlying graph-structured model, i.e., many types of data can be expressed in this format including relational and XML data. For a Personalized Semantic Web Search the semi structured data should be indexed with RDF. This proposed RDF technique not only enhances the privacy and security of the user profile and optimizes query for efficient filtering of data. The user profile access is been avoided by means of placing a proxy in the client side, so profile exposure avoided. The proxy generates a random profile at each time. The contents will be sent back to the proxy and only the relevant contents will be sent over to the client. In this RDF framework the queries are semi structured for personalized web search.
Performance Evaluation of Query Processing Techniques in Information Retrievalidescitation
The first element of the search process is the query.
The user query being on an average restricted to two or three
keywords makes the query ambiguous to the search engine.
Given the user query, the goal of an Information Retrieval
[IR] system is to retrieve information which might be useful
or relevant to the information need of the user. Hence, the
query processing plays an important role in IR system.
The query processing can be divided into four categories
i.e. query expansion, query optimization, query classification and
query parsing. In this paper an attempt is made to evaluate the
performance of query processing algorithms in each of the
category. The evaluation was based on dataset as specified by
Forum for Information Retrieval [FIRE15]. The criteria used
for evaluation are precision and relative recall. The analysis is
based on the importance of each step in query processing. The
experimental results show that the significance of each step
in query processing and also the relevance of web semantics
and spelling correction in the user query.
This document proposes a BOT virtual guide that will extract educational web content based on topics recently taught using web crawling techniques. It will use a domain ontology, DOM parsing, and concept-focused crawling to find relevant documents from the web. The documents will be ranked based on their concept similarity to the topic. The filtered and crawled data will then be provided to students as speech output through a text-to-speech system to serve as an automated virtual guide for supplemental learning materials.
Similar to A survey on various architectures, models and methodologies for information retrieval (20)
Submission Deadline: 30th September 2022
Acceptance Notification: Within Three Days’ time period
Online Publication: Within 24 Hrs. time Period
Expected Date of Dispatch of Printed Journal: 5th October 2022
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...IAEME Publication
White layer thickness (WLT) formed and surface roughness in wire electric discharge turning (WEDT) of tungsten carbide composite has been made to model through response surface methodology (RSM). A Taguchi’s standard Design of experiments involving five input variables with three levels has been employed to establish a mathematical model between input parameters and responses. Percentage of cobalt content, spindle speed, Pulse on-time, wire feed and pulse off-time were changed during the experimental tests based on the Taguchi’s orthogonal array L27 (3^13). Analysis of variance (ANOVA) revealed that the mathematical models obtained can adequately describe performance within the parameters of the factors considered. There was a good agreement between the experimental and predicted values in this study.
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSIAEME Publication
The study explores the reasons for a transgender to become entrepreneurs. In this study transgender entrepreneur was taken as independent variable and reasons to become as dependent variable. Data were collected through a structured questionnaire containing a five point Likert Scale. The study examined the data of 30 transgender entrepreneurs in Salem Municipal Corporation of Tamil Nadu State, India. Simple Random sampling technique was used. Garrett Ranking Technique (Percentile Position, Mean Scores) was used as the analysis for the present study to identify the top 13 stimulus factors for establishment of trans entrepreneurial venture. Economic advancement of a nation is governed upon the upshot of a resolute entrepreneurial doings. The conception of entrepreneurship has stretched and materialized to the socially deflated uncharted sections of transgender community. Presently transgenders have smashed their stereotypes and are making recent headlines of achievements in various fields of our Indian society. The trans-community is gradually being observed in a new light and has been trying to achieve prospective growth in entrepreneurship. The findings of the research revealed that the optimistic changes are taking place to change affirmative societal outlook of the transgender for entrepreneurial ventureship. It also laid emphasis on other transgenders to renovate their traditional living. The paper also highlights that legislators, supervisory body should endorse an impartial canons and reforms in Tamil Nadu Transgender Welfare Board Association.
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSIAEME Publication
Since ages gender difference is always a debatable theme whether caused by nature, evolution or environment. The birth of a transgender is dreadful not only for the child but also for their parents. The pain of living in the wrong physique and treated as second class victimized citizen is outrageous and fully harboured with vicious baseless negative scruples. For so long, social exclusion had perpetuated inequality and deprivation experiencing ingrained malign stigma and besieged victims of crime or violence across their life spans. They are pushed into the murky way of life with a source of eternal disgust, bereft sexual potency and perennial fear. Although they are highly visible but very little is known about them. The common public needs to comprehend the ravaged arrogance on these insensitive souls and assist in integrating them into the mainstream by offering equal opportunity, treat with humanity and respect their dignity. Entrepreneurship in the current age is endorsing the gender fairness movement. Unstable careers and economic inadequacy had inclined one of the gender variant people called Transgender to become entrepreneurs. These tiny budding entrepreneurs resulted in economic transition by means of employment, free from the clutches of stereotype jobs, raised standard of living and handful of financial empowerment. Besides all these inhibitions, they were able to witness a platform for skill set development that ignited them to enter into entrepreneurial domain. This paper epitomizes skill sets involved in trans-entrepreneurs of Thoothukudi Municipal Corporation of Tamil Nadu State and is a groundbreaking determination to sightsee various skills incorporated and the impact on entrepreneurship.
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSIAEME Publication
The banking and financial services industries are experiencing increased technology penetration. Among them, the banking industry has made technological advancements to better serve the general populace. The economy focused on transforming the banking sector's system into a cashless, paperless, and faceless one. The researcher wants to evaluate the user's intention for utilising a mobile banking application. The study also examines the variables affecting the user's behaviour intention when selecting specific applications for financial transactions. The researcher employed a well-structured questionnaire and a descriptive study methodology to gather the respondents' primary data utilising the snowball sampling technique. The study includes variables like performance expectations, effort expectations, social impact, enabling circumstances, and perceived risk. Each of the aforementioned variables has a major impact on how users utilise mobile banking applications. The outcome will assist the service provider in comprehending the user's history with mobile banking applications.
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSIAEME Publication
Technology upgradation in banking sector took the economy to view that payment mode towards online transactions using mobile applications. This system enabled connectivity between banks, Merchant and user in a convenient mode. there are various applications used for online transactions such as Google pay, Paytm, freecharge, mobikiwi, oxygen, phonepe and so on and it also includes mobile banking applications. The study aimed at evaluating the predilection of the user in adopting digital transaction. The study is descriptive in nature. The researcher used random sample techniques to collect the data. The findings reveal that mobile applications differ with the quality of service rendered by Gpay and Phonepe. The researcher suggest the Phonepe application should focus on implementing the application should be user friendly interface and Gpay on motivating the users to feel the importance of request for money and modes of payments in the application.
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOIAEME Publication
The prototype of a voice-based ATM for visually impaired using Arduino is to help people who are blind. This uses RFID cards which contain users fingerprint encrypted on it and interacts with the users through voice commands. ATM operates when sensor detects the presence of one person in the cabin. After scanning the RFID card, it will ask to select the mode like –normal or blind. User can select the respective mode through voice input, if blind mode is selected the balance check or cash withdraw can be done through voice input. Normal mode procedure is same as the existing ATM.
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IAEME Publication
There is increasing acceptability of emotional intelligence as a major factor in personality assessment and effective human resource management. Emotional intelligence as the ability to build capacity, empathize, co-operate, motivate and develop others cannot be divorced from both effective performance and human resource management systems. The human person is crucial in defining organizational leadership and fortunes in terms of challenges and opportunities and walking across both multinational and bilateral relationships. The growing complexity of the business world requires a great deal of self-confidence, integrity, communication, conflict and diversity management to keep the global enterprise within the paths of productivity and sustainability. Using the exploratory research design and 255 participants the result of this original study indicates strong positive correlation between emotional intelligence and effective human resource management. The paper offers suggestions on further studies between emotional intelligence and human capital development and recommends for conflict management as an integral part of effective human resource management.
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYIAEME Publication
Our life journey, in general, is closely defined by the way we understand the meaning of why we coexist and deal with its challenges. As we develop the "inspiration economy", we could say that nearly all of the challenges we have faced are opportunities that help us to discover the rest of our journey. In this note paper, we explore how being faced with the opportunity of being a close carer for an aging parent with dementia brought intangible discoveries that changed our insight of the meaning of the rest of our life journey.
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...IAEME Publication
The main objective of this study is to analyze the impact of aspects of Organizational Culture on the Effectiveness of the Performance Management System (PMS) in the Health Care Organization at Thanjavur. Organizational Culture and PMS play a crucial role in present-day organizations in achieving their objectives. PMS needs employees’ cooperation to achieve its intended objectives. Employees' cooperation depends upon the organization’s culture. The present study uses exploratory research to examine the relationship between the Organization's culture and the Effectiveness of the Performance Management System. The study uses a Structured Questionnaire to collect the primary data. For this study, Thirty-six non-clinical employees were selected from twelve randomly selected Health Care organizations at Thanjavur. Thirty-two fully completed questionnaires were received.
Living in 21st century in itself reminds all of us the necessity of police and its administration. As more and more we are entering into the modern society and culture, the more we require the services of the so called ‘Khaki Worthy’ men i.e., the police personnel. Whether we talk of Indian police or the other nation’s police, they all have the same recognition as they have in India. But as already mentioned, their services and requirements are different after the like 26th November, 2008 incidents, where they without saving their own lives has sacrificed themselves without any hitch and without caring about their respective family members and wards. In other words, they are like our heroes and mentors who can guide us from the darkness of fear, militancy, corruption and other dark sides of life and so on. Now the question arises, if Gandhi would have been alive today, what would have been his reaction/opinion to the police and its functioning? Would he have some thing different in his mind now what he had been in his mind before the partition or would he be going to start some Satyagraha in the form of some improvement in the functioning of the police administration? Really these questions or rather night mares can come to any one’s mind, when there is too much confusion is prevailing in our minds, when there is too much corruption in the society and when the polices working is also in the questioning because of one or the other case throughout the India. It is matter of great concern that we have to thing over our administration and our practical approach because the police personals are also like us, they are part and parcel of our society and among one of us, so why we all are pin pointing towards them.
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...IAEME Publication
The goal of this study was to see how talent management affected employee retention in the selected IT organizations in Chennai. The fundamental issue was the difficulty to attract, hire, and retain talented personnel who perform well and the gap between supply and demand of talent acquisition and retaining them within the firms. The study's main goals were to determine the impact of talent management on employee retention in IT companies in Chennai, investigate talent management strategies that IT companies could use to improve talent acquisition, performance management, career planning and formulate retention strategies that the IT firms could use. The respondents were given a structured close-ended questionnaire with the 5 Point Likert Scale as part of the study's quantitative research design. The target population consisted of 289 IT professionals. The questionnaires were distributed and collected by the researcher directly. The Statistical Package for Social Sciences (SPSS) was used to collect and analyse the questionnaire responses. Hypotheses that were formulated for the various areas of the study were tested using a variety of statistical tests. The key findings of the study suggested that talent management had an impact on employee retention. The studies also found that there is a clear link between the implementation of talent management and retention measures. Management should provide enough training and development for employees, clarify job responsibilities, provide adequate remuneration packages, and recognise employees for exceptional performance.
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...IAEME Publication
Globally, Millions of dollars were spent by the organizations for employing skilled Information Technology (IT) professionals. It is costly to replace unskilled employees with IT professionals possessing technical skills and competencies that aid in interconnecting the business processes. The organization’s employment tactics were forced to alter by globalization along with technological innovations as they consistently diminish to remain lean, outsource to concentrate on core competencies along with restructuring/reallocate personnel to gather efficiency. As other jobs, organizations or professions have become reasonably more appropriate in a shifting employment landscape, the above alterations trigger both involuntary as well as voluntary turnover. The employee view on jobs is also afflicted by the COVID-19 pandemic along with the employee-driven labour market. So, having effective strategies is necessary to tackle the withdrawal rate of employees. By associating Emotional Intelligence (EI) along with Talent Management (TM) in the IT industry, the rise in attrition rate was analyzed in this study. Only 303 respondents were collected out of 350 participants to whom questionnaires were distributed. From the employees of IT organizations located in Bangalore (India), the data were congregated. A simple random sampling methodology was employed to congregate data as of the respondents. Generating the hypothesis along with testing is eventuated. The effect of EI and TM along with regression analysis between TM and EI was analyzed. The outcomes indicated that employee and Organizational Performance (OP) were elevated by effective EI along with TM.
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...IAEME Publication
By implementing talent management strategy, organizations would have the option to retain their skilled professionals while additionally working on their overall performance. It is the course of appropriately utilizing the ideal individuals, setting them up for future top positions, exploring and dealing with their performance, and holding them back from leaving the organization. It is employee performance that determines the success of every organization. The firm quickly obtains an upper hand over its rivals in the event that its employees having particular skills that cannot be duplicated by the competitors. Thus, firms are centred on creating successful talent management practices and processes to deal with the unique human resources. Firms are additionally endeavouring to keep their top/key staff since on the off chance that they leave; the whole store of information leaves the firm's hands. The study's objective was to determine the impact of talent management on organizational performance among the selected IT organizations in Chennai. The study recommends that talent management limitedly affects performance. On the off chance that this talent is appropriately management and implemented properly, organizations might benefit as much as possible from their maintained assets to support development and productivity, both monetarily and non-monetarily.
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...IAEME Publication
Banking regulations act of India, 1949 defines banking as “acceptance of deposits for the purpose of lending or investment from the public, repayment on demand or otherwise and withdrawable through cheques, drafts order or otherwise”, the major participants of the Indian financial system are commercial banks, the financial institution encompassing term lending institutions. Investments institutions, specialized financial institution and the state level development banks, non banking financial companies (NBFC) and other market intermediaries such has the stock brokers and money lenders are among the oldest of the certain variants of NBFC and the oldest market participants. The asset quality of banks is one of the most important indicators of their financial health. The Indian banking sector has been facing severe problems of increasing Non- Performing Assets (NPAs). The NPAs growth directly and indirectly affects the quality of assets and profitability of banks. It also shows the efficiency of banks credit risk management and the recovery effectiveness. NPA do not generate any income, whereas, the bank is required to make provisions for such as assets that why is a double edge weapon. This paper outlines the concept of quality of bank loans of different types like Housing, Agriculture and MSME loans in state Haryana of selected public and private sector banks. This study is highlighting problems associated with the role of commercial bank in financing Small and Medium Scale Enterprises (SME). The overall objective of the research was to assess the effect of the financing provisions existing for the setting up and operations of MSMEs in the country and to generate recommendations for more robust financing mechanisms for successful operation of the MSMEs, in turn understanding the impact of MSME loans on financial institutions due to NPA. There are many research conducted on the topic of Non- Performing Assets (NPA) Management, concerning particular bank, comparative study of public and private banks etc. In this paper the researcher is considering the aggregate data of selected public sector and private sector banks and attempts to compare the NPA of Housing, Agriculture and MSME loans in state Haryana of public and private sector banks. The tools used in the study are average and Anova test and variance. The findings reveal that NPA is common problem for both public and private sector banks and is associated with all types of loans either that is housing loans, agriculture loans and loans to SMES. NPAs of both public and private sector banks show the increasing trend. In 2010-11 GNPA of public and private sector were at same level it was 2% but after 2010-11 it increased in many fold and at present there is GNPA in some more than 15%. It shows the dark area of Indian banking sector.
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...IAEME Publication
An experiment conducted in this study found that BaSO4 changed Nylon 6's mechanical properties. By changing the weight ratios, BaSO4 was used to make Nylon 6. This Researcher looked into how hard Nylon-6/BaSO4 composites are and how well they wear. Experiments were done based on Taguchi design L9. Nylon-6/BaSO4 composites can be tested for their hardness number using a Rockwell hardness testing apparatus. On Nylon/BaSO4, the wear behavior was measured by a wear monitor, pinon-disc friction by varying reinforcement, sliding speed, and sliding distance, and the microstructure of the crack surfaces was observed by SEM. This study provides significant contributions to ultimate strength by increasing BaSO4 content up to 16% in the composites, and sliding speed contributes 72.45% to the wear rate
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...IAEME Publication
The majority of the population in India lives in villages. The village is the back bone of the country. Village or rural industries play an important role in the national economy, particularly in the rural development. Developing the rural economy is one of the key indicators towards a country’s success. Whether it be the need to look after the welfare of the farmers or invest in rural infrastructure, Governments have to ensure that rural development isn’t compromised. The economic development of our country largely depends on the progress of rural areas and the standard of living of rural masses. Village or rural industries play an important role in the national economy, particularly in the rural development. Rural entrepreneurship is based on stimulating local entrepreneurial talent and the subsequent growth of indigenous enterprises. It recognizes opportunity in the rural areas and accelerates a unique blend of resources either inside or outside of agriculture. Rural entrepreneurship brings an economic value to the rural sector by creating new methods of production, new markets, new products and generate employment opportunities thereby ensuring continuous rural development. Social Entrepreneurship has the direct and primary objective of serving the society along with the earning profits. So, social entrepreneurship is different from the economic entrepreneurship as its basic objective is not to earn profits but for providing innovative solutions to meet the society needs which are not taken care by majority of the entrepreneurs as they are in the business for profit making as a sole objective. So, the Social Entrepreneurs have the huge growth potential particularly in the developing countries like India where we have huge societal disparities in terms of the financial positions of the population. Still 22 percent of the Indian population is below the poverty line and also there is disparity among the rural & urban population in terms of families living under BPL. 25.7 percent of the rural population & 13.7 percent of the urban population is under BPL which clearly shows the disparity of the poor people in the rural and urban areas. The need to develop social entrepreneurship in agriculture is dictated by a large number of social problems. Such problems include low living standards, unemployment, and social tension. The reasons that led to the emergence of the practice of social entrepreneurship are the above factors. The research problem lays upon disclosing the importance of role of social entrepreneurship in rural development of India. The paper the tendencies of social entrepreneurship in India, to present successful examples of such business for providing recommendations how to improve situation in rural areas in terms of social entrepreneurship development. Indian government has made some steps towards development of social enterprises, social entrepreneurship, and social in- novation, but a lot remains to be improved.
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...IAEME Publication
Distribution system is a critical link between the electric power distributor and the consumers. Most of the distribution networks commonly used by the electric utility is the radial distribution network. However in this type of network, it has technical issues such as enormous power losses which affect the quality of the supply. Nowadays, the introduction of Distributed Generation (DG) units in the system help improve and support the voltage profile of the network as well as the performance of the system components through power loss mitigation. In this study network reconfiguration was done using two meta-heuristic algorithms Particle Swarm Optimization and Gravitational Search Algorithm (PSO-GSA) to enhance power quality and voltage profile in the system when simultaneously applied with the DG units. Backward/Forward Sweep Method was used in the load flow analysis and simulated using the MATLAB program. Five cases were considered in the Reconfiguration based on the contribution of DG units. The proposed method was tested using IEEE 33 bus system. Based on the results, there was a voltage profile improvement in the system from 0.9038 p.u. to 0.9594 p.u.. The integration of DG in the network also reduced power losses from 210.98 kW to 69.3963 kW. Simulated results are drawn to show the performance of each case.
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...IAEME Publication
Manufacturing industries have witnessed an outburst in productivity. For productivity improvement manufacturing industries are taking various initiatives by using lean tools and techniques. However, in different manufacturing industries, frugal approach is applied in product design and services as a tool for improvement. Frugal approach contributed to prove less is more and seems indirectly contributing to improve productivity. Hence, there is need to understand status of frugal approach application in manufacturing industries. All manufacturing industries are trying hard and putting continuous efforts for competitive existence. For productivity improvements, manufacturing industries are coming up with different effective and efficient solutions in manufacturing processes and operations. To overcome current challenges, manufacturing industries have started using frugal approach in product design and services. For this study, methodology adopted with both primary and secondary sources of data. For primary source interview and observation technique is used and for secondary source review has done based on available literatures in website, printed magazines, manual etc. An attempt has made for understanding application of frugal approach with the study of manufacturing industry project. Manufacturing industry selected for this project study is Mahindra and Mahindra Ltd. This paper will help researcher to find the connections between the two concepts productivity improvement and frugal approach. This paper will help to understand significance of frugal approach for productivity improvement in manufacturing industry. This will also help to understand current scenario of frugal approach in manufacturing industry. In manufacturing industries various process are involved to deliver the final product. In the process of converting input in to output through manufacturing process productivity plays very critical role. Hence this study will help to evolve status of frugal approach in productivity improvement programme. The notion of frugal can be viewed as an approach towards productivity improvement in manufacturing industries.
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTIAEME Publication
In this paper, we investigated a queuing model of fuzzy environment-based a multiple channel queuing model (M/M/C) ( /FCFS) and study its performance under realistic conditions. It applies a nonagonal fuzzy number to analyse the relevant performance of a multiple channel queuing model (M/M/C) ( /FCFS). Based on the sub interval average ranking method for nonagonal fuzzy number, we convert fuzzy number to crisp one. Numerical results reveal that the efficiency of this method. Intuitively, the fuzzy environment adapts well to a multiple channel queuing models (M/M/C) ( /FCFS) are very well.