IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
IRJET- Semantic Web Mining and Semantic Search Engine: A ReviewIRJET Journal
This document provides an overview of the semantic web, semantic web mining, and semantic search engines. It discusses how the semantic web aims to make web data machine-readable through technologies like RDF and ontology. Semantic web mining involves extracting useful knowledge from the semantic web. Semantic search engines then allow users to retrieve more precise and meaningful data from the semantic web through the use of semantic technologies. The document outlines challenges for semantic search engines and opportunities for further research.
IRJET- Structuring Mobile Application for Retrieving Book Data Utilizing Opti...IRJET Journal
This document describes a mobile application that allows users to retrieve book data by taking a photo of the book cover with their Android smartphone camera. The application uses optical character recognition (OCR) technology to extract text from the image. It then separates each word and compares it to words in a database of book titles. The application outputs five potential book titles that match words from the extracted text. This allows users to easily obtain book detail information without the text being readable to the naked eye. The document discusses the system design, related work, proposed method, and experimental results of the application.
Transcoding: A Technique to Transform Digital ContentVideoguy
This document discusses transcoding, which is the process of adapting or customizing digital content for different user environments and preferences. Transcoding involves modifying content without changing its representation, such as selecting a subset for small screens, or translating content to a different representation, such as converting HTML to WML. It allows content to be delivered to a wide range of devices and users by accounting for their constraints and preferences.
The document discusses NHibernate, an open source object-relational mapping framework for .NET. It begins by describing some of the limitations of using ADO.NET datasets for data access and how NHibernate provides a more object-oriented approach. It then provides steps to get started with NHibernate, including configuring NHibernate, defining a domain model, mapping the domain model to database tables, and generating the necessary code.
A language independent web data extraction using vision based page segmentati...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Although of the semantic web technologies utilization in the learning development field is a new research area, some authors have already proposed their idea of how an effective that operate. Specifically, from analysis of the literature in the field, we have identified three different types of existing applications that actually employ these technologies to support learning. These applications aim at: Enhancing the learning objects reusability by linking them to an ontological description of the domain, or, more generally, describe relevant dimension of the learning process in an ontology, then; providing a comprehensive authoring system to retrieve and organize web material into a learning course, and constructing advanced strategies to present annotated resources to the user, in the form of browsing facilities, narrative generation and final rendering of a course. On difference with the approaches cited above, here we propose an approach that is modeled on narrative studies and on their transposition in the digital world. In the rest of the paper, we present the theoretical basis that inspires this approach, and show some examples that are guiding our implementation and testing of these ideas within e-learning. By emerging the idea of the ontologies are recognized as the most important component in achieving semantic interoperability of e-learning resources. The benefits of their use have already been recognized in the learning technology community. In order to better define different aspects of ontology applications in e-learning, researchers have given several classifications of ontologies. We refer to a general one given in that differentiates between three dimensions ontologies can describe: content, context, and structure. Most of the present research has been dedicated to the first group of ontologies. A well-known example of such an ontology is based on the ACM Computer Classification System (ACM CCS) and defined by Resource Description Framework Schema (RDFS). It’s used in the MOODLE to classify learning objects with a goal to improve searching. The chapter will cover the terms of the semantic web and e-learning systems design and management in e-learning (MOODLE) and some of studies depend on e-learning and semantic web, thus the tools will be used in this paper, and lastly we shall discuss the expected contribution. The special attention will be putted on the above topics.
International Journal of Computational Engineering Research(IJCER) ijceronline
nternational Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Semantics in Financial Services -David NewmanPeter Berger
David Newman serves as a Senior Architect in the Enterprise Architecture group at Wells Fargo Bank. He has been following semantic technology for the last 3 years; and has developed several business ontologies. He has been instrumental in thought leadership at Wells Fargo on the application of Semantic Technology and is a representative of the Financial Services Technology Consortium (FSTC)on the W3C SPARQL Working Group.
IRJET- Semantic Web Mining and Semantic Search Engine: A ReviewIRJET Journal
This document provides an overview of the semantic web, semantic web mining, and semantic search engines. It discusses how the semantic web aims to make web data machine-readable through technologies like RDF and ontology. Semantic web mining involves extracting useful knowledge from the semantic web. Semantic search engines then allow users to retrieve more precise and meaningful data from the semantic web through the use of semantic technologies. The document outlines challenges for semantic search engines and opportunities for further research.
IRJET- Structuring Mobile Application for Retrieving Book Data Utilizing Opti...IRJET Journal
This document describes a mobile application that allows users to retrieve book data by taking a photo of the book cover with their Android smartphone camera. The application uses optical character recognition (OCR) technology to extract text from the image. It then separates each word and compares it to words in a database of book titles. The application outputs five potential book titles that match words from the extracted text. This allows users to easily obtain book detail information without the text being readable to the naked eye. The document discusses the system design, related work, proposed method, and experimental results of the application.
Transcoding: A Technique to Transform Digital ContentVideoguy
This document discusses transcoding, which is the process of adapting or customizing digital content for different user environments and preferences. Transcoding involves modifying content without changing its representation, such as selecting a subset for small screens, or translating content to a different representation, such as converting HTML to WML. It allows content to be delivered to a wide range of devices and users by accounting for their constraints and preferences.
The document discusses NHibernate, an open source object-relational mapping framework for .NET. It begins by describing some of the limitations of using ADO.NET datasets for data access and how NHibernate provides a more object-oriented approach. It then provides steps to get started with NHibernate, including configuring NHibernate, defining a domain model, mapping the domain model to database tables, and generating the necessary code.
A language independent web data extraction using vision based page segmentati...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Although of the semantic web technologies utilization in the learning development field is a new research area, some authors have already proposed their idea of how an effective that operate. Specifically, from analysis of the literature in the field, we have identified three different types of existing applications that actually employ these technologies to support learning. These applications aim at: Enhancing the learning objects reusability by linking them to an ontological description of the domain, or, more generally, describe relevant dimension of the learning process in an ontology, then; providing a comprehensive authoring system to retrieve and organize web material into a learning course, and constructing advanced strategies to present annotated resources to the user, in the form of browsing facilities, narrative generation and final rendering of a course. On difference with the approaches cited above, here we propose an approach that is modeled on narrative studies and on their transposition in the digital world. In the rest of the paper, we present the theoretical basis that inspires this approach, and show some examples that are guiding our implementation and testing of these ideas within e-learning. By emerging the idea of the ontologies are recognized as the most important component in achieving semantic interoperability of e-learning resources. The benefits of their use have already been recognized in the learning technology community. In order to better define different aspects of ontology applications in e-learning, researchers have given several classifications of ontologies. We refer to a general one given in that differentiates between three dimensions ontologies can describe: content, context, and structure. Most of the present research has been dedicated to the first group of ontologies. A well-known example of such an ontology is based on the ACM Computer Classification System (ACM CCS) and defined by Resource Description Framework Schema (RDFS). It’s used in the MOODLE to classify learning objects with a goal to improve searching. The chapter will cover the terms of the semantic web and e-learning systems design and management in e-learning (MOODLE) and some of studies depend on e-learning and semantic web, thus the tools will be used in this paper, and lastly we shall discuss the expected contribution. The special attention will be putted on the above topics.
International Journal of Computational Engineering Research(IJCER) ijceronline
nternational Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Semantics in Financial Services -David NewmanPeter Berger
David Newman serves as a Senior Architect in the Enterprise Architecture group at Wells Fargo Bank. He has been following semantic technology for the last 3 years; and has developed several business ontologies. He has been instrumental in thought leadership at Wells Fargo on the application of Semantic Technology and is a representative of the Financial Services Technology Consortium (FSTC)on the W3C SPARQL Working Group.
This document summarizes a research paper that analyzes the movement of metallic particles in gas insulated busduct systems due to electric fields. It presents analytical, finite difference, and finite element methods to model the electric fields and calculate particle trajectories. Simulation results show that aluminum and copper particle movement increases with higher voltages. The maximum radial movement is 10.75mm at 220kV and 35.66mm at 600kV using different field calculation methods.
This document summarizes a research paper that proposes a new DSP controller-based switching configuration for a hybrid distributed energy system using a single input DC-DC buck-boost converter. The system allows two renewable energy sources, such as solar and wind, to supply load either separately or simultaneously based on availability. The DSP controller is used to extract maximum energy from the available source and improve system efficiency. The system can operate in three modes depending on source availability to maximize efficiency. Simulation results using MATLAB/Simulink are presented and analyzed.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document investigates the abrasive wear behavior of coir fiber reinforced epoxy composites using the Taguchi method. Composites were made with 10%, 20%, and 30% treated and untreated coir fiber. Abrasion tests were conducted at loads of 10-30N, speeds of 300-500rpm. Treated fiber composites showed better wear resistance than untreated. Wear rate decreased with increasing fiber content and increased with increasing load and speed. The Taguchi method was used to optimize parameters and minimize experiments. Analysis found treated fiber composites had the lowest wear rate.
The document discusses WiMAX technology and its use of OFDM to address bottleneck problems in high-speed networks. It describes how WiMAX uses OFDM to efficiently allocate bandwidth and share it between connecting nodes based on their data request sizes, in order to prevent data loss. It also provides background on WiMAX architecture and management, and how it can support high-speed point-to-point and point-to-multipoint connections with optimized handover and security features.
1) The document proposes an improvement to the AODV routing protocol called AODV_V for wireless networks.
2) AODV_V uses a table-driven routing approach like DSDV for stationary nodes to quickly find routes, while using an on-demand approach like AODV for active nodes to reduce overhead.
3) Simulation results showed that AODV_V improved delivery rate and reduced delay and packet loss compared to the standard AODV protocol.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document analyzes the properties of superconducting striplines used to interconnect devices in Josephson logic and memory circuits. It investigates properties like inductance, capacitance, propagation delay, attenuation, and phase velocity as functions of temperature and frequency for both low-Tc and high-Tc superconducting materials. The analysis of stripline properties helps optimize switching speed and circuit dimensions before experimental fabrication. It finds that high-Tc superconducting striplines above liquid nitrogen temperature can improve semiconductor and superconducting circuit performance compared to conventional metals or low-Tc materials.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
O documento descreve os principais constituintes do sangue e suas funções. Descreve que o sangue transporta gases respiratórios, nutrientes, excretas e hormônios, além de defender o organismo. Detalha que os glóbulos vermelhos transportam oxigênio, os glóbulos brancos defendem o organismo, e as plaquetas promovem a coagulação do sangue. O plasma transporta nutrientes, substâncias tóxicas e constitui a parte líquida do sangue.
Quero a delicia de poder fotografar as coisasLucas Fonseca
O documento discute a apreciação das coisas simples da vida através da fotografia. Ele descreve como a fotografia pode capturar momentos simples usando os olhos, alma, coração e emoção. Também discute como focar nas coisas positivas ao invés de reclamar constantemente pode trazer mais felicidade.
The document discusses the benefits of meditation for reducing stress and anxiety. Regular meditation practice can calm the mind and body by lowering heart rate and blood pressure. Meditation may also have psychological benefits like improved focus, emotional regulation, and reduced reactivity to stress.
El documento celebra a Herbart por lograr algo que muchos solo sueñan, posiblemente una victoria deportiva o profesional, a pesar de la duda inicial o de ser superados en algún momento, logrando superarse a sí mismos.
El documento presenta la Agenda de Política Económica para el Buen Vivir 2011-2013 del Ecuador. La agenda describe los objetivos y políticas para los sectores financiero y externo con el fin de alcanzar el Buen Vivir. Se busca democratizar el acceso al sistema financiero, canalizar el ahorro interno hacia la inversión productiva, y mantener la viabilidad de la cuenta corriente externa mediante la sustitución de importaciones y el desarrollo endógeno. La agenda guía las políticas económicas del país para los pró
O documento apresenta os resultados operacionais e financeiros da TPI no 1T12. Destaca-se o crescimento de 21,7% na receita líquida e de 22,2% no EBITDA ajustado. O lucro líquido atingiu R$15 milhões e o lucro base de dividendos foi de R$31,9 milhões. Detalha também investimentos realizados e endividamento da companhia.
Sibahle Mdluli is a South African undergraduate student born on June 25, 1997 in Nelspruit, South Africa. She is currently completing her foundation year at Monash South Africa in Johannesburg to enter a Bachelor of Social Science degree program in 2017. Sibahle speaks Isizulu, Isiswati, English, and Afrikaans and has exceptional computer skills.
A empresa de tecnologia anunciou um novo smartphone com câmera aprimorada, tela maior e bateria de longa duração por um preço acessível. O dispositivo tem como objetivo atrair mais consumidores em mercados emergentes com suas especificações equilibradas e preço baixo. Analistas esperam que as melhorias e o preço baixo impulsionem as vendas do novo aparelho.
Semantic Annotation: The Mainstay of Semantic WebEditor IJCATR
Given that semantic Web realization is based on the critical mass of metadata accessibility and the representation of data with formal
knowledge, it needs to generate metadata that is specific, easy to understand and well-defined. However, semantic annotation of the
web documents is the successful way to make the Semantic Web vision a reality. This paper introduces the Semantic Web and its
vision (stack layers) with regard to some concept definitions that helps the understanding of semantic annotation. Additionally, this
paper introduces the semantic annotation categories, tools, domains and models
Topic Modeling : Clustering of Deep Webpagescsandit
The internet is comprised of massive amount of info
rmation in the form of zillions of web
pages.This information can be categorized into the
surface web and the deep web. The existing
search engines can effectively make use of surface
web information.But the deep web remains
unexploited yet. Machine learning techniques have b
een commonly employed to access deep
web content.
Under Machine Learning, topic models provide a simp
le way to analyze large volumes of
unlabeled text. A "topic" consists of a cluster of
words that frequently occur together. Using
contextual clues, topic models can connect words wi
th similar meanings and distinguish
between words with multiple meanings. Clustering is
one of the key solutions to organize the
deep web databases.In this paper, we cluster deep w
eb databases based on the relevance found
among deep web forms by employing a generative prob
abilistic model called Latent Dirichlet
Allocation(LDA) for modeling content representative
of deep web databases. This is
implemented after preprocessing the set of web page
s to extract page contents and form
contents.Further, we contrive the distribution of “
topics per document” and “words per topic”
using the technique of Gibbs sampling. Experimental
results show that the proposed method
clearly outperforms the existing clustering methods
.
This document summarizes a research paper that analyzes the movement of metallic particles in gas insulated busduct systems due to electric fields. It presents analytical, finite difference, and finite element methods to model the electric fields and calculate particle trajectories. Simulation results show that aluminum and copper particle movement increases with higher voltages. The maximum radial movement is 10.75mm at 220kV and 35.66mm at 600kV using different field calculation methods.
This document summarizes a research paper that proposes a new DSP controller-based switching configuration for a hybrid distributed energy system using a single input DC-DC buck-boost converter. The system allows two renewable energy sources, such as solar and wind, to supply load either separately or simultaneously based on availability. The DSP controller is used to extract maximum energy from the available source and improve system efficiency. The system can operate in three modes depending on source availability to maximize efficiency. Simulation results using MATLAB/Simulink are presented and analyzed.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document investigates the abrasive wear behavior of coir fiber reinforced epoxy composites using the Taguchi method. Composites were made with 10%, 20%, and 30% treated and untreated coir fiber. Abrasion tests were conducted at loads of 10-30N, speeds of 300-500rpm. Treated fiber composites showed better wear resistance than untreated. Wear rate decreased with increasing fiber content and increased with increasing load and speed. The Taguchi method was used to optimize parameters and minimize experiments. Analysis found treated fiber composites had the lowest wear rate.
The document discusses WiMAX technology and its use of OFDM to address bottleneck problems in high-speed networks. It describes how WiMAX uses OFDM to efficiently allocate bandwidth and share it between connecting nodes based on their data request sizes, in order to prevent data loss. It also provides background on WiMAX architecture and management, and how it can support high-speed point-to-point and point-to-multipoint connections with optimized handover and security features.
1) The document proposes an improvement to the AODV routing protocol called AODV_V for wireless networks.
2) AODV_V uses a table-driven routing approach like DSDV for stationary nodes to quickly find routes, while using an on-demand approach like AODV for active nodes to reduce overhead.
3) Simulation results showed that AODV_V improved delivery rate and reduced delay and packet loss compared to the standard AODV protocol.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document analyzes the properties of superconducting striplines used to interconnect devices in Josephson logic and memory circuits. It investigates properties like inductance, capacitance, propagation delay, attenuation, and phase velocity as functions of temperature and frequency for both low-Tc and high-Tc superconducting materials. The analysis of stripline properties helps optimize switching speed and circuit dimensions before experimental fabrication. It finds that high-Tc superconducting striplines above liquid nitrogen temperature can improve semiconductor and superconducting circuit performance compared to conventional metals or low-Tc materials.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
O documento descreve os principais constituintes do sangue e suas funções. Descreve que o sangue transporta gases respiratórios, nutrientes, excretas e hormônios, além de defender o organismo. Detalha que os glóbulos vermelhos transportam oxigênio, os glóbulos brancos defendem o organismo, e as plaquetas promovem a coagulação do sangue. O plasma transporta nutrientes, substâncias tóxicas e constitui a parte líquida do sangue.
Quero a delicia de poder fotografar as coisasLucas Fonseca
O documento discute a apreciação das coisas simples da vida através da fotografia. Ele descreve como a fotografia pode capturar momentos simples usando os olhos, alma, coração e emoção. Também discute como focar nas coisas positivas ao invés de reclamar constantemente pode trazer mais felicidade.
The document discusses the benefits of meditation for reducing stress and anxiety. Regular meditation practice can calm the mind and body by lowering heart rate and blood pressure. Meditation may also have psychological benefits like improved focus, emotional regulation, and reduced reactivity to stress.
El documento celebra a Herbart por lograr algo que muchos solo sueñan, posiblemente una victoria deportiva o profesional, a pesar de la duda inicial o de ser superados en algún momento, logrando superarse a sí mismos.
El documento presenta la Agenda de Política Económica para el Buen Vivir 2011-2013 del Ecuador. La agenda describe los objetivos y políticas para los sectores financiero y externo con el fin de alcanzar el Buen Vivir. Se busca democratizar el acceso al sistema financiero, canalizar el ahorro interno hacia la inversión productiva, y mantener la viabilidad de la cuenta corriente externa mediante la sustitución de importaciones y el desarrollo endógeno. La agenda guía las políticas económicas del país para los pró
O documento apresenta os resultados operacionais e financeiros da TPI no 1T12. Destaca-se o crescimento de 21,7% na receita líquida e de 22,2% no EBITDA ajustado. O lucro líquido atingiu R$15 milhões e o lucro base de dividendos foi de R$31,9 milhões. Detalha também investimentos realizados e endividamento da companhia.
Sibahle Mdluli is a South African undergraduate student born on June 25, 1997 in Nelspruit, South Africa. She is currently completing her foundation year at Monash South Africa in Johannesburg to enter a Bachelor of Social Science degree program in 2017. Sibahle speaks Isizulu, Isiswati, English, and Afrikaans and has exceptional computer skills.
A empresa de tecnologia anunciou um novo smartphone com câmera aprimorada, tela maior e bateria de longa duração por um preço acessível. O dispositivo tem como objetivo atrair mais consumidores em mercados emergentes com suas especificações equilibradas e preço baixo. Analistas esperam que as melhorias e o preço baixo impulsionem as vendas do novo aparelho.
Semantic Annotation: The Mainstay of Semantic WebEditor IJCATR
Given that semantic Web realization is based on the critical mass of metadata accessibility and the representation of data with formal
knowledge, it needs to generate metadata that is specific, easy to understand and well-defined. However, semantic annotation of the
web documents is the successful way to make the Semantic Web vision a reality. This paper introduces the Semantic Web and its
vision (stack layers) with regard to some concept definitions that helps the understanding of semantic annotation. Additionally, this
paper introduces the semantic annotation categories, tools, domains and models
Topic Modeling : Clustering of Deep Webpagescsandit
The internet is comprised of massive amount of info
rmation in the form of zillions of web
pages.This information can be categorized into the
surface web and the deep web. The existing
search engines can effectively make use of surface
web information.But the deep web remains
unexploited yet. Machine learning techniques have b
een commonly employed to access deep
web content.
Under Machine Learning, topic models provide a simp
le way to analyze large volumes of
unlabeled text. A "topic" consists of a cluster of
words that frequently occur together. Using
contextual clues, topic models can connect words wi
th similar meanings and distinguish
between words with multiple meanings. Clustering is
one of the key solutions to organize the
deep web databases.In this paper, we cluster deep w
eb databases based on the relevance found
among deep web forms by employing a generative prob
abilistic model called Latent Dirichlet
Allocation(LDA) for modeling content representative
of deep web databases. This is
implemented after preprocessing the set of web page
s to extract page contents and form
contents.Further, we contrive the distribution of “
topics per document” and “words per topic”
using the technique of Gibbs sampling. Experimental
results show that the proposed method
clearly outperforms the existing clustering methods
.
Topic Modeling : Clustering of Deep Webpagescsandit
The internet is comprised of massive amount of information in the form of zillions of web pages.This information can be categorized into the surface web and the deep web. The existing search engines can effectively make use of surface web information.But the deep web remains unexploited yet. Machine learning techniques have been commonly employed to access deep web content.
Under Machine Learning, topic models provide a simple way to analyze large volumes of unlabeled text. A "topic" consists of a cluster of words that frequently occur together. Using
contextual clues, topic models can connect words with similar meanings and distinguish between words with multiple meanings. Clustering is one of the key solutions to organize the deep web databases.In this paper, we cluster deep web databases based on the relevance found among deep web forms by employing a generative probabilistic model called Latent Dirichlet
Allocation(LDA) for modeling content representative of deep web databases. This is implemented after preprocessing the set of web pages to extract page contents and form
contents.Further, we contrive the distribution of “topics per document” and “words per topic”
using the technique of Gibbs sampling. Experimental results show that the proposed method clearly outperforms the existing clustering methods.
Semantic - Based Querying Using Ontology in Relational Database of Library Ma...dannyijwest
The traditional Web stores huge amount of data in the form of Relational Databases (RDB) as it is good at
storing objects and relationships between them. Relational Databases are dynamic in nature which allows
bringing tables together helping user to search for related material across multiple tables. RDB are
scalable to expand as the data grows. The RDB uses a Structured Query Language called SQL to access
the databases for several data retrieval purposes. As the world is moving today from the Syntactic form to
Semantic form and the Web is also taking its new form of Semantic Web. The Structured Query of the RDB
on web can be a Semantic Query on Semantic Web.
Metadata: Towards Machine-Enabled Intelligencedannyijwest
World Wide Web has revolutionized the means of data availability, but with its current structure model , it is becoming increasingly difficult to retrieve relevant information, with reasonable precision and recall, using the major search engines. However, with use of metadata, combined with the use of improved searching techniques, helps to enhance relevant information retrieval .The design of structured, descriptions of Web resources enables greater search precision and a more accurate relevance ranking of retrieved information .One such efforts towards standardization is , Dublin Core standard, which has been developed as Metadata Standard and also other standards which enhances retrieval of a wide range of information resources. This paper discuses the importance of metadata, various metadata schemas and elements, and the need of standardization of Metadata. This paper further discusses how the metadata can be generated using various tools which assist intelligent agents for efficient retrieval
Metadata: Towards Machine-Enabled Intelligence dannyijwest
World Wide Web has revolutionized the means of data availability, but with its current structure model , it
is becoming increasingly difficult to retrieve relevant information, with reasonable precision and recall,
using the major search engines. However, with use of metadata, combined with the use of improved
searching techniques, helps to enhance relevant information retrieval .The design of structured,
descriptions of Web resources enables greater search precision and a more accurate relevance ranking of
retrieved information .One such efforts towards standardization is , Dublin Core standard, which has been
developed as Metadata Standard and also other standards which enhances retrieval of a wide range of
information resources. This paper discuses the importance of metadata, various metadata schemas and
elements, and the need of standardization of Metadata. This paper further discusses how the metadata can
be generated using various tools which assist intelligent agents for efficient retrieval.
The World Wide Web is booming and radically vibrant due to the well established standards and widely accountable framework which guarantees the interoperability at various levels of the application and the society as a whole. So far, the web has been functioning at the random rate on the basis of the human intervention and some manual processing but the next generation web which the researchers called semantic web, edging for automatic processing and machine-level understanding. The well set notion, Semantic Web would be turn possible if only there exists the further levels of interoperability prevails among the applications and networks. In achieving this interoperability and greater functionality among the applications, the W3C standardization has already released the well defined standards such as RDF/RDF Schema and OWL. Using XML as a tool for semantic interoperability has not achieved anything effective and failed to bring the interconnection at the larger level. This leads to the further inclusion of inference layer at the top of the web architecture and its paves the way for proposing the common design for encoding the ontology representation languages in the data models such as RDF/RDFS. In this research article, we have given the clear implication of semantic web research roots and its ontological background process which may help to augment the sheer understanding of named entities in the web.
The document discusses using semantic technologies like XML, RDF, and OWL to represent data on the web in a structured format that is accessible to machines. It describes two main approaches for accessing semantic data on the deep web: ontology plug-in search and deep web service annotation. Both approaches require a semantic web crawler or bot to harvest concepts from deep web forms and iteratively link them to build enriched ontologies that define domain terms and relationships to provide machine-interpretable meaning.
Semantic Query Optimisation with Ontology Simulationdannyijwest
Semantic Web is, without a doubt, gaining momentum in both industry and academia. The word “Semantic” refers to “meaning” – a semantic web is a web of meaning. In this fast changing and result oriented practical world, gone are the days where an individual had to struggle for finding information on the Internet where knowledge management was the major issue. The semantic web has a vision of linking, integrating and analysing data from various data sources and forming a new information stream, hence a web of databases connected with each other and machines interacting with other machines to yield results which are user oriented and accurate. With the emergence of Semantic Web framework the naïve approach of searching information on the syntactic web is cliché. This paper proposes an optimised semantic searching of keywords exemplified by simulation an ontology of Indian universities with a proposed algorithm which ramifies the effective semantic retrieval of information which is easy to access and time saving.
The Semantic Web is a vision of information that is understandable by computers. Although there is great exploitable potential, we are still in "Generation Zero'' of the Semantic Web, since there are few real-world compelling applications. The heterogeneity, the volume of data and the lack of standards are problems that could be addressed through some nature inspired methods. The paper presents the most important aspects of the Semantic Web, as well as its biggest issues; it then describes some methods inspired from nature - genetic algorithms, artificial neural networks, swarm intelligence, and the way these techniques can be used to deal with Semantic Web problems.
Advance Frameworks for Hidden Web Retrieval Using Innovative Vision-Based Pag...IOSR Journals
The document proposes an innovative vision-based page segmentation (IVBPS) algorithm to improve hidden web content extraction. It aims to overcome limitations of existing approaches that rely heavily on HTML structure. IVBPS extracts blocks from the visual representation of a page and clusters them to segment the page semantically. It uses layout features like position and appearance to locate data regions and extract records. The algorithm analyzes the entire page structure rather than local regions, allowing it to retain content DOM tree methods may discard. This is expected to significantly improve hidden web extraction performance.
B-BabelNet: Business-Specific Lexical Database for Improving Semantic Analysi...TELKOMNIKA JOURNAL
B-BabelNet is a new lexical database designed to improve semantic analysis of business process models. It is constructed using the same method as BabelNet but focuses only on terms related to the business domain. B-BabelNet maps Wikipedia pages in business categories and terms from a business dictionary to WordNet synsets. It also constructs relations between synsets to enrich the database. The goal is to address limitations of existing lexical databases for business terms and increase accuracy of semantic analysis for business processes.
The document discusses the Semantic Web, which aims to develop the current web so that machines can understand the meaning of information and not just display it. It outlines some key technologies being used like XML, RDF, and ontologies to add structure and meaning to web content. This will allow software agents to perform more sophisticated tasks by processing structured, machine-readable information based on defined ontologies. The Semantic Web represents an evolution from today's web designed primarily for humans to one where machines can also comprehend and utilize web content.
A web application is one that is invoked due to a web browser over internet. In just a decade, the web has grown from a repository of static web pages to a powerful platform of choice for developing dynamic applications using large number of web technologies and languages. This survey covers five web technologies from different phases of internet development showing their features for web development.
This document discusses methods for measuring semantic similarity between words. It begins by discussing how traditional lexical similarity measurements do not consider semantics. It then discusses several existing approaches that measure semantic similarity using web search engines and text snippets. These approaches calculate word co-occurrence statistics from page counts and analyze lexical patterns extracted from snippets. Pattern clustering is used to group semantically similar patterns. The approaches are evaluated using datasets and metrics like precision and recall. Finally, the document proposes a new method that combines page count statistics, lexical pattern extraction and clustering, and support vector machines to measure semantic similarity.
This document discusses methods for measuring semantic similarity between words. It begins by discussing how traditional lexical similarity measurements do not consider semantics. It then discusses several existing approaches that measure semantic similarity using web search engines and text snippets. These approaches calculate word co-occurrence statistics from page counts and analyze lexical patterns extracted from snippets. Pattern clustering is used to group semantically similar patterns. The approaches are evaluated and combined using support vector machines. Finally, the document proposes a new method that applies lexical pattern extraction and clustering to page count data and text snippets, before using an SVM to measure semantic similarity.
Information on the web is tremendously increasing in
recent years with the faster rate. This massive or voluminous data
has driven intricate problems for information retrieval and
knowledge management. As the data resides in a web with several
forms, the Knowledge management in the web is a challenging
task. Here the novel 'Semantic Web' concept may be used for
understanding the web contents by the machine to offer
intelligent services in an efficient way with a meaningful
knowledge representation. The data retrieval in the traditional
web source is focused on 'page ranking' techniques, whereas in
the semantic web the data retrieval processes are based on the
‘concept based learning'. The proposed work is aimed at the
development of a new framework for automatic generation of
ontology and RDF to some real time Web data, extracted from
multiple repositories by tracing their URI’s and Text Documents.
Improved inverted indexing technique is applied for ontology
generation and turtle notation is used for RDF notation. A
program is written for validating the extracted data from
multiple repositories by removing unwanted data and considering
only the document section of the web page.
Amit P. Sheth, “Relationships at the Heart of Semantic Web: Modeling, Discovering, Validating and Exploiting Complex Semantic Relationships,” Keynote at the 29th Conference on Current Trends in Theory and Practice of Informatics (SOFSEM 2002), Milovy, Czech Republic, November 22–29, 2002.
Keynote: http://www.sofsem.cz/sofsem02/keynote.html
Related paper: http://knoesis.wright.edu/?q=node/2063
Relationships at the Heart of Semantic Web: Modeling, Discovering, Validating...
En24877880
1. Sukanta Sinha, Rana Dattagupta, Debajyoti Mukhopadhyay / International Journal of
Engineering Research and Applications (IJERA) ISSN: 2248-9622
www.ijera.com Vol. 2, Issue 4, July-August 2012, pp.877-880
Identify Web-page Content meaning using Knowledge based
System for Dual Meaning Words
Sukanta Sinha1, 4, Rana Dattagupta2, Debajyoti Mukhopadhyay3, 4
1
(TATA Consultancy Services, Victoria Park, Kolkata 700091, India)
2
(Computer Sc. Dept., Jadavpur University, Kolkata 700032, India)
3
(Dept. of Information Technology, Maharashtra Institute of Technology, Pune 411038, India)
4
(WIDiCoReL, Green Tower C- 9/1, Golf Green, Kolkata 700095, India)
Abstract
holds only dual meaning words in their Web-page
Meaning of Web-page content plays a big
content. To identify the meaning, we have created a
role while produced a search result from a search
knowledge based system by collecting various types
engine. Most of the cases Web-page meaning
of data patterns.
stored in title or meta-tag area but those
Our paper is not intended to provide a complete
meanings do not always match with Web-page
survey of techniques. According to our knowledge,
content. To overcome this situation we need to go
we have applied these techniques on few examples.
through the Web-page content to identify the
Now a day‟s research on search engine has been
Web-page meaning. In such cases, where Web-
carried out in universities and open laboratories,
page content holds dual meaning words that time
many dot-com companies. Unfortunately, many of
it is really difficult to identify the meaning of the
these techniques are used by dot-coms, and
Web-page. In this paper, we are introducing a
especially the resulting performance, are kept
new design and development mechanism of
private behind company walls, or are disclosed in
identifying the Web-page content meaning which
patents that can be comprehended and appreciate by
holds dual meaning words in their Web-page
the lawyers. Therefore, we believe that the overview
content.
of problems and techniques that we presented here
can be useful.
Keywords – Dual meaning word, Knowledge This paper discusses survey of the problem area in
based system, Search engine, Web-page content, section 2. Section 3 discusses about the XML
Web resources schema. Section 4 depicts the proposed approach.
Section 5 shows some experimental analyses.
1. Introduction Finally, section 6 concludes the paper.
Web search engine is a tool that produces
search results based on the user given query. World 2. The Problem Area
Wide Web (WWW) is a huge reservoir of Web- Web-page content meaning identification is
pages. Search engine crawler crawls down the Web- an essential part of a search engine to produce
pages from WWW and creates a database of Web relevant search result. Most of the cases we can get
resources for the search engine [1, 2]. the Web-page content meaning from title or meta-
In the present era of Internet, WWW is an tag area of that Web-page content but they do not
accumulated and interactive medium for accessing always match with the actual Web-page content. On
an enormous conglomeration of information [3]. the other hand, a few cases where Web-page content
The information in the Web-page content consists of holding dual meaning words are really difficult to
diverse data types such as structured data, semi identify the meaning of the Web-page content.
structured data and lack of structure of Web data, In general, our main goal is to identify the Web-
etc. [4]. Few cases we also found holds dual page content meaning which holds dual meaning
meaning words are exists in Web-page content. words in their Web-page content. The example
Meaning identification of those Web-page contents illustrates the difficulty to identify the meaning of a
which holds dual meaning words is a challenging Web-page content, which can be overcome by using
task. our proposed system.
The dual meaning word means a word which
contains two meanings like „bank‟ represents Example 1: John is looking for a bank to open a
„financial institute‟ as well as „river side‟. We need savings account on the other hand Alex is looking
to identify the meaning based on the full sentence. for a bank of the river for a get together. Here, both
In our approach, we have mainly focused on the bank represents different meaning, one for
identifying the Web-page content meaning, which financial institutes and other one for river side. If
both the sentence exists in different Web-page
877 | P a g e
2. Sukanta Sinha, Rana Dattagupta, Debajyoti Mukhopadhyay / International Journal of
Engineering Research and Applications (IJERA) ISSN: 2248-9622
www.ijera.com Vol. 2, Issue 4, July-August 2012, pp.877-880
content then the meaning of the Web-page content complex type element which holds similar types of
need to be retrieved based on their content. key elements with their meaning. „names‟ is a
complex type element which holds key element
Example 2: Peter found a bank which located on the names that represent same meaning. „name‟ and
bank of the river. This is a single sentence which „meaning‟ are simple type element holds key values
represents financial institutions as well as river side. and their meaning. Each XML holds a „dmw_id‟.
This time any one of the meanings is valid for the We have maintained dual meaning word with a
sentence. In our approach, we assumed that one corresponding „dmw_id‟. Key words are taken from
Web-page has only one meaning. Hence, for this dual meaning word holding sentence. For example
type of situation we will assign any one meaning “John is looking for a bank to open a savings
based on our programming logic. account” and “Alex is looking for a bank of the river
for a get together” holds „account‟, „river‟ key
3. XML Schema words. All the key word meaning is taken care while
design the XML. In Fig.2 we have shown a part of
An XML Schema describes the structure of an XML
an XML for „bank‟.
document [5, 6]. The XML Schema language refers
to an XML Schema Definition (XSD). The purpose
of an XML Schema is to define the legal building
blocks of an XML document. An XML Schema
defines elements, attributes that can appear in a
document [7, 8]. It also expressed data types, default
and fixed values for elements and attributes. One of
the greatest strengths of XML Schemas is the
support for data types and written in XML. XML
Schemas are extensible because they are written in
XML.
XML Schema holds simple and complex elements
[9, 10, 11]. A simple element is an XML element
that contains only text. It cannot contain any other
elements or attributes. A complex element is an
XML element that contains other elements and/or
attributes. There are four kinds of complex
elements; they are empty elements, elements that
contain only other elements, elements that contain
only text, elements that contain both other elements figure 1. A sample XSD
and text. The <schema> element is the root element
of every XML Schema. The <schema> element may
contain some attributes [12, 13, 14].
4. Proposed Approach
In our approach, we have proposed a mechanism
which identifies meaning of Web-page content for
those who holds dual meaning word in their Web-
page content. Section 4.1 explains an overview of
creating knowledge based system and section 4.2
depicts our algorithm.
4.1. Knowledge Based System Generation
To create a knowledge based system we have
collected dual meaning words from various sources
like internet, dictionary, etc. Now for each dual
figure 2. A part of an XML (for bank)
meaning word, we have created one XML which
link with Fig.1 given XSD. The considered XSD 4.2. Algorithm
holds both simple and complex type of elements.
To identify Web-page content meaning we are using
„dualMeaningWordName‟ attribute holds the dual
below given algorithm. This algorithm mainly
meaning word name. „keywords‟ is a complex
focused on identifying the Web-page content
element which holds various sets of keyword, which
meaning, which holds dual meaning words in their
classified based on their meaning. „keyword‟ also a
878 | P a g e
3. Sukanta Sinha, Rana Dattagupta, Debajyoti Mukhopadhyay / International Journal of
Engineering Research and Applications (IJERA) ISSN: 2248-9622
www.ijera.com Vol. 2, Issue 4, July-August 2012, pp.877-880
Web-page content. In our approach, we have used a proper meaning, XML schema, etc. Initially, we
knowledge based system for identifying the have created the knowledge based system with the
meaning of dual meaning words. The knowledge help of internet, dictionary. Then we have tuned the
based system stores the information in XML form. knowledge based system through our experiments.
Input : Web-page content In our experiment, we have taken a Web-page from
Output : Meaning of the Web-page our repository and pass it through our system and
content check the database for the meaning of that Web-
page. If the Web-page holds dual meaning words
1. Extract dual meaning words from the
Web-page content. then the meaning will identified otherwise update
2. get count of dual meaning words in 'isDualMeaningFlag' as false.
the Web-page content
3. if count = 0 then 5.2. Experimental Results
set isDualMeaningFlag:=False and It is very difficult to compare our system
exit
with any existing system. Anyhow we have
4. if count = 1 then
a) set isDualMeaningFlag:=True produced few data to measure our proposed system
b) Extract key words in the dual performance. As a part of experimental results, we
meaning word sentence have produced a statistic, which given in Table 1.
c) Based on the key word traverse
XML (knowledge based system) for Table1. Performance Report of Our System
dual meaning word
d) Retrieve the meaning of that key
No. of Web-page No. of Web-page Correct Meaning No. of Correct Meaning
and store it in a temporary
Taken / hold Dual Identified in 1st XML Identified after
table.
Repository Size Meaning Words Run Modified XML Modification
e) Go to step 6.
1000 30 22 6 28
5. if count > 1 then 2000 50 43 5 47
a) set isDualMeaningFlag:=True 3000 80 71 6 76
b) select the max occurred dual 4000 110 99 9 104
meaning word in the Web-page 5000 140 127 10 134
content
c) if there exists multiple dual
meaning word with same number of 6. Conclusion
occurrence then select dual Web-page content meaning identification is
meaning word which appeared
a very difficult job for any system. The human brain
first in the Web-page content
d) Extract key words in the dual can find it easily but need to go through each and
meaning word sentence every Web-page contents, which is really
e) Based on the key word traverse impossible. We found that approximate 30% - 40%
XML (knowledge based system) for
dual meaning word Web-pages are representing unique meaning; out of
f) Retrieve the meaning of that key those 30% - 40% approximate 8% - 10% Web-pages
and store it in a temporary are holding dual meaning words. Hence, we are
table.
g) Go to step 6. concentrating to create those 8% - 10% Web-page
6. Choose the meaning from temporary meaning XML. We found approximate 95%
table which count is maximized. successful cases achieved to identify Web-page
7. end
content meaning those held dual meaning words in
their Web-page content. Our approach is highly
5. Experimental Analysis scalable. Suppose, we encountered a new pattern
In this section, we have given some and want to support that pattern, then we just
experimental study as well as discussed how to set introduce the meaning XML and the system will
up our system. Section 5.1 explains our work. We have tested our system by taking a sub-set
experimental procedure, and section 5.2 shows the of Web-pages shown in experimental results section.
experimental results of our system. In this paper, we are mainly focused on our
approach, which will work for large volume of data.
5.1. Experimental Procedure
Performance of our system depends on
various parameters and those parameters need to be REFERENCES
set up before running our system. The considered [1] C. H. Yu, and S. J. Lin, Parallel Crawling and
parameters are Web-page repository, knowledge Capturing for On-Line Auction, Lecture Notes
based system, i.e., dual meaning word XML with
879 | P a g e
4. Sukanta Sinha, Rana Dattagupta, Debajyoti Mukhopadhyay / International Journal of
Engineering Research and Applications (IJERA) ISSN: 2248-9622
www.ijera.com Vol. 2, Issue 4, July-August 2012, pp.877-880
In Computer Science, Springer-Verlag, Berlin, Databases, WebDB 2004 Proceedings, Maison
Heidelberg, 5075, 2008, 455-466. de la Chimie, Paris, France, June 17-18, 2004,
79–84.
[2] D. Mukhopadhyay, A. Biswas, S. Sinha, A New
Approach to Design Domain Specific Ontology [13] B. Chidlovskii, Schema extraction from xml: A
Based Web Crawler, 10th International grammatical inference approach, In
Conference on Information Technology, ICIT Proceedings of the International Workshop on
2007 Proceedings, Rourkela, India, IEEE Knowledge Representation Meets Databases
Computer Society Press, California, USA, (KRDB), 2001.
December 17-20, 2007, 289-291.
[14] Y. Papakonstantinou and V. Vianu, DTD
[3] W. Willinger, R. Govindan, S. Jamin, V. Inference for Views of XML Data, In the
Paxson and S. Shenker, Scaling phenomena in Procedings of 19th ACM Symposium on
the Internet, In Proceedings of the National Principles of Database Systems (PODS),
Academy of Sciences, 1999, suppl. 1, 2573– Dallas, Texas, USA, 2000, 35-46.
2580.
[4] J. J. Rehmeyer, Mapping a medusa: The
Internet spreads its tentacles, Science News,
171, June 2007, 387-388.
[5] M. Murata, D. Lee, M. Mani and K.
Kawaguchi, Taxonomy of XML Schema
Languages using Formal Language Theory, In
ACM Trans. on Internet Technology (TOIT),
5(4), November 2005, 1-45.
[6] I. Stuart, XML Schema, a brief introduction
(Internet archived by WayBack Machine,
October 26, 2001).
[7] D. Lee and W. W. Chu, Comparative Analysis
of Six XML Schema Languages, In ACM
SIGMOD Record, 29(3), September 2000, 76-
87.
[8] C. Binstock, D. Peterson, M. Smith, M.
Wooding and C. Dix, The XML Schema
Complete Reference (Published by Addison-
Wesley, 2002).
[9] J. Hegewald, F. Naumann and M. Weis,
XStruct: efficient schema extraction from
multiple and large XML documents, The 22nd
International Conference on Data Engineering,
ICDE Workshops, IEEE Computer Society,
Atlanta, GA, April 3-8, 2006, 81-91.
[10] G. J. Bex, W. Martens, F. Neven and T.
Schwentick, Expressiveness of XSDs: from
practice to theory, there and back again, In
Proceedings of the 14th international World
Wide Web Conference, Chiba, Japan, 2005,
712–721.
[11] G. J. Bex, F. Neven, T. Schwentick and K.
Tuyls, Inference of concise DTDs from XML
data, Proceedings of the 32nd International
Conference on Very Large Data Bases (VLDB),
Seoul, Korea, September 12-15, 2006.
[12] G. J. Bex, F. Neven and J. V. Bussche, DTDs
versus XML Schema: a practical study, 7th
International Workshop on the Web and
880 | P a g e