There is an explosive growth of information in
the World Wide Web thus posing a challenge to Web users
to extract essential knowledge from the Web. Search
engines help us to narrow down the search in the form of
Search Engine Result Pages (SERP). Web Content Mining
is one of the techniques that help users to extract useful
information from these SERPs. In this paper, we propose
two similarity based mechanisms; WDES, to extract desired
SERPs and store them in the local depository for offline
browsing and WDICS, to integrate the requested contents
and enable the user to perform the intended analysis and
extract the desired information. Our experimental results
show that WDES and WDICS outperform DEPTA [1] in
terms of Precision and Recall.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document discusses improving web performance through prefetching frequently accessed pages. It begins by introducing the concept of prefetching web pages to reduce latency. Next, it reviews related work on predictive prefetching using techniques like Markov models and association rules to predict future page access. Finally, it proposes an approach to increase web performance by analyzing user access logs and website structure to predict pages for prefetching. The goal is to reduce latency and improve user experience by prefetching relevant pages in the background.
This document presents a heuristic approach for extracting web content using tag trees and heuristics. The approach first parses the web page into a tag tree using an HTML parser. Objects and separators are then extracted from the nested tag tree using heuristics like pattern repeating, standard deviation, and sibling tag heuristics. The main content is identified and implemented using the extracted separators. Experimental results showed the technique outperformed existing methods by extracting only the relevant content for the user's query.
The document discusses methods for improving web navigation efficiency through reconciling website structure based on user browsing patterns. It involves mining the website structure and user logs to determine browsing behaviors. Efficiency is calculated as the shortest path from the start page to the target page divided by the operating cost, defined as the number of pages visited. The approach was tested on a website and was able to reorganize the structure based on user navigation analysis from logs to increase browsing efficiency.
This document describes a prototype system that combines images located through textual metadata with a visual collaging technique to provide a leisurely browsing experience of a digital music collection. The system retrieves relevant images from the web based on metadata from MP3 files in the collection. These images are dynamically added and faded in a collage visualization. Users can click images to explore related web pages or play the associated songs. The system was implemented using the open source Greenstone digital library software.
IRJET - Re-Ranking of Google Search ResultsIRJET Journal
This document summarizes a research paper that proposes a hybrid personalized re-ranking approach to search results. It models a user's search interests using a conceptual user profile containing categories and concepts extracted from clicked results and a concept hierarchy. The user profile contains two types of documents - taxonomy documents representing general interests and viewed documents representing specific interests. A hybrid re-ranking process then semantically integrates the user's general and specific interests from their profile with search engine rankings to improve result relevance.
PTAR's library website provides several useful features and services to users. It has a simple and easy to navigate homepage. The website also offers bookmarks, maps and directions to help users virtually visit the library. Training guides and online publications are available to aid users in researching and accessing library resources. While some features work well, improvements could be made to the search functionality and providing clearer guidance on services.
Web personalization using clustering of web usage dataijfcstjournal
The exponential growth in the number and the complexity of information resources and services on the Web
has made log data an indispensable resource to characterize the users for Web-based environment. It
creates information of related web data in the form of hierarchy structure through approximation. This
hierarchy structure can be used as the input for a variety of data mining tasks such as clustering,
association rule mining, sequence mining etc.
In this paper, we present an approach for personalizing web user environment dynamically when he
interacting with web by clustering of web usage data using concept hierarchy. The system is inferred from
the web server’s access logs by means of data and web usage mining techniques to extract the information
about users. The extracted knowledge is used for the purpose of offering a personalized view of the
services to users.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document discusses improving web performance through prefetching frequently accessed pages. It begins by introducing the concept of prefetching web pages to reduce latency. Next, it reviews related work on predictive prefetching using techniques like Markov models and association rules to predict future page access. Finally, it proposes an approach to increase web performance by analyzing user access logs and website structure to predict pages for prefetching. The goal is to reduce latency and improve user experience by prefetching relevant pages in the background.
This document presents a heuristic approach for extracting web content using tag trees and heuristics. The approach first parses the web page into a tag tree using an HTML parser. Objects and separators are then extracted from the nested tag tree using heuristics like pattern repeating, standard deviation, and sibling tag heuristics. The main content is identified and implemented using the extracted separators. Experimental results showed the technique outperformed existing methods by extracting only the relevant content for the user's query.
The document discusses methods for improving web navigation efficiency through reconciling website structure based on user browsing patterns. It involves mining the website structure and user logs to determine browsing behaviors. Efficiency is calculated as the shortest path from the start page to the target page divided by the operating cost, defined as the number of pages visited. The approach was tested on a website and was able to reorganize the structure based on user navigation analysis from logs to increase browsing efficiency.
This document describes a prototype system that combines images located through textual metadata with a visual collaging technique to provide a leisurely browsing experience of a digital music collection. The system retrieves relevant images from the web based on metadata from MP3 files in the collection. These images are dynamically added and faded in a collage visualization. Users can click images to explore related web pages or play the associated songs. The system was implemented using the open source Greenstone digital library software.
IRJET - Re-Ranking of Google Search ResultsIRJET Journal
This document summarizes a research paper that proposes a hybrid personalized re-ranking approach to search results. It models a user's search interests using a conceptual user profile containing categories and concepts extracted from clicked results and a concept hierarchy. The user profile contains two types of documents - taxonomy documents representing general interests and viewed documents representing specific interests. A hybrid re-ranking process then semantically integrates the user's general and specific interests from their profile with search engine rankings to improve result relevance.
PTAR's library website provides several useful features and services to users. It has a simple and easy to navigate homepage. The website also offers bookmarks, maps and directions to help users virtually visit the library. Training guides and online publications are available to aid users in researching and accessing library resources. While some features work well, improvements could be made to the search functionality and providing clearer guidance on services.
Web personalization using clustering of web usage dataijfcstjournal
The exponential growth in the number and the complexity of information resources and services on the Web
has made log data an indispensable resource to characterize the users for Web-based environment. It
creates information of related web data in the form of hierarchy structure through approximation. This
hierarchy structure can be used as the input for a variety of data mining tasks such as clustering,
association rule mining, sequence mining etc.
In this paper, we present an approach for personalizing web user environment dynamically when he
interacting with web by clustering of web usage data using concept hierarchy. The system is inferred from
the web server’s access logs by means of data and web usage mining techniques to extract the information
about users. The extracted knowledge is used for the purpose of offering a personalized view of the
services to users.
DataCite and Campus Data Services
Paul Bracke, Associate Dean for Digital Programs and Information Services, Purdue University
Research libraries are increasingly interested in developing data services for their campuses. There are many perspectives, however, on how to develop services that are responsive to the many needs of scientists; sensitive to the concerns of scientists who are not always accustomed to sharing their data; and that are attractive to campus administrators. This presentation will discuss the development of campus-based data services programs, the centrality of data citation to these efforts, and the ways in which engagement with DataCite can enhance local programs.
This document provides an introduction to web structure mining and discusses two popular methods: HITS and PageRank. It begins with an overview of web mining categories including web content mining, web structure mining, and web usage mining. Web structure mining focuses on the hyperlink structure of the web and analyzes link relationships between pages. HITS and PageRank are two algorithms that have been proposed to handle potential correlations between linked pages and improve predictive accuracy.
The Perpustakaan Tun Abdul Razak (PTAR) library website provides access to many of the library's services and resources online. It contains features like bookmarks, practical information about the library's location, training guides, a search function, access to online publications and databases, an online catalogue, links to library friends and other libraries, an "Ask a librarian" function, subject folders, and information about the library's own publications. The website aims to be the main gateway for users to access the library's resources and services virtually. It provides concise and easy navigation to key information, resources, and assistance functions to support users' needs.
Kunal Punera is seeking a full-time position in research labs working on web/data mining, information retrieval, and machine learning. He has a Ph.D. in computer engineering from UT Austin with a focus on these areas. His research interests include web data analysis, data mining, machine learning, and information retrieval. He has published numerous papers in top conferences and journals and has worked with Yahoo! and IBM on related research projects.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
call for paper 2012, hard copy of journal, research paper publishing, where to publish research paper,
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Scientific discovery and innovation in an era of data-intensive science
William (Bill) Michener, Professor and Director of e-Science Initiatives for University Libraries, University of New Mexico; DataONE Principal Investigator
The scope and nature of biological, environmental and earth sciences research are evolving rapidly in response to environmental challenges such as global climate change, invasive species and emergent diseases. Scientific studies are increasingly focusing on long-term, broad-scale, and complex questions that require massive amounts of diverse data collected by remote sensing platforms and embedded environmental sensor networks; collaborative, interdisciplinary science teams; and new tools that promote scientific data preservation, discovery, and innovation. This talk describes the challenges facing scientists as they transition into this new era of data intensive science, presents current solutions, and lays out a roadmap to the future where new information technologies significantly increase the pace of scientific discovery and innovation.
IRJET-A Survey on Web Personalization of Web Usage MiningIRJET Journal
S.Jagan, Dr.S.P.Rajagopalan "A Survey on Web Personalization of Web Usage Mining", International Research Journal of Engineering and Technology (IRJET),Volume 2,issue-01 Mar-2015. e-ISSN:2395-0056, p-ISSN:2395-0072. www.irjet.net , published by Fast Track Publications
Abstract
Now a day, World Wide Web (www) is a rich and most powerful source of information. Day by day it is becoming more complex and expanding in size to get maximum information details online. However, it is becoming more complex and critical task to retrieve exact information expected by its users. To deal with this problem one more powerful concept is personalization which is becoming more powerful now days. Personalization is a subclass of information filtering system that seek to predict the 'ratings' or 'preferences' that a user would give to an items, they had not yet considered, using a model built from the characteristics of an item (content-based approaches or collaborative filtering approaches). Web mining is an emerging field of data mining used to provide personalization on the web. It consist three major categories i.e. Web Content Mining, Web Usage Mining, and Web Structure Mining. This paper focuses on web usage mining and algorithms used for providing personalization on the web.
WEB MINING: PATTERN DISCOVERY ON THE WORLD WIDE WEB - 2011Mustafa TURAN
Web Mining: Pattern Discovery on the World Wide Web
The study covers auto Turkish text content gathering from most popular social media platforms (twitter, facebook, blogs, news, etc...) and auto classifying sentimentally those contents.
Data Equivalence
Mark Parsons, Lead Project Manager, Senior Associate Scientist, National Snow and Ice Data Center
Data citation, especially using persistent identifiers like Digital Object Identifiers (DOIs), is an increasingly accepted scientific practice. Recently, several, respected organizations have developed guidelines for data citation. The different guidelines are largely congruent in that they agree on the basic practice and elements of data citation, especially for relatively static, whole data collections. There is less agreement on the more subtle nuances of data citation that are sometimes necessary to ensure precise reference and scientific reproducibility--the core purpose of data citation. We need to be sure that if you follow a data reference you get to the precise data that were used or at least their scientific equivalent. Identifiers such as DOIs are necessary but not sufficient for the precise, detailed, references necessary. This talk discusses issues around data set versioning, micro-citation, and scientific equivalence. I propose some interim solutions and suggest research strategies for the future.
Mobile Visual Search (MVS) allows users to search for visual data like images and videos using mobile devices. It faces challenges of low latency, limited computational power and bandwidth. The basic MVS pipeline extracts features from a query image, finds similar images in a database, and verifies matches. Research opportunities exist to improve interactive search under mobile constraints and enable robust recognition with varying image conditions. Datasets are also needed to advance MVS research.
Web Mining Research Issues and Future Directions – A SurveyIOSR Journals
This document summarizes research on web mining techniques. It begins with an abstract describing how web mining aims to extract useful information from vast amounts of unstructured web data. It then reviews various web mining techniques including web content mining, web structure mining, and web usage mining. The document surveys literature on pattern extraction techniques such as association rule mining, clustering, classification, and sequential pattern mining. It also discusses challenges in pre-processing web data and issues related to scaling up data mining algorithms for large web datasets. In closing, the document outlines future research directions in web mining including dealing with unstructured data and multimedia content.
This document discusses fuzzy clustering techniques for web mining. It proposes a fuzzy hierarchical clustering method to create clusters of web documents using fuzzy equivalence relations. The method aims to improve information retrieval by grouping similar documents into clusters. It describes how fuzzy clustering is suitable for web mining given the fuzzy nature of the web. It also provides background on related topics like web mining taxonomy, document clustering algorithms, and challenges of information retrieval on the web.
Web mining involves applying data mining techniques to discover useful information from web data. There are three types of web mining: web content mining analyzes data within web pages, web structure mining examines the hyperlink structure between pages, and web usage mining involves analyzing server logs to discover patterns in user behavior and interactions with websites. Web mining has applications in website design, web traffic analysis, e-commerce personalization, and security/crime investigation.
This document provides an annotated bibliography on user-friendly database interface design. It summarizes 5 sources that discuss various aspects of interface organization and design such as using intuitive layouts, clear navigation, and consistency. One source describes an idealized "perfect" database interface. Another discusses 7 principles for usable web design like having an intuitive structure. A third addresses database usability issues identified through research. The document provides high-level summaries of the sources and their relevance to interface design.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This document provides a literature survey and comparison of different techniques for web mining, including web structure mining, web usage mining, and web content mining. It summarizes various page ranking algorithms and models like PageRank, Weighted PageRank, HITS, General Utility Mining, and Topological Frequency Utility Mining. The document compares these algorithms and models based on the type of web mining activity, whether they consider website topology, their processing approach, and limitations. It aims to help compare techniques for analyzing the structure, usage, and content of websites.
Advance Frameworks for Hidden Web Retrieval Using Innovative Vision-Based Pag...IOSR Journals
The document proposes an innovative vision-based page segmentation (IVBPS) algorithm to improve hidden web content extraction. It aims to overcome limitations of existing approaches that rely heavily on HTML structure. IVBPS extracts blocks from the visual representation of a page and clusters them to segment the page semantically. It uses layout features like position and appearance to locate data regions and extract records. The algorithm analyzes the entire page structure rather than local regions, allowing it to retain content DOM tree methods may discard. This is expected to significantly improve hidden web extraction performance.
This document discusses extracting main content from deep web pages that contain multiple data regions. It proposes a hybrid approach with two steps: 1) Using visual features to identify the different data regions in the DOM tree. 2) Independently mining positive data records and data items from each data region using vision-based page segmentation. Related work on single-region deep web page extraction is also reviewed. The technique aims to automatically extract information from complex pages containing multiple, independent data listings.
This document proposes a rule-based system for classifying web pages. It discusses challenges with classifying unstructured web content and outlines various machine learning approaches used for web page classification, including association rule mining, naive Bayes, support vector machines, logistic regression, and decision trees. The proposed system uses a rule-based classifier to assign HTML documents to predefined categories based on checking for rule occurrences in the document content. The system is designed to improve web page classification by enabling online classification of pages.
This document discusses web page classification using a rule-based system. It begins by introducing the problem of classifying the large amount of unstructured information on the web. It then discusses various approaches to web page classification, including text content-based categorization and link and content analysis. The document focuses on using a rule-based classifier to assign HTML documents to predefined categories by checking for occurrences of system rules in the HTML content. Finally, it discusses common machine learning algorithms that have been used for web page classification, such as association rule mining, naive Bayes, support vector machines, logistic regression, and decision trees. The goal of the proposed system is to enhance other web page classification systems by enabling online classification of web pages.
DataCite and Campus Data Services
Paul Bracke, Associate Dean for Digital Programs and Information Services, Purdue University
Research libraries are increasingly interested in developing data services for their campuses. There are many perspectives, however, on how to develop services that are responsive to the many needs of scientists; sensitive to the concerns of scientists who are not always accustomed to sharing their data; and that are attractive to campus administrators. This presentation will discuss the development of campus-based data services programs, the centrality of data citation to these efforts, and the ways in which engagement with DataCite can enhance local programs.
This document provides an introduction to web structure mining and discusses two popular methods: HITS and PageRank. It begins with an overview of web mining categories including web content mining, web structure mining, and web usage mining. Web structure mining focuses on the hyperlink structure of the web and analyzes link relationships between pages. HITS and PageRank are two algorithms that have been proposed to handle potential correlations between linked pages and improve predictive accuracy.
The Perpustakaan Tun Abdul Razak (PTAR) library website provides access to many of the library's services and resources online. It contains features like bookmarks, practical information about the library's location, training guides, a search function, access to online publications and databases, an online catalogue, links to library friends and other libraries, an "Ask a librarian" function, subject folders, and information about the library's own publications. The website aims to be the main gateway for users to access the library's resources and services virtually. It provides concise and easy navigation to key information, resources, and assistance functions to support users' needs.
Kunal Punera is seeking a full-time position in research labs working on web/data mining, information retrieval, and machine learning. He has a Ph.D. in computer engineering from UT Austin with a focus on these areas. His research interests include web data analysis, data mining, machine learning, and information retrieval. He has published numerous papers in top conferences and journals and has worked with Yahoo! and IBM on related research projects.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
call for paper 2012, hard copy of journal, research paper publishing, where to publish research paper,
journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJERD, journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, publishing of research paper, reserach and review articles, IJERD Journal, How to publish your research paper, publish research paper, open access engineering journal, Engineering journal, Mathemetics journal, Physics journal, Chemistry journal, Computer Engineering, Computer Science journal, how to submit your paper, peer reviw journal, indexed journal, reserach and review articles, engineering journal, www.ijerd.com, research journals,
yahoo journals, bing journals, International Journal of Engineering Research and Development, google journals, hard copy of journal
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Scientific discovery and innovation in an era of data-intensive science
William (Bill) Michener, Professor and Director of e-Science Initiatives for University Libraries, University of New Mexico; DataONE Principal Investigator
The scope and nature of biological, environmental and earth sciences research are evolving rapidly in response to environmental challenges such as global climate change, invasive species and emergent diseases. Scientific studies are increasingly focusing on long-term, broad-scale, and complex questions that require massive amounts of diverse data collected by remote sensing platforms and embedded environmental sensor networks; collaborative, interdisciplinary science teams; and new tools that promote scientific data preservation, discovery, and innovation. This talk describes the challenges facing scientists as they transition into this new era of data intensive science, presents current solutions, and lays out a roadmap to the future where new information technologies significantly increase the pace of scientific discovery and innovation.
IRJET-A Survey on Web Personalization of Web Usage MiningIRJET Journal
S.Jagan, Dr.S.P.Rajagopalan "A Survey on Web Personalization of Web Usage Mining", International Research Journal of Engineering and Technology (IRJET),Volume 2,issue-01 Mar-2015. e-ISSN:2395-0056, p-ISSN:2395-0072. www.irjet.net , published by Fast Track Publications
Abstract
Now a day, World Wide Web (www) is a rich and most powerful source of information. Day by day it is becoming more complex and expanding in size to get maximum information details online. However, it is becoming more complex and critical task to retrieve exact information expected by its users. To deal with this problem one more powerful concept is personalization which is becoming more powerful now days. Personalization is a subclass of information filtering system that seek to predict the 'ratings' or 'preferences' that a user would give to an items, they had not yet considered, using a model built from the characteristics of an item (content-based approaches or collaborative filtering approaches). Web mining is an emerging field of data mining used to provide personalization on the web. It consist three major categories i.e. Web Content Mining, Web Usage Mining, and Web Structure Mining. This paper focuses on web usage mining and algorithms used for providing personalization on the web.
WEB MINING: PATTERN DISCOVERY ON THE WORLD WIDE WEB - 2011Mustafa TURAN
Web Mining: Pattern Discovery on the World Wide Web
The study covers auto Turkish text content gathering from most popular social media platforms (twitter, facebook, blogs, news, etc...) and auto classifying sentimentally those contents.
Data Equivalence
Mark Parsons, Lead Project Manager, Senior Associate Scientist, National Snow and Ice Data Center
Data citation, especially using persistent identifiers like Digital Object Identifiers (DOIs), is an increasingly accepted scientific practice. Recently, several, respected organizations have developed guidelines for data citation. The different guidelines are largely congruent in that they agree on the basic practice and elements of data citation, especially for relatively static, whole data collections. There is less agreement on the more subtle nuances of data citation that are sometimes necessary to ensure precise reference and scientific reproducibility--the core purpose of data citation. We need to be sure that if you follow a data reference you get to the precise data that were used or at least their scientific equivalent. Identifiers such as DOIs are necessary but not sufficient for the precise, detailed, references necessary. This talk discusses issues around data set versioning, micro-citation, and scientific equivalence. I propose some interim solutions and suggest research strategies for the future.
Mobile Visual Search (MVS) allows users to search for visual data like images and videos using mobile devices. It faces challenges of low latency, limited computational power and bandwidth. The basic MVS pipeline extracts features from a query image, finds similar images in a database, and verifies matches. Research opportunities exist to improve interactive search under mobile constraints and enable robust recognition with varying image conditions. Datasets are also needed to advance MVS research.
Web Mining Research Issues and Future Directions – A SurveyIOSR Journals
This document summarizes research on web mining techniques. It begins with an abstract describing how web mining aims to extract useful information from vast amounts of unstructured web data. It then reviews various web mining techniques including web content mining, web structure mining, and web usage mining. The document surveys literature on pattern extraction techniques such as association rule mining, clustering, classification, and sequential pattern mining. It also discusses challenges in pre-processing web data and issues related to scaling up data mining algorithms for large web datasets. In closing, the document outlines future research directions in web mining including dealing with unstructured data and multimedia content.
This document discusses fuzzy clustering techniques for web mining. It proposes a fuzzy hierarchical clustering method to create clusters of web documents using fuzzy equivalence relations. The method aims to improve information retrieval by grouping similar documents into clusters. It describes how fuzzy clustering is suitable for web mining given the fuzzy nature of the web. It also provides background on related topics like web mining taxonomy, document clustering algorithms, and challenges of information retrieval on the web.
Web mining involves applying data mining techniques to discover useful information from web data. There are three types of web mining: web content mining analyzes data within web pages, web structure mining examines the hyperlink structure between pages, and web usage mining involves analyzing server logs to discover patterns in user behavior and interactions with websites. Web mining has applications in website design, web traffic analysis, e-commerce personalization, and security/crime investigation.
This document provides an annotated bibliography on user-friendly database interface design. It summarizes 5 sources that discuss various aspects of interface organization and design such as using intuitive layouts, clear navigation, and consistency. One source describes an idealized "perfect" database interface. Another discusses 7 principles for usable web design like having an intuitive structure. A third addresses database usability issues identified through research. The document provides high-level summaries of the sources and their relevance to interface design.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This document provides a literature survey and comparison of different techniques for web mining, including web structure mining, web usage mining, and web content mining. It summarizes various page ranking algorithms and models like PageRank, Weighted PageRank, HITS, General Utility Mining, and Topological Frequency Utility Mining. The document compares these algorithms and models based on the type of web mining activity, whether they consider website topology, their processing approach, and limitations. It aims to help compare techniques for analyzing the structure, usage, and content of websites.
Advance Frameworks for Hidden Web Retrieval Using Innovative Vision-Based Pag...IOSR Journals
The document proposes an innovative vision-based page segmentation (IVBPS) algorithm to improve hidden web content extraction. It aims to overcome limitations of existing approaches that rely heavily on HTML structure. IVBPS extracts blocks from the visual representation of a page and clusters them to segment the page semantically. It uses layout features like position and appearance to locate data regions and extract records. The algorithm analyzes the entire page structure rather than local regions, allowing it to retain content DOM tree methods may discard. This is expected to significantly improve hidden web extraction performance.
This document discusses extracting main content from deep web pages that contain multiple data regions. It proposes a hybrid approach with two steps: 1) Using visual features to identify the different data regions in the DOM tree. 2) Independently mining positive data records and data items from each data region using vision-based page segmentation. Related work on single-region deep web page extraction is also reviewed. The technique aims to automatically extract information from complex pages containing multiple, independent data listings.
This document proposes a rule-based system for classifying web pages. It discusses challenges with classifying unstructured web content and outlines various machine learning approaches used for web page classification, including association rule mining, naive Bayes, support vector machines, logistic regression, and decision trees. The proposed system uses a rule-based classifier to assign HTML documents to predefined categories based on checking for rule occurrences in the document content. The system is designed to improve web page classification by enabling online classification of pages.
This document discusses web page classification using a rule-based system. It begins by introducing the problem of classifying the large amount of unstructured information on the web. It then discusses various approaches to web page classification, including text content-based categorization and link and content analysis. The document focuses on using a rule-based classifier to assign HTML documents to predefined categories by checking for occurrences of system rules in the HTML content. Finally, it discusses common machine learning algorithms that have been used for web page classification, such as association rule mining, naive Bayes, support vector machines, logistic regression, and decision trees. The goal of the proposed system is to enhance other web page classification systems by enabling online classification of web pages.
This document proposes a new technique to enhance the learning capabilities and reduce the computation intensity of a competitive learning multi-layered neural network using the K-means clustering algorithm. The proposed model uses a multi-layered network architecture with backpropagation learning to analyze web log data. Data preprocessing steps like cleaning, user identification, and transaction identification are applied to prepare the enterprise proxy log data for analysis. The proposed framework aims to discover useful patterns from web log data through a combination of K-means clustering and a feedforward neural network.
An Improved Annotation Based Summary Generation For Unstructured DataMelinda Watson
This document discusses annotation-based summarization of unstructured data. It begins with an introduction to annotation and information retrieval. Current annotation processes cannot maintain modifications due to frequent document updates. The document then reviews literature on automatic text classification, applying annotations to linked open data sets, and using domain ontologies for automatic document annotation. Keywords, sentences and contexts are extracted from documents for annotation. Different annotation models are discussed. The goal is to develop an improved annotation approach for summarizing unstructured data that can handle frequent document changes.
Web Content Mining Based on Dom Intersection and Visual Features Conceptijceronline
Structured Data extraction from deep Web pages is a challenging task due to the underlying complex structures of such pages. Also website developer generally follows different web page design technique. Data extraction from webpage is highly useful to build our own database from number applications. A large number of techniques have been proposed to address this problem, but all of them have inherent limitations because they present different limitations and constraints for extracting data from such webpages. This paper presents two different approaches to get structured data extraction. The first approach is non-generic solution which is based on template detection using intersection of Document Object Model Tree of various webpages from the same website. This approach is giving better result in terms of efficiency and accurately locating the main data at the particular webpage. The second approach is based on partial tree alignment mechanism based on using important visual features such as length, size, and position of web table available on the webpages. This approach is a generic solution as it does not depend on one particular website and its webpage template. It is perfectly locating the multiple data regions, data records and data items within a given web page. We have compared our work's result with existing mechanism and found our result much better for number webpage
Data Processing in Web Mining Structure by Hyperlinks and Pagerankijtsrd
Creating a quick and effective page ranking system for web crawling and retrieval is still a difficult problem. We suggest constructing a set of PageRank vectors biased using a collection of representative subjects in order to better capture the idea of relevance with regard to a certainty of topic in order to produce more accurate for search results. The outcome of the experiment demonstrates that the suggested algorithm improves the degree of relevance compared to the original one and reduces the topic sensitive PageRanks query time efforts. This paper offers an overview of Web mining as well as a review of its various categories. Next, we concentrate on one of these subcategories Web structure mining. In this area, we describe link mining and examine PageRank, two well liked techniques used in web structure mining. Ku Nalesh | Ghanshyam Sahu | Lalit Kumar P Bhaiya "Data Processing in Web Mining Structure by Hyperlinks and Pagerank" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-7 | Issue-6 , December 2023, URL: https://www.ijtsrd.com/papers/ijtsrd60083.pdf Paper Url: https://www.ijtsrd.com/computer-science/data-miining/60083/data-processing-in-web-mining-structure-by-hyperlinks-and-pagerank/ku-nalesh
This document discusses parsing HTML documents to extract data from websites. It proposes an automated system to parse HTML pages from the SEC website and extract specific data fields, like company financial information, to insert into databases of financial companies. The system will use Java parser libraries to identify patterns in SEC forms, including data in plain text and tables. It analyzes sample SEC forms to understand the structure and focus on extracting data from table sections.
Nowadays, the explosive growth of the World Wide Web generates tremendous amount of web data and consequently web data mining has become an important technique for discovering useful information and knowledge. Web mining is a vivid research area closely related to Information Extraction IE . Automatic content extraction from web pages is a challenging yet significant problem in the fields of information retrieval and data mining. Web Content mining refers to the discovery of useful information from web content such as text, images videos etc. Web content extraction is the process of organizing data instances into groups whose members are similar in some way. Content Extraction helps the user to easily select the topic of interest. Web Content Ming technology is useful in management information system. Web content mining extracts or mines useful information or knowledge from web page contents. This paper aims to study on web content extraction techniques. Aye Pwint Phyu | Khaing Khaing Wai "Study on Web Content Extraction Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd27931.pdfPaper URL: https://www.ijtsrd.com/computer-science/data-miining/27931/study-on-web-content-extraction-techniques/aye-pwint-phyu
This document summarizes research on improving web performance through integrated web prefetching and caching. It discusses how web prefetching can reduce latency by predicting and fetching web pages before they are requested. An integrated architecture reserves cache space for prefetched pages. By analyzing web logs to build prediction models of frequent paths, it aims to improve performance over caching alone. The tradeoff between reduced latency and potential increased network load is analyzed. Several previous works studying prefetching and caching algorithms individually are reviewed. The goal is a seamless prefetching system that works with existing caching systems.
This document summarizes research on improving web performance through integrated web prefetching and caching. It discusses how web prefetching can reduce latency by predicting and fetching web pages before they are requested. An integrated architecture reserves cache space for prefetched pages. By analyzing web logs to build prediction models of frequent paths, it aims to improve performance over caching alone. The tradeoff between reduced latency and potential increased network load is analyzed. Several previous works studying prefetching and caching algorithms individually are reviewed. The goal is a seamless prefetching system that works with existing caching systems.
Web Page Recommendation Using Web MiningIJERA Editor
On World Wide Web various kind of content are generated in huge amount, so to give relevant result to user web recommendation become important part of web application. On web different kind of web recommendation are made available to user every day that includes Image, Video, Audio, query suggestion and web page. In this paper we are aiming at providing framework for web page recommendation. 1) First we describe the basics of web mining, types of web mining. 2) Details of each web mining technique.3)We propose the architecture for the personalized web page recommendation.
AN EXTENSIVE LITERATURE SURVEY ON COMPREHENSIVE RESEARCH ACTIVITIES OF WEB US...James Heller
This document summarizes an extensive literature review on web usage mining. It begins by defining web mining and its three types: web content mining, web structure mining, and web usage mining. It then describes the main stages of web usage mining in detail: pre-processing, storage models, pattern discovery, optimization techniques, and pattern analysis. Finally, it discusses the characteristics of web server log data, including that it is unstructured, heterogeneous, distributed, contains different data types and dynamic content, and is voluminous, non-scalable, and incremental in nature.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
The document discusses web mining techniques for web personalization. It defines web mining as extracting useful information from web data, including web usage mining, web content mining, and web structure mining. Web usage mining involves data gathering, preparation, pattern discovery, analysis, visualization and application. Web content mining extracts information from web document contents. The document then discusses how these web mining techniques can be applied to web personalization by learning about user interactions and interests to customize web page content and presentations.
Optimization of Search Results with Duplicate Page Elimination using Usage DataIDES Editor
The performance and scalability of search engines
are greatly affected by the presence of enormous amount of
duplicate data on the World Wide Web. The flooded search
results containing a large number of identical or near identical
web pages affect the search efficiency and seek time of the users
to find the desired information within the search results. When
navigating through the results, the only information left behind
by the users is the trace through the pages they accessed. This
data is recorded in the query log files and usually referred to
as Web Usage Data. In this paper, a novel technique for
optimizing search efficiency by removing duplicate data from
search results is being proposed, which utilizes the usage data
stored in the query logs. The duplicate data detection is
performed by the proposed Duplicate Data Detection (D3)
algorithm, which works offline on the basis of favored user
queries found by pre-mining the logs with query clustering.
The proposed result optimization technique is supposed to
enhance the search engine efficiency and effectiveness to a large
scale.
COST-SENSITIVE TOPICAL DATA ACQUISITION FROM THE WEBIJDKP
The cost of acquiring training data instances for induction of data mining models is one of the main concerns in real-world problems. The web is a comprehensive source for many types of data which can be used for data mining tasks. But the distributed and dynamic nature of web dictates the use of solutions which can handle these characteristics. In this paper, we introduce an automatic method for topical data acquisition from the web. We propose a new type of topical crawlers that use a hybrid link context extraction method for topical crawling to acquire on-topic web pages with minimum bandwidth usage and with the lowest cost. The new link context extraction method which is called Block Text Window (BTW), combines a text window method with a block-based method and overcomes challenges of each of these methods using the advantages of the other one. Experimental results show the predominance of BTW in comparison with state of the art automatic topical web data acquisition methods based on standard metrics.
IRJET-Multi -Stage Smart Deep Web Crawling Systems: A ReviewIRJET Journal
This document summarizes research on multi-stage smart deep web crawling systems. It discusses challenges in efficiently locating deep web interfaces due to their large numbers and dynamic nature. It proposes a three-stage crawling framework to address these challenges. The first stage performs site-based searching to prioritize relevant sites. The second stage explores sites to efficiently search for forms. An adaptive learning algorithm selects features and constructs link rankers to prioritize relevant links for fast searching. Evaluation on real web data showed the framework achieves substantially higher harvest rates than existing approaches.
The document summarizes techniques for web mining, which involves mining web content, structure, and usage data. Web content mining extracts useful information from web page content and structures. Web structure mining analyzes the hyperlink structure between pages to determine important pages and group similar pages. Web usage mining analyzes server logs to discover general access patterns and customize websites for individual users based on their behavior. Text mining extends traditional data mining to unstructured text data through features like word occurrences and relationships.
Similar to Similarity based Dynamic Web Data Extraction and Integration System from Search Engine Result Pages for Web Content Mining (20)
Power System State Estimation - A ReviewIDES Editor
This document provides a review of power system state estimation techniques. It discusses both static and dynamic state estimation algorithms. For static state estimation, it covers weighted least squares, decoupled, and robust estimation methods. Weighted least squares is commonly used but can have numerical instability issues. Decoupled state estimation approximates the gain matrix for faster computation. Robust estimation uses M-estimators and other techniques to handle outliers and bad data. Dynamic state estimation applies Kalman filtering, leapfrog algorithms, and other methods to continuously monitor system states over time.
Artificial Intelligence Technique based Reactive Power Planning Incorporating...IDES Editor
This document summarizes a research paper that proposes using artificial intelligence techniques and FACTS controllers for reactive power planning in real-time power transmission systems. The paper formulates the reactive power planning problem and incorporates flexible AC transmission system (FACTS) devices like static VAR compensators (SVC), thyristor controlled series capacitors (TCSC), and unified power flow controllers (UPFC). Evolutionary algorithms like evolutionary programming (EP) and differential evolution (DE) are applied to find the optimal locations and settings of the FACTS controllers to minimize losses and costs. Simulation results on IEEE 30-bus and 72-bus Indian test systems show that UPFC performs best in reducing losses compared to SVC and TCSC.
Design and Performance Analysis of Genetic based PID-PSS with SVC in a Multi-...IDES Editor
Damping of power system oscillations with the help
of proposed optimal Proportional Integral Derivative Power
System Stabilizer (PID-PSS) and Static Var Compensator
(SVC)-based controllers are thoroughly investigated in this
paper. This study presents robust tuning of PID-PSS and
SVC-based controllers using Genetic Algorithms (GA) in
multi machine power systems by considering detailed model
of the generators (model 1.1). The effectiveness of FACTSbased
controllers in general and SVC-based controller in
particular depends upon their proper location. Modal
controllability and observability are used to locate SVC–based
controller. The performance of the proposed controllers is
compared with conventional lead-lag power system stabilizer
(CPSS) and demonstrated on 10 machines, 39 bus New England
test system. Simulation studies show that the proposed genetic
based PID-PSS with SVC based controller provides better
performance.
Optimal Placement of DG for Loss Reduction and Voltage Sag Mitigation in Radi...IDES Editor
This paper presents the need to operate the power
system economically and with optimum levels of voltages has
further led to an increase in interest in Distributed
Generation. In order to reduce the power losses and to improve
the voltage in the distribution system, distributed generators
(DGs) are connected to load bus. To reduce the total power
losses in the system, the most important process is to identify
the proper location for fixing and sizing of DGs. It presents a
new methodology using a new population based meta heuristic
approach namely Artificial Bee Colony algorithm(ABC) for
the placement of Distributed Generators(DG) in the radial
distribution systems to reduce the real power losses and to
improve the voltage profile, voltage sag mitigation. The power
loss reduction is important factor for utility companies because
it is directly proportional to the company benefits in a
competitive electricity market, while reaching the better power
quality standards is too important as it has vital effect on
customer orientation. In this paper an ABC algorithm is
developed to gain these goals all together. In order to evaluate
sag mitigation capability of the proposed algorithm, voltage
in voltage sensitive buses is investigated. An existing 20KV
network has been chosen as test network and results are
compared with the proposed method in the radial distribution
system.
Line Losses in the 14-Bus Power System Network using UPFCIDES Editor
Controlling power flow in modern power systems
can be made more flexible by the use of recent developments
in power electronic and computing control technology. The
Unified Power Flow Controller (UPFC) is a Flexible AC
transmission system (FACTS) device that can control all the
three system variables namely line reactance, magnitude and
phase angle difference of voltage across the line. The UPFC
provides a promising means to control power flow in modern
power systems. Essentially the performance depends on proper
control setting achievable through a power flow analysis
program. This paper presents a reliable method to meet the
requirements by developing a Newton-Raphson based load
flow calculation through which control settings of UPFC can
be determined for the pre-specified power flow between the
lines. The proposed method keeps Newton-Raphson Load Flow
(NRLF) algorithm intact and needs (little modification in the
Jacobian matrix). A MATLAB program has been developed to
calculate the control settings of UPFC and the power flow
between the lines after the load flow is converged. Case studies
have been performed on IEEE 5-bus system and 14-bus system
to show that the proposed method is effective. These studies
indicate that the method maintains the basic NRLF properties
such as fast computational speed, high degree of accuracy and
good convergence rate.
Study of Structural Behaviour of Gravity Dam with Various Features of Gallery...IDES Editor
The size and shape of opening in dam causes the
stress concentration, it also causes the stress variation in the
rest of the dam cross section. The gravity method of the analysis
does not consider the size of opening and the elastic property
of dam material. Thus the objective of study is comprises of
the Finite Element Method which considers the size of
opening, elastic property of material, and stress distribution
because of geometric discontinuity in cross section of dam.
Stress concentration inside the dam increases with the opening
in dam which results in the failure of dam. Hence it is
necessary to analyses large opening inside the dam. By making
the percentage area of opening constant and varying size and
shape of opening the analysis is carried out. For this purpose
a section of Koyna Dam is considered. Dam is defined as a
plane strain element in FEM, based on geometry and loading
condition. Thus this available information specified our path
of approach to carry out 2D plane strain analysis. The results
obtained are then compared mutually to get most efficient
way of providing large opening in the gravity dam.
Assessing Uncertainty of Pushover Analysis to Geometric ModelingIDES Editor
Pushover Analysis a popular tool for seismic
performance evaluation of existing and new structures and is
nonlinear Static procedure where in monotonically increasing
loads are applied to the structure till the structure is unable
to resist the further load .During the analysis, whatever the
strength of concrete and steel is adopted for analysis of
structure may not be the same when real structure is
constructed and the pushover analysis results are very sensitive
to material model adopted, geometric model adopted, location
of plastic hinges and in general to procedure followed by the
analyzer. In this paper attempt has been made to assess
uncertainty in pushover analysis results by considering user
defined hinges and frame modeled as bare frame and frame
with slab modeled as rigid diaphragm and results compared
with experimental observations. Uncertain parameters
considered includes the strength of concrete, strength of steel
and cover to the reinforcement which are randomly generated
and incorporated into the analysis. The results are then
compared with experimental observations.
Secure Multi-Party Negotiation: An Analysis for Electronic Payments in Mobile...IDES Editor
This document summarizes and analyzes secure multi-party negotiation protocols for electronic payments in mobile computing. It presents a framework for secure multi-party decision protocols using lightweight implementations. The main focus is on synchronizing security features to avoid agreement manipulation and reduce user traffic. The paper describes negotiation between an auctioneer and bidders, showing multiparty security is better than existing systems. It analyzes the performance of encryption algorithms like ECC, XTR, and RSA for use in the multiparty negotiation protocols.
Selfish Node Isolation & Incentivation using Progressive ThresholdsIDES Editor
The problems associated with selfish nodes in
MANET are addressed by a collaborative watchdog approach
which reduces the detection time for selfish nodes thereby
improves the performance and accuracy of watchdogs[1]. In
the related works they make use of credit based systems, reputation
based mechanisms, pathrater and watchdog mechanism
to detect such selfish nodes. In this paper we follow an approach
of collaborative watchdog which reduces the detection
time for selfish nodes and also involves the removal of such
selfish nodes based on some progressively assessed thresholds.
The threshold gives the nodes a chance to stop misbehaving
before it is permanently deleted from the network.
The node passes through several isolation processes before it
is permanently removed. Another version of AODV protocol
is used here which allows the simulation of selfish nodes in
NS2 by adding or modifying log files in the protocol.
Various OSI Layer Attacks and Countermeasure to Enhance the Performance of WS...IDES Editor
Wireless sensor networks are networks having non
wired infrastructure and dynamic topology. In OSI model each
layer is prone to various attacks, which halts the performance
of a network .In this paper several attacks on four layers of
OSI model are discussed and security mechanism is described
to prevent attack in network layer i.e wormhole attack. In
Wormhole attack two or more malicious nodes makes a covert
channel which attracts the traffic towards itself by depicting a
low latency link and then start dropping and replaying packets
in the multi-path route. This paper proposes promiscuous mode
method to detect and isolate the malicious node during
wormhole attack by using Ad-hoc on demand distance vector
routing protocol (AODV) with omnidirectional antenna. The
methodology implemented notifies that the nodes which are
not participating in multi-path routing generates an alarm
message during delay and then detects and isolate the
malicious node from network. We also notice that not only
the same kind of attacks but also the same kind of
countermeasures can appear in multiple layer. For example,
misbehavior detection techniques can be applied to almost all
the layers we discussed.
Responsive Parameter based an AntiWorm Approach to Prevent Wormhole Attack in...IDES Editor
The recent advancements in the wireless technology
and their wide-spread deployment have made remarkable
enhancements in efficiency in the corporate and industrial
and Military sectors The increasing popularity and usage of
wireless technology is creating a need for more secure wireless
Ad hoc networks. This paper aims researched and developed
a new protocol that prevents wormhole attacks on a ad hoc
network. A few existing protocols detect wormhole attacks but
they require highly specialized equipment not found on most
wireless devices. This paper aims to develop a defense against
wormhole attacks as an Anti-worm protocol which is based on
responsive parameters, that does not require as a significant
amount of specialized equipment, trick clock synchronization,
no GPS dependencies.
Cloud Security and Data Integrity with Client Accountability FrameworkIDES Editor
This document summarizes a proposed cloud security and data integrity framework that provides client accountability. The framework aims to address issues like lack of user control over cloud data, need for data transparency and tracking, and ensuring data integrity. It proposes using JAR (Java Archive) files for data sharing due to benefits like portability. The framework incorporates client-side verification using MD5 hashing, digital signature-based authentication of JAR files, and use of HMAC to ensure data integrity. It also uses password-based encryption of log files to keep them tamper-proof. The framework is intended to provide both accountability and security for data sharing in cloud environments.
Genetic Algorithm based Layered Detection and Defense of HTTP BotnetIDES Editor
A System state in HTTP botnet uses HTTP protocol
for the creation of chain of Botnets thereby compromising
other systems. By using HTTP protocol and port number 80,
attacks can not only be hidden but also pass through the
firewall without being detected. The DPR based detection
leads to better analysis of botnet attacks [3]. However, it
provides only probabilistic detection of the attacker and also
time consuming and error prone. This paper proposes a Genetic
algorithm based layered approach for detecting as well as
preventing botnet attacks. The paper reviews p2p firewall
implementation which forms the basis of filtering.
Performance evaluation is done based on precision, F-value
and probability. Layered approach reduces the computation
and overall time requirement [7]. Genetic algorithm promises
a low false positive rate.
Enhancing Data Storage Security in Cloud Computing Through SteganographyIDES Editor
This document summarizes a research paper that proposes a method for enhancing data security in cloud computing through steganography. The method hides user data in digital images stored on cloud servers. When data needs to be accessed, it is extracted from the images. The document outlines the cloud architecture and security issues addressed. It then describes the proposed system architecture, security model, and data storage and retrieval process. Data is partitioned and hidden in multiple images to improve security. The goal is to prevent unauthorized access to user data stored on cloud servers.
The main tasks of a Wireless Sensor Network
(WSN) are data collection from its nodes and communication
of this data to the base station (BS). The protocols used for
communication among the WSN nodes and between the WSN
and the BS, must consider the resource constraints of nodes,
battery energy, computational capabilities and memory. The
WSN applications involve unattended operation of the network
over an extended period of time. In order to extend the lifetime
of a WSN, efficient routing protocols need to be adopted. The
proposed low power routing protocol based on tree-based
network structure reliably forwards the measured data towards
the BS using TDMA. An energy consumption analysis of the
WSN making use of this protocol is also carried out. It is
found that the network is energy efficient with an average
duty cycle of 0:7% for the WSN nodes. The OmNET++
simulation platform along with MiXiM framework is made
use of.
Permutation of Pixels within the Shares of Visual Cryptography using KBRP for...IDES Editor
The security of authentication of internet based
co-banking services should not be susceptible to high risks.
The passwords are highly vulnerable to virus attacks due to
the lack of high end embedding of security methods. In order
for the passwords to be more secure, people are generally
compelled to select jumbled up character based passwords
which are not only less memorable but are also equally prone
to insecurity. Multiple use of distributed shares has been
studied to solve the problem of authentication by algorithms
based on thresholding of pixels in image processing and visual
cryptography concepts where the subset of shares is considered
for the recovery of the original image for authentication using
correlation function[1][2].The main disadvantage in the above
study is the plain storage of shares and also one of the shares
is being supplied to the customer, which will lead to the
possibility of misuse by a third party. This paper proposes a
technique for scrambling of pixels by key based random
permutation (KBRP) within the shares before the
authentication has been attempted. Total number of shares to
be created is dependent on the multiplicity of ownership of
the account. By this method the problem of uncertainty among
the customers with regard to security, storage, retrieval of
holding of half of the shares is minimized.
This paper presents a trifocal Rotman Lens Design
approach. The effects of focal ratio and element spacing on
the performance of Rotman Lens are described. A three beam
prototype feeding 4 element antenna array working in L-band
has been simulated using RLD v1.7 software. Simulated
results show that the simulated lens has a return loss of –
12.4dB at 1.8GHz. Beam to array port phase error variation
with change in the focal ratio and element spacing has also
been investigated.
Band Clustering for the Lossless Compression of AVIRIS Hyperspectral ImagesIDES Editor
Hyperspectral images can be efficiently compressed
through a linear predictive model, as for example the one
used in the SLSQ algorithm. In this paper we exploit this
predictive model on the AVIRIS images by individuating,
through an off-line approach, a common subset of bands, which
are not spectrally related with any other bands. These bands
are not useful as prediction reference for the SLSQ 3-D
predictive model and we need to encode them via other
prediction strategies which consider only spatial correlation.
We have obtained this subset by clustering the AVIRIS bands
via the clustering by compression approach. The main result
of this paper is the list of the bands, not related with the
others, for AVIRIS images. The clustering trees obtained for
AVIRIS and the relationship among bands they depict is also
an interesting starting point for future research.
Microelectronic Circuit Analogous to Hydrogen Bonding Network in Active Site ...IDES Editor
A microelectronic circuit of block-elements
functionally analogous to two hydrogen bonding networks is
investigated. The hydrogen bonding networks are extracted
from â-lactamase protein and are formed in its active site.
Each hydrogen bond of the network is described in equivalent
electrical circuit by three or four-terminal block-element.
Each block-element is coded in Matlab. Static and dynamic
analyses are performed. The resultant microelectronic circuit
analogous to the hydrogen bonding network operates as
current mirror, sine pulse source, triangular pulse source as
well as signal modulator.
Texture Unit based Monocular Real-world Scene Classification using SOM and KN...IDES Editor
In this paper a method is proposed to discriminate
real world scenes in to natural and manmade scenes of similar
depth. Global-roughness of a scene image varies as a function
of image-depth. Increase in image depth leads to increase in
roughness in manmade scenes; on the contrary natural scenes
exhibit smooth behavior at higher image depth. This particular
arrangement of pixels in scene structure can be well explained
by local texture information in a pixel and its neighborhood.
Our proposed method analyses local texture information of a
scene image using texture unit matrix. For final classification
we have used both supervised and unsupervised learning using
K-Nearest Neighbor classifier (KNN) and Self Organizing
Map (SOM) respectively. This technique is useful for online
classification due to very less computational complexity.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Building RAG with self-deployed Milvus vector database and Snowpark Container...Zilliz
This talk will give hands-on advice on building RAG applications with an open-source Milvus database deployed as a docker container. We will also introduce the integration of Milvus with Snowpark Container Services.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).