Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
An Enhanced Approach for Detecting User's Behavior Applying Country-Wise Loca...IJSRD
The development of the web in past few years has created a lot of challenge in this field. The new work in this field is the search of the data in a search tree pattern based on tree. Various sequential mining algorithms have been devoloped till date. Web usage mining is used to operate the web server logs, that contains the navigation history of the user. Recommendater system is explained properly with the explanation of whole procedure of the recommendater system. The search results of the data leads to the proper ad efficient search. But the problem was the time utilization and the search results generated from them. So, a new local search algorithm is proposed for country-wise search that makes the searching more efficient on local results basis. This approach has lead to an advancement in the search based methods and the results generated.
Identifying the Number of Visitors to improve Website Usability from Educatio...Editor IJCATR
Web usage mining deals with understanding the Visitor’s behaviour with a Website. It helps in understanding the concerns
such as present and future probability of every website user, relationship between behaviour and website usability. It has different
branches such as web content mining, web structure and web usage mining. The focus of this paper is on web mining usage patterns of
an educational institution web log data. There are three types of web related log data namely web access log, error log and proxy log
data. In this paper web access log data has been used as dataset because the web access log data is the typical source of navigational
behaviour of the website visitor. The study of web server log analysis is helpful in applying the web mining techniques.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Recommendation generation by integrating sequential pattern mining and semanticseSAT Journals
Abstract As the Internet usage keeps increasing, the number of web sites and hence the number of web pages also keeps increasing. A recommendation system can be used to provide personalized web service by suggesting the pages that are likely to be accessed in future. Most of the recommendation systems are based on association rule mining or based on keywords. Using the association rule mining the prediction rate is less as it doesn’t take into account the order of access of the web pages by the users. The recommendation systems that are key-word based provides lesser relevant results. This paper proposes a recommendation system that uses the advantages of sequential pattern mining and semantics over the association rule mining and keyword based systems respectively. Keywords: Sequential Pattern Mining, Taxonomy, Apriori-All, CS-Mine, Semantic, Clustering
Web Page Recommendation Using Web MiningIJERA Editor
On World Wide Web various kind of content are generated in huge amount, so to give relevant result to user web recommendation become important part of web application. On web different kind of web recommendation are made available to user every day that includes Image, Video, Audio, query suggestion and web page. In this paper we are aiming at providing framework for web page recommendation. 1) First we describe the basics of web mining, types of web mining. 2) Details of each web mining technique.3)We propose the architecture for the personalized web page recommendation.
An Enhanced Approach for Detecting User's Behavior Applying Country-Wise Loca...IJSRD
The development of the web in past few years has created a lot of challenge in this field. The new work in this field is the search of the data in a search tree pattern based on tree. Various sequential mining algorithms have been devoloped till date. Web usage mining is used to operate the web server logs, that contains the navigation history of the user. Recommendater system is explained properly with the explanation of whole procedure of the recommendater system. The search results of the data leads to the proper ad efficient search. But the problem was the time utilization and the search results generated from them. So, a new local search algorithm is proposed for country-wise search that makes the searching more efficient on local results basis. This approach has lead to an advancement in the search based methods and the results generated.
Identifying the Number of Visitors to improve Website Usability from Educatio...Editor IJCATR
Web usage mining deals with understanding the Visitor’s behaviour with a Website. It helps in understanding the concerns
such as present and future probability of every website user, relationship between behaviour and website usability. It has different
branches such as web content mining, web structure and web usage mining. The focus of this paper is on web mining usage patterns of
an educational institution web log data. There are three types of web related log data namely web access log, error log and proxy log
data. In this paper web access log data has been used as dataset because the web access log data is the typical source of navigational
behaviour of the website visitor. The study of web server log analysis is helpful in applying the web mining techniques.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Recommendation generation by integrating sequential pattern mining and semanticseSAT Journals
Abstract As the Internet usage keeps increasing, the number of web sites and hence the number of web pages also keeps increasing. A recommendation system can be used to provide personalized web service by suggesting the pages that are likely to be accessed in future. Most of the recommendation systems are based on association rule mining or based on keywords. Using the association rule mining the prediction rate is less as it doesn’t take into account the order of access of the web pages by the users. The recommendation systems that are key-word based provides lesser relevant results. This paper proposes a recommendation system that uses the advantages of sequential pattern mining and semantics over the association rule mining and keyword based systems respectively. Keywords: Sequential Pattern Mining, Taxonomy, Apriori-All, CS-Mine, Semantic, Clustering
Web Page Recommendation Using Web MiningIJERA Editor
On World Wide Web various kind of content are generated in huge amount, so to give relevant result to user web recommendation become important part of web application. On web different kind of web recommendation are made available to user every day that includes Image, Video, Audio, query suggestion and web page. In this paper we are aiming at providing framework for web page recommendation. 1) First we describe the basics of web mining, types of web mining. 2) Details of each web mining technique.3)We propose the architecture for the personalized web page recommendation.
A Review on Pattern Discovery Techniques of Web Usage MiningIJERA Editor
In the recent years with the development of Internet technology the growth of World Wide Web exceeded all expectations. A lot of information is available in different formats and retrieving interesting content has become a very difficult task. One possible approach to solve this problem is Web Usage Mining (WUM), the important application of Web Mining. Extracting the hidden knowledge in the log files of a web server, recognizing various interests of web users, discovering customer behavior while at the site are normally referred as the applications of web usage mining. In this paper we provide an updated focused survey on techniques of web usage mining.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
BIDIRECTIONAL GROWTH BASED MINING AND CYCLIC BEHAVIOUR ANALYSIS OF WEB SEQUEN...ijdkp
Web sequential patterns are important for analyzing and understanding users’ behaviour to improve the
quality of service offered by the World Wide Web. Web Prefetching is one such technique that utilizes
prefetching rules derived through Cyclic Model Analysis of the mined Web sequential patterns. The more
accurate the prediction and more satisfying the results of prefetching if we use a highly efficient and
scalable mining technique such as the Bidirectional Growth based Directed Acyclic Graph. In this paper,
we propose a novel algorithm called Bidirectional Growth based mining Cyclic behavior Analysis of web
sequential Patterns (BGCAP) that effectively combines these strategies to generate prefetching rules in the
form of 2-sequence patterns with Periodicity and threshold of Cyclic Behaviour that can be utilized to
effectively prefetch Web pages, thus reducing the users’ perceived latency. As BGCAP is based on
Bidirectional pattern growth, it performs only (log n+1) levels of recursion for mining n Web sequential
patterns. Our experimental results show that prefetching rules generated using BGCAP is 5-10% faster for
different data sizes and 10-15% faster for a fixed data size than TD-Mine. In addition, BGCAP generates
about 5-15% more prefetching rules than TD-Mine.
In this world of information technology, everyone has the tendency to do business electronically. Today
lot of businesses are happening on World Wide Web (WWW), it is very important for the website owner to
provide a better platform to attract more customers for their site. Providing information in a better way is
the solution to bring more customers or users. Customer is the end-user, who accessing the information
in a way it yields some credit to the web site owners. In this paper we define web mining and present a
method to utilize web mining in a better way to know the users and website behaviour which in turn
enhance the web site information to attract more users. This paper also presents an overview of the
various researches done on pattern extraction, web content mining and how it can be taken as a catalyst
for E-business.
AN INTELLIGENT OPTIMAL GENETIC MODEL TO INVESTIGATE THE USER USAGE BEHAVIOUR ...ijdkp
The unexpected wide spread use of WWW and dynamically increasing nature of the web creates new
challenges in the web mining since the data in the web inherently unlabelled, incomplete, non linear, and
heterogeneous. The investigation of user usage behaviour on WWW is real time problem which involves
multiple conflicting measures of performance. These measures make not only computational intensive but
also needs to the possibility of be unable to find the exact solution. Unfortunately, the conventional methods
are limited to optimization problems due to the absence of semantic certainty and presence of human
intervention. In handling such data and overcome the limitations of conventional methodologies it is
necessary to use a soft computing model that can work intelligently to attain optimal solution.
Comparable Analysis of Web Mining Categoriestheijes
Web Data Mining is the current field of analysis which is a combination of two research area known as Data Mining and World Wide Web. Web Data Mining research associates with various research diversities like Database, Artificial Intelligence and Information redeem. The mining techniques are categorized into various categories namely Web Content Mining, Web Structure Mining and Web Usage Mining. In this work, analysis of mining techniques are done. From the analysis it has been concluded that Web Content Mining has unstructured or semi- structure view of data whereas Web Structure Mining have linked structure and Web Usage Mining mainly includes interaction.
A Comparative Study of Recommendation System Using Web Usage Mining Editor IJMTER
Web Mining is one of the Developing field in research. Exact custom of the Web is to get the
beneficial material in the sites. To reduce the work time of user the Web Usage Mining (WUM) technique
is introduced. In this Technique use Web Page recommendation for the Web request from the user. For
the recommendation system in Web Usage Mining (WUM) variousauthor has introduce different
Algorithm and technique to improve the user interest in surfing the Web. Web log files are used todefine
the user interest and there next recommend page to view.The data stored in the web log file consist of
large amount oferoded, incomplete, and unnecessary information. So, the Web log files have to preprocess, customize, and to clean the data. In this paper we will survey different recommendation technique
to identify the issues in web surfing and to improve web usagemining (WUM) pre-processing for pattern
mining and analysis.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
A Review on Pattern Discovery Techniques of Web Usage MiningIJERA Editor
In the recent years with the development of Internet technology the growth of World Wide Web exceeded all expectations. A lot of information is available in different formats and retrieving interesting content has become a very difficult task. One possible approach to solve this problem is Web Usage Mining (WUM), the important application of Web Mining. Extracting the hidden knowledge in the log files of a web server, recognizing various interests of web users, discovering customer behavior while at the site are normally referred as the applications of web usage mining. In this paper we provide an updated focused survey on techniques of web usage mining.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
BIDIRECTIONAL GROWTH BASED MINING AND CYCLIC BEHAVIOUR ANALYSIS OF WEB SEQUEN...ijdkp
Web sequential patterns are important for analyzing and understanding users’ behaviour to improve the
quality of service offered by the World Wide Web. Web Prefetching is one such technique that utilizes
prefetching rules derived through Cyclic Model Analysis of the mined Web sequential patterns. The more
accurate the prediction and more satisfying the results of prefetching if we use a highly efficient and
scalable mining technique such as the Bidirectional Growth based Directed Acyclic Graph. In this paper,
we propose a novel algorithm called Bidirectional Growth based mining Cyclic behavior Analysis of web
sequential Patterns (BGCAP) that effectively combines these strategies to generate prefetching rules in the
form of 2-sequence patterns with Periodicity and threshold of Cyclic Behaviour that can be utilized to
effectively prefetch Web pages, thus reducing the users’ perceived latency. As BGCAP is based on
Bidirectional pattern growth, it performs only (log n+1) levels of recursion for mining n Web sequential
patterns. Our experimental results show that prefetching rules generated using BGCAP is 5-10% faster for
different data sizes and 10-15% faster for a fixed data size than TD-Mine. In addition, BGCAP generates
about 5-15% more prefetching rules than TD-Mine.
In this world of information technology, everyone has the tendency to do business electronically. Today
lot of businesses are happening on World Wide Web (WWW), it is very important for the website owner to
provide a better platform to attract more customers for their site. Providing information in a better way is
the solution to bring more customers or users. Customer is the end-user, who accessing the information
in a way it yields some credit to the web site owners. In this paper we define web mining and present a
method to utilize web mining in a better way to know the users and website behaviour which in turn
enhance the web site information to attract more users. This paper also presents an overview of the
various researches done on pattern extraction, web content mining and how it can be taken as a catalyst
for E-business.
AN INTELLIGENT OPTIMAL GENETIC MODEL TO INVESTIGATE THE USER USAGE BEHAVIOUR ...ijdkp
The unexpected wide spread use of WWW and dynamically increasing nature of the web creates new
challenges in the web mining since the data in the web inherently unlabelled, incomplete, non linear, and
heterogeneous. The investigation of user usage behaviour on WWW is real time problem which involves
multiple conflicting measures of performance. These measures make not only computational intensive but
also needs to the possibility of be unable to find the exact solution. Unfortunately, the conventional methods
are limited to optimization problems due to the absence of semantic certainty and presence of human
intervention. In handling such data and overcome the limitations of conventional methodologies it is
necessary to use a soft computing model that can work intelligently to attain optimal solution.
Comparable Analysis of Web Mining Categoriestheijes
Web Data Mining is the current field of analysis which is a combination of two research area known as Data Mining and World Wide Web. Web Data Mining research associates with various research diversities like Database, Artificial Intelligence and Information redeem. The mining techniques are categorized into various categories namely Web Content Mining, Web Structure Mining and Web Usage Mining. In this work, analysis of mining techniques are done. From the analysis it has been concluded that Web Content Mining has unstructured or semi- structure view of data whereas Web Structure Mining have linked structure and Web Usage Mining mainly includes interaction.
A Comparative Study of Recommendation System Using Web Usage Mining Editor IJMTER
Web Mining is one of the Developing field in research. Exact custom of the Web is to get the
beneficial material in the sites. To reduce the work time of user the Web Usage Mining (WUM) technique
is introduced. In this Technique use Web Page recommendation for the Web request from the user. For
the recommendation system in Web Usage Mining (WUM) variousauthor has introduce different
Algorithm and technique to improve the user interest in surfing the Web. Web log files are used todefine
the user interest and there next recommend page to view.The data stored in the web log file consist of
large amount oferoded, incomplete, and unnecessary information. So, the Web log files have to preprocess, customize, and to clean the data. In this paper we will survey different recommendation technique
to identify the issues in web surfing and to improve web usagemining (WUM) pre-processing for pattern
mining and analysis.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
International conference On Computer Science And technologyanchalsinghdm
ICGCET 2019 | 5th International Conference on Green Computing and Engineering Technologies. The conference will be held on 7th September - 9th September 2019 in Morocco. International Conference On Engineering Technology
The conference aims to promote the work of researchers, scientists, engineers and students from across the world on advancement in electronic and computer systems.
Web Usage Mining: A Survey on User's Navigation Pattern from Web Logsijsrd.com
With an expontial growth of World Wide Web, there are so many information overloaded and it became hard to find out data according to need. Web usage mining is a part of web mining, which deal with automatic discovery of user navigation pattern from web log. This paper presents an overview of web mining and also provide navigation pattern from classification and clustering algorithm for web usage mining. Web usage mining contain three important task namely data preprocessing, pattern discovery and pattern analysis based on discovered pattern. And also contain the comparative study of web mining techniques.
Web is a collection of inter-related files on one or more web servers while web mining means extracting
valuable information from web databases. Web mining is one of the data mining domains where data
mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer
behaviour, evaluate a particular website based on the information which is stored in web log files. Web
mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library.
Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase
the complexity of dealing information from different web service providers. The collection of information
becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web is a collection of inter-related files on one or more web servers while web mining means extracting
valuable information from web databases. Web mining is one of the data mining domains where data
mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer
behaviour, evaluate a particular website based on the information which is stored in web log files. Web
mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library.
Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase
the complexity of dealing information from different web service providers. The collection of information
becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web is a collection of inter-related files on one or more web servers while web mining means extracting
valuable information from web databases. Web mining is one of the data mining domains where data
mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer
behaviour, evaluate a particular website based on the information which is stored in web log files. Web
mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library.
Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase
the complexity of dealing information from different web service providers. The collection of information
becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web is a collection of inter-related files on one or more web servers while web mining means extracting valuable information from web databases. Web mining is one of the data mining domains where data mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer behaviour, evaluate a particular website based on the information which is stored in web log files. Web mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library. Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase the complexity of dealing information from different web service providers. The collection of information becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web is a collection of inter-related files on one or more web servers while web mining means extracting
valuable information from web databases. Web mining is one of the data mining domains where data
mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer
behaviour, evaluate a particular website based on the information which is stored in web log files. Web
mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library.
Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase
the complexity of dealing information from different web service providers. The collection of information
becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web is a collection of inter-related files on one or more web servers while web mining means extracting valuable information from web databases. Web mining is one of the data mining domains where data mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer behaviour, evaluate a particular website based on the information which is stored in web log files. Web mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library. Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase the complexity of dealing information from different web service providers. The collection of information becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web is a collection of inter-related files on one or more web servers while web mining means extracting valuable information from web databases. Web mining is one of the data mining domains where data mining techniques are used for extracting information from the web servers. The web data includes web
pages, web links, objects on the web and web logs. Web mining is used to understand the customer behaviour, evaluate a particular website based on the information which is stored in web log files. Web mining is evaluated by using data mining techniques, namely classification, clustering, and association
rules. It has some beneficial areas or applications such as Electronic commerce, E-learning, Egovernment, E-policies, E-democracy, Electronic business, security, crime investigation and digital library. Retrieving the required web page from the web efficiently and effectively becomes a challenging task
because web is made up of unstructured data, which delivers the large amount of information and increase the complexity of dealing information from different web service providers. The collection of information becomes very hard to find, extract, filter or evaluate the relevant information for the users. In this paper,
we have studied the basic concepts of web mining, classification, processes and issues. In addition to this,
this paper also analyzed the web mining research challenges.
Web personalization using clustering of web usage dataijfcstjournal
The exponential growth in the number and the complexity of information resources and services on the Web
has made log data an indispensable resource to characterize the users for Web-based environment. It
creates information of related web data in the form of hierarchy structure through approximation. This
hierarchy structure can be used as the input for a variety of data mining tasks such as clustering,
association rule mining, sequence mining etc.
In this paper, we present an approach for personalizing web user environment dynamically when he
interacting with web by clustering of web usage data using concept hierarchy. The system is inferred from
the web server’s access logs by means of data and web usage mining techniques to extract the information
about users. The extracted knowledge is used for the purpose of offering a personalized view of the
services to users.
A Study of Pattern Analysis Techniques of Web Usageijbuiiir1
Web mining is the most important application of data mining techniques to extract knowledge from web data including web document, hyperlinks between documents, usage logs of web sites etc. Web mining has been explored to a vast degree and different techniques have been proposed for a huge variety of applications that includes search engine enhancement, optimization of web services, Business Intelligence, B2B and B2C business etc. Most research on web mining has been from a �process-centric� point of view which defined web mining as a sequence of tasks. In this paper, we highlight the significance of studying the evolving nature of the web pattern analysis (WPA). Web usage mining is used to discover interesting user navigation patterns and can be applied to many real-world problems, such as improving web sites/pages. A Web usage mining system performs five major tasks: i) data collection ii) information filtering iii) pattern discovery iv) pattern analysis and visualization techniques, and v) Knowledge Query Mechanism (KQM). Each task is explained in detail and its related technologies are introduced. The web mining research is a converging research area from several research communities, such as database system, information retrieval, information extraction and artificial intelligence. In this paper we implement how web usage mining techniques can be applied for the customization i.e. web visualization
Data mining refers to the process of analysing the data from different perspectives and summarizing it into useful information.
Data mining software is one of the number of tools used for analysing data. It allows users to analyse from many different dimensions and angles, categorize it, and summarize the relationship identified.
Data mining is about technique for finding and describing Structural Patterns in data.
Data mining is the process of finding correlation or patterns among fields in large relational databases.
The process of extracting valid, previously unknown, comprehensible , and actionable information from large databases and using it to make crucial business decisions.
Semantically enriched web usage mining for predicting user future movementsIJwest
Explosive and quick growth of the World Wide Web has resulted in intricate Web sites, demanding
enhanced user skills and sophisticated tools to help the Web user to find the desi
red information. Finding
desired information on the Web has become a critical ingredient of everyday personal, educational, and
business life. Thus, there is a demand for more sophisticated tools to help the user to navigate a Web site
and find the desired
information. The users must be provided with information and services specific to
their needs, rather than an undiffere
ntiated mass of information.
For discovering interesting and frequent
navigation patterns from Web server logs many Web usage mining te
chniques have been applied. The
recommendation accuracy of solely usage based techniques can be improved by integrating Web site
content and site structure in the personalization process.
Herein, we propose Semantically enriched Web Usage Mining method (S
WUM), which combines the fields
of Web Usage Mining and Semantic Web. In the proposed method, the undirected graph derived from
usage data is enriched with rich semantic information extracted from the Web pages and the Web site
structure. The experimental
results show that the SWUM generates accurate recommendations with
integration of usage, semantic data and Web site structure. The results shows that proposed method is able
to achieve 10
-
20%
better accuracy than the solely usage based model, and 5
-
8% bet
ter than an ontology
based model.
ANALYTICAL IMPLEMENTATION OF WEB STRUCTURE MINING USING DATA ANALYSIS IN ONLI...IAEME Publication
In today ’s global business, the web has been the most important means of communication. Clients and customers may find their products online, which is a benefit of doing business online. Web mining is the process of using data mining tools to analyse and extract the information from a Web pages and applications autonomously. Many firms use web structure mining to generate suitable predictions and judgments for business growth, productivity, manufacturing techniques, and more utilizing data mining business strategies. In the online booking domain, optimum web data mining analysis of web structure is a crucial component that gives a systematic manner of new application towards real-time data with various levels of implications. Web structure mining emphases on the construction of the web's hyperlinks. Linkage administration that is done correctly can lead to future connections, which can therefore increase the prediction performance of learnt models. A increased interest in Web mining, structural analysis research has expanded, resulting in a new research area that sits at the crossroads of work in the network analysis, hyperlink and the web mining, structural training, and empirical software design techniques, as well as graph mining. Web structure mining is the development of determining structure data from the web. The proposed WSM approach is a system of finding the structure of data stored over the Web. Web structure mining can encourage the clients to recover the significant records by breaking down the connection situated structure of Web content. Web structure mining has been one of the most important resources for information extraction and the knowledge discovery as the amount of data available online has increased.
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksIJERD Editor
Distributed Denial of Service (DDoS) Attacks became a massive threat to the Internet. Traditional
Architecture of internet is vulnerable to the attacks like DDoS. Attacker primarily acquire his army of Zombies,
then that army will be instructed by the Attacker that when to start an attack and on whom the attack should be
done. In this paper, different techniques which are used to perform DDoS Attacks, Tools that were used to
perform Attacks and Countermeasures in order to detect the attackers and eliminate the Bandwidth Distributed
Denial of Service attacks (B-DDoS) are reviewed. DDoS Attacks were done by using various Flooding
techniques which are used in DDoS attack.
The main purpose of this paper is to design an architecture which can reduce the Bandwidth
Distributed Denial of service Attack and make the victim site or server available for the normal users by
eliminating the zombie machines. Our Primary focus of this paper is to dispute how normal machines are
turning into zombies (Bots), how attack is been initiated, DDoS attack procedure and how an organization can
save their server from being a DDoS victim. In order to present this we implemented a simulated environment
with Cisco switches, Routers, Firewall, some virtual machines and some Attack tools to display a real DDoS
attack. By using Time scheduling, Resource Limiting, System log, Access Control List and some Modular
policy Framework we stopped the attack and identified the Attacker (Bot) machines
Hearing loss is one of the most common human impairments. It is estimated that by year 2015 more
than 700 million people will suffer mild deafness. Most can be helped by hearing aid devices depending on the
severity of their hearing loss. This paper describes the implementation and characterization details of a dual
channel transmitter front end (TFE) for digital hearing aid (DHA) applications that use novel micro
electromechanical- systems (MEMS) audio transducers and ultra-low power-scalable analog-to-digital
converters (ADCs), which enable a very-low form factor, energy-efficient implementation for next-generation
DHA. The contribution of the design is the implementation of the dual channel MEMS microphones and powerscalable
ADC system.
Influence of tensile behaviour of slab on the structural Behaviour of shear c...IJERD Editor
-A composite beam is composed of a steel beam and a slab connected by means of shear connectors
like studs installed on the top flange of the steel beam to form a structure behaving monolithically. This study
analyzes the effects of the tensile behavior of the slab on the structural behavior of the shear connection like slip
stiffness and maximum shear force in composite beams subjected to hogging moment. The results show that the
shear studs located in the crack-concentration zones due to large hogging moments sustain significantly smaller
shear force and slip stiffness than the other zones. Moreover, the reduction of the slip stiffness in the shear
connection appears also to be closely related to the change in the tensile strain of rebar according to the increase
of the load. Further experimental and analytical studies shall be conducted considering variables such as the
reinforcement ratio and the arrangement of shear connectors to achieve efficient design of the shear connection
in composite beams subjected to hogging moment.
Gold prospecting using Remote Sensing ‘A case study of Sudan’IJERD Editor
Gold has been extracted from northeast Africa for more than 5000 years, and this may be the first
place where the metal was extracted. The Arabian-Nubian Shield (ANS) is an exposure of Precambrian
crystalline rocks on the flanks of the Red Sea. The crystalline rocks are mostly Neoproterozoic in age. ANS
includes the nations of Israel, Jordan. Egypt, Saudi Arabia, Sudan, Eritrea, Ethiopia, Yemen, and Somalia.
Arabian Nubian Shield Consists of juvenile continental crest that formed between 900 550 Ma, when intra
oceanic arc welded together along ophiolite decorated arc. Primary Au mineralization probably developed in
association with the growth of intra oceanic arc and evolution of back arc. Multiple episodes of deformation
have obscured the primary metallogenic setting, but at least some of the deposits preserve evidence that they
originate as sea floor massive sulphide deposits.
The Red Sea Hills Region is a vast span of rugged, harsh and inhospitable sector of the Earth with
inimical moon-like terrain, nevertheless since ancient times it is famed to be an abode of gold and was a major
source of wealth for the Pharaohs of ancient Egypt. The Pharaohs old workings have been periodically
rediscovered through time. Recent endeavours by the Geological Research Authority of Sudan led to the
discovery of a score of occurrences with gold and massive sulphide mineralizations. In the nineties of the
previous century the Geological Research Authority of Sudan (GRAS) in cooperation with BRGM utilized
satellite data of Landsat TM using spectral ratio technique to map possible mineralized zones in the Red Sea
Hills of Sudan. The outcome of the study mapped a gossan type gold mineralization. Band ratio technique was
applied to Arbaat area and a signature of alteration zone was detected. The alteration zones are commonly
associated with mineralization. The alteration zones are commonly associated with mineralization. A filed check
confirmed the existence of stock work of gold bearing quartz in the alteration zone. Another type of gold
mineralization that was discovered using remote sensing is the gold associated with metachert in the Atmur
Desert.
Reducing Corrosion Rate by Welding DesignIJERD Editor
The paper addresses the importance of welding design to prevent corrosion at steel. Welding is
used to join pipe, profiles at bridges, spindle, and a lot more part of engineering construction. The
problems happened associated with welding are common issues in these fields, especially corrosion.
Corrosion can be reduced with many methods, they are painting, controlling humidity, and also good
welding design. In the research, it can be found that reducing residual stress on the welding can be
solved in corrosion rate reduction problem.
Preheating on 500oC and 600oC give better condition to reduce corosion rate than condition after
preheating 400oC. For all welding groove type, material with 500oC and 600oC preheating after 14 days
corrosion test is 0,5%-0,69% lost. Material with 400oC preheating after 14 days corrosion test is 0,57%-0,76%
lost.
Welding groove also influence corrosion rate. X and V type welding groove give better condition to reduce
corrosion rate than use 1/2V and 1/2 X welding groove. After 14 days corrosion test, the samples with
X welding groove type is 0,5%-0,57% lost. The samples with V welding groove after 14 days corrosion test is
0,51%-0,59% lost. The samples with 1/2V and 1/2X welding groove after 14 days corrosion test is 0,58%-
0,71% lost.
Router 1X3 – RTL Design and VerificationIJERD Editor
Routing is the process of moving a packet of data from source to destination and enables messages
to pass from one computer to another and eventually reach the target machine. A router is a networking device
that forwards data packets between computer networks. It is connected to two or more data lines from different
networks (as opposed to a network switch, which connects data lines from one single network). This paper,
mainly emphasizes upon the study of router device, it‟s top level architecture, and how various sub-modules of
router i.e. Register, FIFO, FSM and Synchronizer are synthesized, and simulated and finally connected to its top
module.
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...IJERD Editor
This paper presents a component within the flexible ac-transmission system (FACTS) family, called
distributed power-flow controller (DPFC). The DPFC is derived from the unified power-flow controller (UPFC)
with an eliminated common dc link. The DPFC has the same control capabilities as the UPFC, which comprise
the adjustment of the line impedance, the transmission angle, and the bus voltage. The active power exchange
between the shunt and series converters, which is through the common dc link in the UPFC, is now through the
transmission lines at the third-harmonic frequency. DPFC multiple small-size single-phase converters which
reduces the cost of equipment, no voltage isolation between phases, increases redundancy and there by
reliability increases. The principle and analysis of the DPFC are presented in this paper and the corresponding
simulation results that are carried out on a scaled prototype are also shown.
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRIJERD Editor
Power quality has been an issue that is becoming increasingly pivotal in industrial electricity
consumers point of view in recent times. Modern industries employ Sensitive power electronic equipments,
control devices and non-linear loads as part of automated processes to increase energy efficiency and
productivity. Voltage disturbances are the most common power quality problem due to this the use of a large
numbers of sophisticated and sensitive electronic equipment in industrial systems is increased. This paper
discusses the design and simulation of dynamic voltage restorer for improvement of power quality and
reduce the harmonics distortion of sensitive loads. Power quality problem is occurring at non-standard
voltage, current and frequency. Electronic devices are very sensitive loads. In power system voltage sag,
swell, flicker and harmonics are some of the problem to the sensitive load. The compensation capability
of a DVR depends primarily on the maximum voltage injection ability and the amount of stored
energy available within the restorer. This device is connected in series with the distribution feeder at
medium voltage. A fuzzy logic control is used to produce the gate pulses for control circuit of DVR and the
circuit is simulated by using MATLAB/SIMULINK software.
Study on the Fused Deposition Modelling In Additive ManufacturingIJERD Editor
Additive manufacturing process, also popularly known as 3-D printing, is a process where a product
is created in a succession of layers. It is based on a novel materials incremental manufacturing philosophy.
Unlike conventional manufacturing processes where material is removed from a given work price to derive the
final shape of a product, 3-D printing develops the product from scratch thus obviating the necessity to cut away
materials. This prevents wastage of raw materials. Commonly used raw materials for the process are ABS
plastic, PLA and nylon. Recently the use of gold, bronze and wood has also been implemented. The complexity
factor of this process is 0% as in any object of any shape and size can be manufactured.
Spyware triggering system by particular string valueIJERD Editor
This computer programme can be used for good and bad purpose in hacking or in any general
purpose. We can say it is next step for hacking techniques such as keylogger and spyware. Once in this system if
user or hacker store particular string as a input after that software continually compare typing activity of user
with that stored string and if it is match then launch spyware programme.
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...IJERD Editor
This paper presents a blind steganalysis technique to effectively attack the JPEG steganographic
schemes i.e. Jsteg, F5, Outguess and DWT Based. The proposed method exploits the correlations between
block-DCTcoefficients from intra-block and inter-block relation and the statistical moments of characteristic
functions of the test image is selected as features. The features are extracted from the BDCT JPEG 2-array.
Support Vector Machine with cross-validation is implemented for the classification.The proposed scheme gives
improved outcome in attacking.
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeIJERD Editor
- Data over the cloud is transferred or transmitted between servers and users. Privacy of that
data is very important as it belongs to personal information. If data get hacked by the hacker, can be
used to defame a person’s social data. Sometimes delay are held during data transmission. i.e. Mobile
communication, bandwidth is low. Hence compression algorithms are proposed for fast and efficient
transmission, encryption is used for security purposes and blurring is used by providing additional
layers of security. These algorithms are hybridized for having a robust and efficient security and
transmission over cloud storage system.
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...IJERD Editor
A thorough review of existing literature indicates that the Buckley-Leverett equation only analyzes
waterflood practices directly without any adjustments on real reservoir scenarios. By doing so, quite a number
of errors are introduced into these analyses. Also, for most waterflood scenarios, a radial investigation is more
appropriate than a simplified linear system. This study investigates the adoption of the Buckley-Leverett
equation to estimate the radius invasion of the displacing fluid during waterflooding. The model is also adopted
for a Microbial flood and a comparative analysis is conducted for both waterflooding and microbial flooding.
Results shown from the analysis doesn’t only records a success in determining the radial distance of the leading
edge of water during the flooding process, but also gives a clearer understanding of the applicability of
microbes to enhance oil production through in-situ production of bio-products like bio surfactans, biogenic
gases, bio acids etc.
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraIJERD Editor
- Gesture gaming is a method by which users having a laptop/pc/x-box play games using natural or
bodily gestures. This paper presents a way of playing free flash games on the internet using an ordinary webcam
with the help of open source technologies. Emphasis in human activity recognition is given on the pose
estimation and the consistency in the pose of the player. These are estimated with the help of an ordinary web
camera having different resolutions from VGA to 20mps. Our work involved giving a 10 second documentary to
the user on how to play a particular game using gestures and what are the various kinds of gestures that can be
performed in front of the system. The initial inputs of the RGB values for the gesture component is obtained by
instructing the user to place his component in a red box in about 10 seconds after the short documentary before
the game is finished. Later the system opens the concerned game on the internet on popular flash game sites like
miniclip, games arcade, GameStop etc and loads the game clicking at various places and brings the state to a
place where the user is to perform only gestures to start playing the game. At any point of time the user can call
off the game by hitting the esc key and the program will release all of the controls and return to the desktop. It
was noted that the results obtained using an ordinary webcam matched that of the Kinect and the users could
relive the gaming experience of the free flash games on the net. Therefore effective in game advertising could
also be achieved thus resulting in a disruptive growth to the advertising firms.
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...IJERD Editor
-LLC resonant frequency converter is basically a combo of series as well as parallel resonant ckt. For
LCC resonant converter it is associated with a disadvantage that, though it has two resonant frequencies, the
lower resonant frequency is in ZCS region[5]. For this application, we are not able to design the converter
working at this resonant frequency. LLC resonant converter existed for a very long time but because of
unknown characteristic of this converter it was used as a series resonant converter with basically a passive
(resistive) load. . Here, it was designed to operate in switching frequency higher than resonant frequency of the
series resonant tank of Lr and Cr converter acts very similar to Series Resonant Converter. The benefit of LLC
resonant converter is narrow switching frequency range with light load[6] . Basically, the control ckt plays a
very imp. role and hence 555 Timer used here provides a perfect square wave as the control ckt provides no
slew rate which makes the square wave really strong and impenetrable. The dead band circuit provides the
exclusive dead band in micro seconds so as to avoid the simultaneous firing of two pairs of IGBT’s where one
pair switches off and the other on for a slightest period of time. Hence, the isolator ckt here is associated with
each and every ckt used because it acts as a driver and an isolation to each of the IGBT is provided with one
exclusive transformer supply[3]. The IGBT’s are fired using the appropriate signal using the previous boards
and hence at last a high frequency rectifier ckt with a filtering capacitor is used to get an exact dc
waveform .The basic goal of this particular analysis is to observe the wave forms and characteristics of
converters with differently positioned passive elements in the form of tank circuits.
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...IJERD Editor
LLC resonant frequency converter is basically a combo of series as well as parallel resonant ckt. For
LCC resonant converter it is associated with a disadvantage that, though it has two resonant frequencies, the
lower resonant frequency is in ZCS region [5]. For this application, we are not able to design the converter
working at this resonant frequency. LLC resonant converter existed for a very long time but because of
unknown characteristic of this converter it was used as a series resonant converter with basically a passive
(resistive) load. . Here, it was designed to operate in switching frequency higher than resonant frequency of the
series resonant tank of Lr and Cr converter acts very similar to Series Resonant Converter. The benefit of LLC
resonant converter is narrow switching frequency range with light load[6] . Basically, the control ckt plays a
very imp. role and hence 555 Timer used here provides a perfect square wave as the control ckt provides no
slew rate which makes the square wave really strong and impenetrable. The dead band circuit provides the
exclusive dead band in micro seconds so as to avoid the simultaneous firing of two pairs of IGBT’s where one
pair switches off and the other on for a slightest period of time. Hence, the isolator ckt here is associated with
each and every ckt used because it acts as a driver and an isolation to each of the IGBT is provided with one
exclusive transformer supply[3]. The IGBT’s are fired using the appropriate signal using the previous boards
and hence at last a high frequency rectifier ckt with a filtering capacitor is used to get an exact dc
waveform .The basic goal of this particular analysis is to observe the wave forms and characteristics of
converters with differently positioned passive elements in the form of tank circuits. The supported simulation
is done through PSIM 6.0 software tool
Amateurs Radio operator, also known as HAM communicates with other HAMs through Radio
waves. Wireless communication in which Moon is used as natural satellite is called Moon-bounce or EME
(Earth -Moon-Earth) technique. Long distance communication (DXing) using Very High Frequency (VHF)
operated amateur HAM radio was difficult. Even with the modest setup having good transceiver, power
amplifier and high gain antenna with high directivity, VHF DXing is possible. Generally 2X11 YAGI antenna
along with rotor to set horizontal and vertical angle is used. Moon tracking software gives exact location,
visibility of Moon at both the stations and other vital data to acquire real time position of moon.
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...IJERD Editor
Simple Sequence Repeats (SSR), also known as Microsatellites, have been extensively used as
molecular markers due to their abundance and high degree of polymorphism. The nucleotide sequences of
polymorphic forms of the same gene should be 99.9% identical. So, Microsatellites extraction from the Gene is
crucial. However, Microsatellites repeat count is compared, if they differ largely, he has some disorder. The Y
chromosome likely contains 50 to 60 genes that provide instructions for making proteins. Because only males
have the Y chromosome, the genes on this chromosome tend to be involved in male sex determination and
development. Several Microsatellite Extractors exist and they fail to extract microsatellites on large data sets of
giga bytes and tera bytes in size. The proposed tool “MS-Extractor: An Innovative Approach to extract
Microsatellites on „Y‟ Chromosome” can extract both Perfect as well as Imperfect Microsatellites from large
data sets of human genome „Y‟. The proposed system uses string matching with sliding window approach to
locate Microsatellites and extracts them.
Importance of Measurements in Smart GridIJERD Editor
- The need to get reliable supply, independence from fossil fuels, and capability to provide clean
energy at a fixed and lower cost, the existing power grid structure is transforming into Smart Grid. The
development of a smart energy distribution grid is a current goal of many nations. A Smart Grid should have
new capabilities such as self-healing, high reliability, energy management, and real-time pricing. This new era
of smart future grid will lead to major changes in existing technologies at generation, transmission and
distribution levels. The incorporation of renewable energy resources and distribution generators in the existing
grid will increase the complexity, optimization problems and instability of the system. This will lead to a
paradigm shift in the instrumentation and control requirements for Smart Grids for high quality, stable and
reliable electricity supply of power. The monitoring of the grid system state and stability relies on the
availability of reliable measurement of data. In this paper the measurement areas that highlight new
measurement challenges, development of the Smart Meters and the critical parameters of electric energy to be
monitored for improving the reliability of power systems has been discussed.
Study of Macro level Properties of SCC using GGBS and Lime stone powderIJERD Editor
One of the major environmental concerns is the disposal of the waste materials and utilization of
industrial by products. Lime stone quarries will produce millions of tons waste dust powder every year. Having
considerable high degree of fineness in comparision to cement this material may be utilized as a partial
replacement to cement. For this purpose an experiment is conducted to investigate the possibility of using lime
stone powder in the production of SCC with combined use GGBS and how it affects the fresh and mechanical
properties of SCC. First SCC is made by replacing cement with GGBS in percentages like 10, 20, 30, 40, 50 and
by taking the optimum mix with GGBS lime stone powder is blended to mix in percentages like 5, 10, 15, 20 as
a partial replacement to cement. Test results shows that the SCC mix with combination of 30% GGBS and 15%
limestone powder gives maximum compressive strength and fresh properties are also in the limits prescribed by
the EFNARC.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Elizabeth Buie - Older adults: Are we really designing for our future selves?
International Journal of Engineering Research and Development
1. International Journal of Engineering Research and Development
e-ISSN: 2278-067X, p-ISSN: 2278-800X, www.ijerd.com
Volume 10, Issue 5 (May 2014), PP.32-40
32
A Novel Approach for Improving the Recommendation System
by Knowledge of Semantic Web in Web Usage Mining
1
Nirali N. Madhak, 2
Chintan R. Varnagar, 3
Shahida G. Chauhan
1
Post Graduate Student, 2, 3
Assistant Professor,
1, 2, 3
Department of Computer Engineering,
Atmiya Institute of Technology and Science, Rajkot, Gujarat, India
Abstract:- The World Wide, We has influenced a lot to both users as well as the web site owners. Massive
growth of World Wide Web increases the complexity for users to browse effectively. For increasing web site
uses and to achieve desired goal efficiently, web server activities are hypothetical to be changed as per users’
interests. To achieve this they have to record and analyze user access pattern which are captured in the form of
log files. Web usage mining refers to process of analyzing interaction of user with different web application and
deriving some important knowledge out of it. Web usage mining process produces result based on maximum
usage history, stored in the web server access logs. So we propose the system, which uses semantic knowledge,
derived from each page, along with knowledge derived from WUM. The system generates the most efficient and
applicable recommendation list, and also provides equal opportunity for the content, which is added recently, to
be incorporated in the list, if this best matches with users interest.
Keywords:- Data mining, Web usage mining (WUM), Web log mining, Content Recommendation,
Recommender System, Web server log.
I. INTRODUCTION
The richness of information on the World Wide Web has attracted users to seek and retrieve
information from the World Wide Web (WWW). When user is trying to access the content on web site, he/she is
facing difficulty in finding fruitful content, that utmost match with the user interest. Recommendation systems
are intelligent system that suggests and assist in selecting right content heuristically. Recommendation system is
one of the applications of Web Usage Mining (WUM).
The main goal of the recommendation system in commercial domain is to improve website usability and
thereby increasing the profitability of website owner. The patterns/ knowledge discovered provided as an input
to the recommendation system, which recommends appropriate pages relevant to user interest.
Web access log file that resides in the web server, notes the client activity – request for accessing file
on server initiated by client through the web browser. Web usage mining refers to the process of extracting
knowledge/patterns by applying various data mining techniques like association rule mining (ARM), clustering,
classification etc on the web access log files.
Recommendation systems, that makes suggestion based on only previous web access history or one
that considers the similarity between item currently being accessed and items similar to that, fights with its own
perils, and hence the recommendation generated does not guaranteed to be qualitative. The content of page i.e.
overall theme, keywords, and its density should be considered in deciding the candidature of the content, to be
presented in the recommendation list.
Here we propose content recommendation system which gives, suggestion based on not only access
history of current and other users, but also considers the semantic knowledge achieved, from web content
mining. The architecture presented, proved, to be giving better and optimized results. Additionally the system is
prone with the addition of newly added page or unvisited pages which matches with user interest.
So that proposed solution is combination of two web mining approaches that are Web Usage Mining
and Web Content Mining (WCM), Section IV.
Section II explains the brief about work done so far. Section III explains the input for web usage
mining and on various Pattern discovery techniques that can be applied on preprocessed web access logs
gathered to mine knowledge from it. Section IV discusses on proposed content matching algorithm which uses
the heuristic function to build the recommendation list, Section V concludes with the merits of proposed
approach and future scope.
2. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
33
II. LITERATURE SURVEY
Data mining is the process of extracting previously unknown information from different types of data
like text, audio, video etc., which leads to fruitful knowledge. Now a day’s web has proved to be as affluent
sources of data where multiple domains are accessed and mined, mining web data is referred as Web Mining.
Web Mining can be broadly divided into three phases: [7] Web Usage Mining (WUM), Web Content
Mining (WCM) and Web Structure Mining (WSM). In this paper, we propose a recommendation system, which
not only considers user navigation history but it also considers content of web page. So, here proposed approach
is a combination of WUM and WCM, which in combination can give better recommendations.
Fig. 1: Classification of WCM
As shown in figure 2.1, WCM aims to extract information from different type of data of web page. By
applying various techniques of web content mining on unstructured data ,a HTML pages, is challenging task
because, HTML web pages have multiple tags which make the web pages highly unstructured. Topic tracking is
the technique by which a registered user can track the topic according to his/her interest. He/she have to register
with the topic, whenever there is an update or news regarding the interest of the user happens, he/she will
intimate by message.
Structured data extraction is the process of extracting information from the web pages; programs are
written which helps to extract such information, which is surrounded by Wrapper. By the keyword extraction
technique which is applied on the structured data gives the information about the number of keywords which is
of one phrases, two phrases etc.
Multimedia data extraction is the process of finding interesting knowledge from data like audio, video,
image and text. By developing methods and tools to organize, supervise, search and to perform domain specific
tasks for data of different domains such as surveillance, meetings, broadcast news, sports, archives, movies,
medical data, personal and online media collections
There are various approaches used for web content mining like, by using the ontology’s which specify
the domain specific knowledge and finding its relevant content, by applying the various clustering techniques.
Ontology is a domain specific knowledge that could be used to describe information on the Web. Ontologies
based web mining can be used to improve search to web data by adding Ontology explanation, better browsing
capabilities and personalization of Web data from the user’s interest profile. With the use of various domain
specific knowledge (ontology) ,we can improve the user interest profile integrating with the semantic web
approach.
Semantic web is all about integrating and extracting intelligent information from structured or
unstructured data from the web and the knowledge which is embedded in web applications, thus providing a
semantic-based access to the Internet. Ontology is one of the layers of Semantic Web’s architecture as proposed
by Sir Tim Berner’s Lee.
Accordingly [], Jayatilaka A.D.S!, and Wimalarathne, It provide idea how ontologies are used in
semantic web. Semantic web mining is relatively new sub-field of data mining. It has a vast scope for
investigation keeping in view of the availability of tons of unstructured data on WWW.A user-oriented semantic
web search is the need of today and days to come. This field if explored in a right manner will provide unlimited
opportunities to extract knowledge to ultimately improve the profitability of an individual or company by
mining knowledge from unstructured and/or structured data available across the internet.
Jayntilaka et al. [14], insisted the concept of semantic web using both web author's view point and web
user's perspectives in the ontology learning process. It also eliminates the use of web site dependent
characteristics in extracting semantics.
The extracted concepts with help of ontology web language and the conceptual relationships gives rise
to a semantic network and form ontology. There are three main stages in this method which are:
Web Content Mining
*()
Unstructured Structured
Multi-Media
-Information
Extraction
-Topic Tracking
-Summarization
-Wrapper Generation
-Page Content Mining
-Keyword Extraction
-Multi-Media
like audio-video
etc Mining
3. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
34
i) Concept and conceptual relationship extraction through web content mining,
ii) Conceptual relationships identification through web usage mining
iii) Refining/Merging the conceptual relationships obtained through the web content mining process
[15]Web mining based on semantic networks is used the new semantic to improve the Web mining
result. With the analysis of semantic level on RDFMS resource description, RDFMS hierarchical clustering
method based on semantic distance data is proposed. The inductive logic programming design is proposed for
the semantic Web data mining technique and how to apply this technological algorithm description.
Jenice Aroma R.et al.[16], proposed the semantic discovery algorithm combines the method of
semantic similarity measure between words to be applied with the queries supplied, in order to retrieve the
semantically matched results. To optimize the retrieved results, ranking be applied over the matched results. It
brings more relevant results to be ranked highest. Thus, Intelligence on Information retrieval for achieving more
relevant results can be implemented on applying this proposed semantic discovery algorithm with semantic
similarity measure. But for that we have to classify each document into its appropriate ontology classes.
Content Recommendation Systems are one of the applications of WUM as well as WCM. Content
recommender system, gives suggestion based on access history of current user and others.
The main objectives of the system is to improve usability of web site by dynamically and automatically
understanding and modeling visitors navigational behavior to build user profile ,second exploiting thus created
knowledge base by the application of most suited heuristic techniques, often embedded within a special
component called intelligent agent. There by recommending appropriate pages and gaining user satisfaction and
easy of surfing [9]. Identifying and selecting proper recommendation to user is complex process and mostly the
techniques applied are heuristic in nature.
A Recommender System, merely applying various data mining techniques to the web server log data
fight with its own perils, as it will not help to derive complete, accurate and efficient recommendations. First
Information in the web log is very limited, Second it assumes that the requests are fulfilled sequentially, which
is not true for con-current information need, it means user can fire the requests parallel. Finally if user leaves
after visiting few links, without competing transactions, it might mislead our results. But visitor may have left
because of non fulfillment of his information need, poor link navigation or bad navigation also [17].
Ting CHEN et al. [18] suggested a system, which does recommendation, consisting of three tiers
(layer). L-1(Layer-1) is row information collection agent, which collects data from client machine. L-2, a logic
layer uses this data to create Dynamic User Profile (DUP), L-3 is responsible for presentation and customized
UI. [18] Suggested to build such a dynamic profile from various hardware level events like keyboard, mouse
etc.
According to [19] to discover sequential access patterns first association rule mining algorithm (A-
priori) and its associated modified algorithm, second methods based on stochastic probability, Markov chains,
third to use weighted association rule mining, which allows different weights to be assigned to different items,
hence improving AR model.
[4] Suggested similar concept, which creates user groups (clusters) and uses intelligent
recommendation agent, covering all type of multimedia contents. This is perfect combination for the integration
of user specific activity (service) recommendation in social networking and its associated web services, in the
environment of web 2.0.By the above literature survey we will finding some of techniques for web usage mining
and web content mining. It is possible to optimize the recommendation system by discovery of new algorithm
and new approach to measure the result of combination of both technique WUM and WCM as discussed in
Section IV.
III. INPUT FOR WEB USAGE MINING
In the last decade, many researchers have developed many different kinds of approaches of web
system to achieve web personalization. While surfing the web sites, users interaction with web sites are recorded
in web server access log file.
There are three main sources to get the row log data [1], which are namely:
1) Web server log file data like Access Log, Agent Log, Error Log, Referrer Log.
2) Client side log files, which are more authenticate and accurate.
3) Proxy server or firewall log files, which contains access log captured at organizational gateway, may
be varied in format, content and in other aspect from server to server or across different software and hence very
difficult to get useful information from it.
4. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
35
Web Server access log data:
The most frequently used source for web usage mining is web server access log data. This web log data
is generated automatically within web server when it services any request sent by the user, which contains all
information about visitor’s activity.
Accordingly Suneetha K R et al.[1] proposed various preprocessing techniques apply on web access log.
Many different formats for web access log data are available like:-
1. Common log format
2. Extended common log format,
3. Centralized log format
4. NCSA common log format
5. ODBC logging
Among all common or extended file format is mainly implemented by web server due to many reasons.
Fig. 2: Attributes of ECFL [7].
Extended Common Log File Format (ECFL), Figure 2.2 is important in web usage mining, as it can be
customized as per the requirements and is followed by most of the web server. The additional attributes that are
captured are
i) REFERER_URL define the URL where visitor came from
ii) HTTP_Command reflects GET/POST method
iii) PROTO--type of protocol used for the request
iv) HTTP_Status-status code generated by the request
v) GMToffset-signed offset from Greenwich Mean Time
[1] Web access logs may be used to increase the effectiveness of web portals or for better
understanding of user behavior. The main research challenges of this field are identifying the non human entries
made by web robots, design efficient heuristics for user session identification and finding association among
different user’s access patterns.
The system takes raw log data for the website http://eyuva.com for the periods between 2014-03-20
[05:35:14] TO 2014-04-23 [4:41:59].As shown below:
Client side log data:
It is refer to recording of activities, events that happens within the premises of client machine. Like mouse
wheel rotation, scrolling within a particular page, mouse clicks, content selection. In some case it is
advantageous, as it eliminates necessity of session identification, caching [11]. This can be recorded by number
of ways:
1) By integrating java applet with web site: Java applet records each of the activity of users. But for that
java plug in need to be installed on each client side browser. Also user may experience delay in page loading
time, when applet is loaded for the first time [11]. Additionally all current pages of the website need to be
redesigned and recreated in terms of Applet.
2) By writing Java Scripts: Java Script needs to be inserted at appropriate places, need to be invoked as
and when required (need to be associated with appropriate event handler), in each page of web site. This will
record these interactions of user with web page and report it to server when transaction is complete.
3) By developing a browser plug-in(extension): Which need to be installed only once which can record this
kind of interaction and will send the record at finite interval of time or just before when user is about to close
the connection with website or when user is quitting from browser. This can be done without changing the
underlying design, architecture or technology of web site.
Proxy Server Log Data:
At many places network traffic is routed through a dedicated machine known as a proxy server, all the
request and response are serviced through this proxy server. Study of this proxy server log files, whose format is
same as of web log file may reveal the actual HTTP requests coming from multiple clients to multiple web
5. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
36
servers and characterizes, reveals the browsing behavior for a group of anonymous users sharing a common
proxy server [11].
Some web sites use n-tier architecture to have reliable, efficient and secure web applications. Log data
that are gathered at application server while servicing the users request can also be used for web usage mining.
They peculiarly show how user requests are serviced and may assist in identifying and understanding the
internal calls-page access resulted to fulfill a single request.
IV. PROPOSED CONTENT RECOMMENDATION SYSTEM
In the literature survey till now I have not came across the method, technique, or approach which uses
the content of the page ,as a one of the parameter to decide whether it should be made available – get listed in
the recommendation list or not. So by considering this parameter in building recommendation list will not only
improve the quality of recommendations made by system for user – depending on his till now browsing patterns
(interest) but will provide a equal chance for the pages (content), which are added recently in the system.
At no place up to authors knowledge it is think of, no argument is given in support or favor for using
both ,i.e. web server access log and content of each page to build the recommendation set. Hence the system that
knows the contents of the pages that user has browsed till time, will surely be able to predict the most interesting
page – content, that user is about to like or will be interested in or he is looking for.
So content can be more precisely be predicated by modifying heuristic algorithm by considering this
semantic knowledge – that reside within a page , when combined with other traditional methods of content
recommendation like frequent access pattern mining using Apriori algorithm or some other techniques to mine a
association rule and combine that all results according to heuristic function.
Proposed system architecture for content recommendation system can be logically divided in to TWO
co-related PHASES named. Each phase is logically co-related with each other in a sense it accepts output of
other as an input. But from the viewpoint of their functions and their relative execution domain it can be
classified in above two phases.
1) Back End (Offline) and
2) Front End (Online).
Phase-I (Back End):
This Module is responsible for capturing and storing the content of the pages, in a meaningful and
easily accessible – retrievable way known as web content mining. I.e. getting to know the semantic of each page
and to store it in a way, that is most easy to access as and when it is required.
It also processes web server access logs to mine frequent access patterns, which in itself is not an easy
task and requires a series of steps known as data preprocessing, pattern discovery and pattern analysis
Back end phase processes the data which are relatively static in nature like web pages – content of the
web site and web server access logs. They are static in a way as it does not need to compute or have to perform
any processing-each time user request is fired or recommendation is to be generated. However it need to be
updated periodically, so as to get qualitative and most accurate, up to date results. So frequent run of this phase
is necessary especially when new content is added in to the web site and also after finite time even though it is
not added or updated so as to accommodate and mimic the correct current trend of the website user.
Back End Phase further can be classified based on the function and their operation domain, in
following Modules.
A) Web usage mining on web server log
B) Web content mining
6. International Journal of Engineering Research and Development
e-ISSN: 2278-067X, p-ISSN: 2278-800X, www.ijerd.com
Volume 10, Issue 5 (May 2014), PP.32-40
37
Fig. 3: Proposed Architecture of Recommendation System
A) Web usage mining on web server log:
This module uses web log files as input and derives some meaningful knowledge, frequent access
patterns out of it. Entire process of web usage mining can be logically divided into four significant and co-
related steps, which are Data Collection, Data Preprocessing, Discovery of Pattern, and Analysis of pattern. For
that first web raw log data need to be cleaned, it means removal of unwanted rows, known as data
preprocessing. Secondly to mine frequent access patterns, using technique of Association Rule Mining,
specifically a simple Apriori algorithm is applied and results (patterns) are stored in some permanent storage for
referring to it later.
Data Preprocessing: Data Preparation is the most complicated and time consuming task. About 80 percentages
[2] of time is given on this process to strengthen quality of data because as qualitative the data is better the
results. For this data preparation task which mainly includes various sub-task namely data cleaning, user
identification, session identification, path completion and transaction identification [6].
Web Server
Web Server
Access Log Data
Preprocessing
Pattern
Discovery
Pattern
Analysis
Web Pages
Extracting Frequent
Keywords –per Page
Generate two and three
phrases from the
previous step
Capture User
Interaction
Build & Track
Active Session –
per User
Store and
update in
database
Store it in database
for further
references
Generate and build
recommendation list by
considering appropriate
weightage to each parameter
Browser
USER
Retrieve partial session
for current userDB – WUM
(Frequent
Patterns)
Result stored
in Database
DB - Patterns
Recommendation List
DB – WCM
(Frequent
Keywords &
Phrases)
Back End Front End
7. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
38
Discovery of pattern: It is the ultimate stage where some useful knowledge will be derived by applying various
statistical and/or data mining techniques at hand from various research areas like data mining, machine learning,
statistical method and pattern recognition. Frequently used techniques are classification, clustering, association
rule, sequential pattern etc [12].
Association Rules are able to discover related item occurring together in same transaction, and is used to find
interdependency, co-relation among the pages. Such number of rules generated could be very large so two
measures support and confidence is employed, which determines importance and quality of rules [1]. A-Priori
and its many versions are developed to mine association rule.
B) Web Content Mining: This step is responsible for generating keywords – which are appearing most of the
times from the contents of each web pages of a particular web site, and then it generates two word phrases and
three word phrases by applying following proposed User_Interest_Content_Matching algorithm, which helps to
optimize the results of recommendation list.
Algorithm: User_Interest_Content_Matching
By applying the above algorithm that helps to find heuristic function, that will help to combine two
different approaches web content mining and web usage mining.
Phase-II (Front End):
It is responsible for processing of a URL request of user by Personalized & Interest Specific Intelligent
Recommendation Agent. These results (Recommendation Set) can be made available to the user, which is based
on the current browsing history of current user that too without changing neither the page code (to insert the
recommendation results) nor browsing experience of a user is affected for the website under consideration. This
data will be supplied in the side bar, also known as a slider, without restructuring the web page. To build the
final recommendation list algorithm Generate_ recommendation _list is applied.
Algorithm: Generate_ Recommendation List
Step-1: For each webpage extract top N keywords, whose frequency is maximum.
Step-2: Generate two words phrases and three words phrases from the keywords generated at step-1.
Step-3: Store these result in database for further reference.
Step-4: Apply the member function and compute the value for each web page from the set of
recommendation list (web pages) as follows:
Fwi= A * no. of occurrences of single phrase keywords that is common across all the visited page till now
+
B * no. of occurrences of two words phrase keywords that is common across all the visited page till now
+
C * no. of occurrences of three words phrase keywords that is common across all the visited page till now
+
X * support count of (% times) the page is referred in history
Where A, B, C Є [0, 1] & X Є [1, 100]
Step-5: Choose the best promising page (content) based on the combination of usage history obtained and
appropriateness of content as follow:
Compare the values computed by above heuristic function, computed for each page and choose the
maximum value – which indicates most likeliness of the value to be presented i.e. which user might like the
most.
Step-6: Display Recommendation set based on descending value of heuristic function.
Step-7: Exit
Step – 1: Retrieve IP address of client requesting for the resource.
Step – 2: Capture and build the active session i.e. list of web pages traversed so far, for all the currently
active user dynamically, by applying session identification technique.
Step – 3: The new current, partial session will be compared with already existing aggregate usage profile
from the frequent access pattern set stored in knowledge base.
Step – 4: Build a recommendation set based on ascending value returned by Heuristic function Fwi of
procedure User_Interest_Content_Matching.
Step – 5: If more than one pattern found with same support value for current partial session, second
measures like page weight, computed by considering various parameters, stored in database will be used to
decide the priority for displaying the results.
8. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
39
If we are going to consider only the web usage mining behavior of user ,how many pages are frequently access
by multiple user we are getting result by considering active time spent by user on the web page as below:
Fig.4: URL vs Accesstime
By applying the for generating recommendation list according the ascending value returned by heuristic
function. Following figure shows the how to index url with the value of Fwi.
Fig .5: URL vs Fwi
http://eyuva.com/2012/10/karbonn-…
http://eyuva.com/2012/03/srs-on-…
http://eyuva.com/2013/12/how-to-…
http://eyuva.com/2013/12/first-…
http://eyuva.com/2011/03/oreo-…
http://eyuva.com/
http://eyuva.com/2011/05/nokia-n9-…
http://eyuva.com/2011/07/len-less-…
http://eyuva.com/2011/08/reliance-…
http://eyuva.com/2010/10/finally-…
http://eyuva.com/2011/09/how-to-…
http://eyuva.com/2013/06/playstation…
http://eyuva.com/2010/02/koochie-…
http://eyuva.com/category/gadget/pa…
WUM
0 100 200 300 400 500
http://eyuva.com/category/gadget…
http://eyuva.com/category/gadget…
http://eyuva.com/2013/12/first-…
http://eyuva.com/2013/06/how-…
http://eyuva.com/2012/12/how-…
http://eyuva.com/2012/03/srs-on-…
http://eyuva.com/2011/10/sony-…
http://eyuva.com/2011/09/how-…
http://eyuva.com/2011/07/playsta…
http://eyuva.com/2011/06/nokia-…
http://eyuva.com/2011/05/nokia-…
http://eyuva.com/2011/03/oreo-…
http://eyuva.com/2010/10/finally-…
http://eyuva.com/
Url vs Fwi
fwi
9. A Novel Approach for Improving the Recommendation System by Knowledge of Semantic…
40
By comparing results of the only web usage behavior and combination of WCM and WUM, Second approach
gives us fruitful results that are beneficial to finding user interest web pages that best matches with user interest
profile.
V. CONCLUSION
Web sites are great amount of use for the user. Web sites are built, deployed and maintained to serve
with various function to user. Web is increasing its importance in each possible aspect and is becoming an
expected part of one’s routine and regular resources. Hence there are sufficient opportunities and wide scope
and requirement to study this field in the depth. Systems incorporating knowledge from ONLY navigational
history (WUM) often produce incomplete, inefficient result – as it is based on single parameter .So newly
added content(web page),will not be listed in Recommendation List , although it best matches with users interest
– just because so far it is not visited or visited very less time. So a system can be improved, if it considers
semantic knowledge of each page and incorporates this factor, with knowledge achieved from WUM,
dynamically – for each possible element (page) from recommendation set.
REFERENCES
[1] Hauqiang zhou and Hongxia Gao et al. “Research on improving Methods of processing in Web log
Mining”, IEEE, 2010.
[2] R. Cooley, B. Mobasher, J. Srivastava, "Web mining: information and pattern discovery on World
Wide web”, tools with artificial intelligence, Ninth IEEE International November 1997.
[3] J. Srivasta, R.Cooley, M.Deshpande, P.Tan, "Web usage mining: discovery and applications of usage
patterns from Web data", ACM SIGKDD Vol.7, No.2, Jan-2000.
[4] S. K. Pani et al.,” Web Usage Mining: A survey on pattern extraction from web logs”, International
Journal of Instrumentation, Control &Automation, Vol.1, Issue 1, 2011.
[5] Ting Chen et al., “Content Recommendation System based on Private Dynamic User Profile”, VIth
International Conference on Machine Learning and Cybernetics, IEEE, August-2007.
[6] J. Srivasta, R.Cooley, M.Deshpande, P.Tan, "Web usage mining: discovery and applications of usage
patterns from Web data", ACM SIGKDD Vol.7, No.2, Jan-2000.
[7] Chintan.R.Varnagar et al.,”Web Usage Mining:Asurvey on Pattern extraction using web
logs”,IEEE,2013
[8] Sang-il hwa-sung kim et al.,”Ontology Modeling for provision of semantic based open API”, IEEE,
2013
[9] Rana Forsati et.al, ”An Efficient Algorithm for Web Recommendation Systems”, IEEE, 2009.
[10] Ravi Bhushan and Dr.Rajendra Nath,et al., “Automatic Recommedation of web pages for online users
using Web Usage Mining”, IEEE, 2012
[11] Saim Shin et al,” The User-group based Recommendation for the Diverse Multimedia Contents in the
Social Network Environments” , IEEE, 2011.
[12 Jenice Arona R,Mathew kurian,”A semantic Web:Intelligence in information Retrieval”,IEEE-2013.
[13] Yanjing Zou,”Personalized Automatic Recommedation for the web based Autonomous language
learning system based on Data Mining Technology”,IEEE-2011.
[14] Jayatilaka A,D.S, ”Knowledge Extraction for semantic web using Web Mining”, IEEE, 2013.
[15] Zahid Ansari,A.Vinay Babu, “A Fuzzy set theoretic approach to Discover user Session from Web
Navigational Data ”, International Conference on Control and Automation, IEEE, 2011.
[16] Liu Kewen, “Analysis of Preprocessing methods for web usage mining”, International Conference on
measurement, Information and Control, IEEE, 2012