This document discusses a hierarchical fuzzy rule-based classification system using genetic rule selection to filter unwanted messages from online social networks. It aims to improve performance on imbalanced data sets by increasing granularity of fuzzy partitions at class boundaries. The system uses a neural network learning model and genetic algorithm for rule selection to build an accurate and compact fuzzy rule-based model. It analyzes challenges in classifying short texts from social media posts and reviews related work on content-based filtering and policy-based personalization for social networks. The document also discusses issues with imbalanced data sets and proposes oversampling the minority class using SMOTE (Synthetic Minority Over-sampling Technique) as a preprocessing step to address class imbalance problems.
Iaetsd efficient filteration of unwanted messagesIaetsd Iaetsd
This document discusses an efficient filteration system for unwanted messages on social networking sites. It proposes a Trust Evaluation System (TES) that uses a reputation metric to evaluate new messages submitted by users and assign a confidence level based on the trustworthiness of the reporter. TES rewards reporters whose feedback agrees with highly trusted users and penalizes those who disagree. It also continuously updates the confidence level of messages based on additional feedback. The system aims to induct a community of trusted reporters and automatically filter future messages matching fingerprints that have been cataloged as spam.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Pharmaceutical Science Invention (IJPSI) is an international journal intended for professionals and researchers in all fields of Pahrmaceutical Science. IJPSI publishes research articles and reviews within the whole field Pharmacy and Pharmaceutical Science, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
A system to filter unwanted messages from OSN user wallsGajanand Sharma
The document presents a system to filter unwanted messages from user walls on online social networks. It uses machine learning techniques like text classification and radial basis function networks to categorize messages as neutral or non-neutral, and further classify non-neutral messages. Users can define custom filtering rules and blacklists to automatically filter messages on their walls based on content, user relationships, and other criteria. The system aims to give users more control over their timeline posts while maintaining flexibility.
Filtering Unwanted Messages from Online Social Networks (OSN) using Rule Base...IOSR Journals
Online Social Networks (OSNs) are today one of the most popular interactive medium to share,
communicate, and distribute a significant amount of human life information. In OSNs, information filtering can
also be used for a different, more responsive, function. This is owing to the fact that in OSNs there is the
possibility of posting or commenting other posts on particular public/private regions, called in general walls.
Information filtering can therefore be used to give users the ability to automatically control the messages
written on their own walls, by filtering out unwanted messages. OSNs provide very little support to prevent
unwanted messages on user walls. For instance, Facebook permits users to state who is allowed to insert
messages in their walls (i.e., friends, defined groups of friends or friends of friends). Though, no content-based
partialities are preserved and therefore it is not possible to prevent undesired communications, for instance
political or offensive ones, no matter of the user who posts them. To propose and experimentally evaluate an
automated system, called Filtered Wall (FW), able to filter unwanted messages from OSN user walls
To Get any Project for CSE, IT ECE, EEE Contact Me @ 09849539085, 09966235788 or mail us - ieeefinalsemprojects@gmail.co¬m-Visit Our Website: www.finalyearprojects.org
This document outlines a proposed system to filter unwanted messages from online social networks. It discusses the existing problems of misuse on social media platforms. The proposed system would use machine learning techniques like SVM for text categorization and identification of fake profiles to filter content by category (e.g. abusive, vulgar, sexual). It presents the system architecture as a three-tier structure and provides results of testing the filtering mechanism and classifier. The conclusion is that the "Filtered wall" system could address concerns around unwanted content on social media walls.
The document proposes a system called Filtered Wall (FW) to filter unwanted messages from users' walls in Online Social Networks (OSNs). FW uses machine learning techniques to automatically categorize short text messages. It also provides flexible filtering rules that allow users to customize which content is displayed on their walls based on message categorization, user profiles, and relationships. The system was experimentally evaluated on its ability to accurately categorize messages and effectively apply the filtering rules. A prototype was implemented for Facebook to demonstrate the system.
Iaetsd efficient filteration of unwanted messagesIaetsd Iaetsd
This document discusses an efficient filteration system for unwanted messages on social networking sites. It proposes a Trust Evaluation System (TES) that uses a reputation metric to evaluate new messages submitted by users and assign a confidence level based on the trustworthiness of the reporter. TES rewards reporters whose feedback agrees with highly trusted users and penalizes those who disagree. It also continuously updates the confidence level of messages based on additional feedback. The system aims to induct a community of trusted reporters and automatically filter future messages matching fingerprints that have been cataloged as spam.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Pharmaceutical Science Invention (IJPSI) is an international journal intended for professionals and researchers in all fields of Pahrmaceutical Science. IJPSI publishes research articles and reviews within the whole field Pharmacy and Pharmaceutical Science, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
A system to filter unwanted messages from OSN user wallsGajanand Sharma
The document presents a system to filter unwanted messages from user walls on online social networks. It uses machine learning techniques like text classification and radial basis function networks to categorize messages as neutral or non-neutral, and further classify non-neutral messages. Users can define custom filtering rules and blacklists to automatically filter messages on their walls based on content, user relationships, and other criteria. The system aims to give users more control over their timeline posts while maintaining flexibility.
Filtering Unwanted Messages from Online Social Networks (OSN) using Rule Base...IOSR Journals
Online Social Networks (OSNs) are today one of the most popular interactive medium to share,
communicate, and distribute a significant amount of human life information. In OSNs, information filtering can
also be used for a different, more responsive, function. This is owing to the fact that in OSNs there is the
possibility of posting or commenting other posts on particular public/private regions, called in general walls.
Information filtering can therefore be used to give users the ability to automatically control the messages
written on their own walls, by filtering out unwanted messages. OSNs provide very little support to prevent
unwanted messages on user walls. For instance, Facebook permits users to state who is allowed to insert
messages in their walls (i.e., friends, defined groups of friends or friends of friends). Though, no content-based
partialities are preserved and therefore it is not possible to prevent undesired communications, for instance
political or offensive ones, no matter of the user who posts them. To propose and experimentally evaluate an
automated system, called Filtered Wall (FW), able to filter unwanted messages from OSN user walls
To Get any Project for CSE, IT ECE, EEE Contact Me @ 09849539085, 09966235788 or mail us - ieeefinalsemprojects@gmail.co¬m-Visit Our Website: www.finalyearprojects.org
This document outlines a proposed system to filter unwanted messages from online social networks. It discusses the existing problems of misuse on social media platforms. The proposed system would use machine learning techniques like SVM for text categorization and identification of fake profiles to filter content by category (e.g. abusive, vulgar, sexual). It presents the system architecture as a three-tier structure and provides results of testing the filtering mechanism and classifier. The conclusion is that the "Filtered wall" system could address concerns around unwanted content on social media walls.
The document proposes a system called Filtered Wall (FW) to filter unwanted messages from users' walls in Online Social Networks (OSNs). FW uses machine learning techniques to automatically categorize short text messages. It also provides flexible filtering rules that allow users to customize which content is displayed on their walls based on message categorization, user profiles, and relationships. The system was experimentally evaluated on its ability to accurately categorize messages and effectively apply the filtering rules. A prototype was implemented for Facebook to demonstrate the system.
A system to filter unwanted messages from theMadan Golla
This document presents a system to filter unwanted messages from social network users' walls. It consists of three main components: filtering rules, thresholds for applying the rules which are customized for each user, and a blacklist mechanism. The filtering rules allow users to control what types of messages are allowed on their walls based on attributes of the message creator and their relationship to the user. The system aims to provide flexible and transparent filtering of messages while minimizing mistakes.
1) The document discusses challenges with achieving interoperability between ultra large scale systems due to heterogeneity in platforms, data, and semantics.
2) It proposes a three-layered model for interoperability using web service technologies and semantic web approaches to address these challenges.
3) Key aspects of interoperability discussed include different levels (e.g. syntactic, semantic), use of ontologies to provide common understandings and resolve conflicts, and semantic web service approaches like OWL-S that semantically annotate service descriptions.
Semantic Massage Addressing based on Social Cloud Actor's InterestsCSCJournals
Wireless communication with Mobile Terminals has become popular tools for collecting and sending information and data. With mobile communication comes the Short Message Service (SMS) technology which is an ideal way to stay connected with anyone, anywhere anytime to help maintain business relationships with customers. Sending individual SMS messages to long list of mobile numbers can be very time consuming, and face problems of wireless communications such as variable and asymmetric bandwidth, geographical mobility and high usage costs and face the rigidity of lists. This paper proposes a technique that assures sending the message to semantically specified group of recipients. A recipient group is automatically identified based on personal information (interests, work place, publications, social relationships, etc.) and behavior based on a populated ontology created by integrating the publicly available FOAF (Friend-of-a-Friend) documents. We demonstrate that our simple technique can first, ensure extracting groups effectively according to the descriptive attributes and second send SMS effectively and can help combat unintentional spam and preserve the privacy of mobile numbers and even individual identities. The technique provides fast, effective, and dynamic solution to save time in constructing lists and sending group messages which can be applied both on personal level or in business.
This document summarizes a research paper that proposes and compares fuzzy and Naive Bayes models for detecting obfuscated plagiarism in Marathi language texts. It first provides background on plagiarism detection and describes different types of plagiarism, including obfuscated plagiarism. It then presents the fuzzy semantic similarity model, which uses fuzzy logic rules and semantic relatedness between words to calculate similarity scores between texts. Next, it describes the Naive Bayes model for plagiarism detection using Bayes' theorem. The paper compares the performance of the fuzzy and Naive Bayes models on precision, recall, F-measure and granularity. It finds that the Naive Bayes model provides more accurate detection of obfuscated plagiar
A Review: Text Classification on Social Media DataIOSR Journals
This document provides a review of different classifiers used for text classification on social media data. It discusses how social media data is often unstructured and contains users' opinions and sentiments. Various machine learning algorithms can be used to classify this social media text data, extracting meaningful information. The document focuses on describing Naive Bayes classifiers, which are commonly used for text classification tasks. It explains how Naive Bayes classifiers work by calculating the posterior probability that a document belongs to a certain class, based on applying Bayes' theorem with an independence assumption between features.
Context Driven Technique for Document ClassificationIDES Editor
In this paper we present an innovative hybrid Text
Classification (TC) system that bridges the gap between
statistical and context based techniques. Our algorithm
harnesses contextual information at two stages. First it extracts
a cohesive set of keywords for each category by using lexical
references, implicit context as derived from LSA and wordvicinity
driven semantics. And secondly, each document is
represented by a set of context rich features whose values are
derived by considering both lexical cohesion as well as the extent
of coverage of salient concepts via lexical chaining. After
keywords are extracted, a subset of the input documents is
apportioned as training set. Its members are assigned categories
based on their keyword representation. These labeled
documents are used to train binary SVM classifiers, one for
each category. The remaining documents are supplied to the
trained classifiers in the form of their context-enhanced feature
vectors. Each document is finally ascribed its appropriate
category by an SVM classifier.
Classification-based Retrieval Methods to Enhance Information Discovery on th...IJMIT JOURNAL
The widespread adoption of the World-Wide Web (the Web) has created challenges both for society as a whole and for the technology used to build and maintain the Web. The ongoing struggle of information retrieval systems is to wade through this vast pile of data and satisfy users by presenting them with information that most adequately it’s their needs. On a societal level, the Web is expanding faster than we can comprehend its implications or develop rules for its use. The ubiquitous use of the Web has raised important social concerns in the areas of privacy, censorship, and access to information. On a technical level, the novelty of the Web and the pace of its growth have created challenges not only in the development of new applications that realize the power of the Web, but also in the technology needed to scale applications to accommodate the resulting large data sets and heavy loads. This thesis presents searching algorithms and hierarchical classification techniques for increasing a search service's understanding of web queries. Existing search services rely solely on a query's occurrence in the document collection to locate relevant documents. They typically do not perform any task or topic-based analysis of queries using other available resources, and do not leverage changes in user query patterns over time. Provided within are a set of techniques and metrics for performing temporal analysis on query logs. Our log analyses are shown to be reasonable and informative, and can be used to detect changing trends and patterns in the query stream, thus providing valuable data to a search service.
QUERY EXPANSION WITH ENRICHED USER PROFILES FOR PERSONALIZED SEARCH UTILIZING...Prasadu Peddi
The document proposes two novel techniques for personalized query expansion using folksonomy data. It introduces a model that constructs enriched user profiles by integrating word embeddings with topic models from user annotations and an external corpus. The first technique selects expansion terms using topical weights-enhanced word embeddings. The second calculates topical relevance between the query and user profile terms. An evaluation shows the approaches outperform existing non-personalized and personalized query expansion methods.
Scaling Down Dimensions and Feature Extraction in Document Repository Classif...ijdmtaiir
-In this study a comprehensive evaluation of two
supervised feature selection methods for dimensionality
reduction is performed - Latent Semantic Indexing (LSI) and
Principal Component Analysis (PCA). This is gauged against
unsupervised techniques like fuzzy feature clustering using
hard fuzzy C-means (FCM) . The main objective of the study is
to estimate the relative efficiency of two supervised techniques
against unsupervised fuzzy techniques while reducing the
feature space. It is found that clustering using FCM leads to
better accuracy in classifying documents in the face of
evolutionary algorithms like LSI and PCA. Results show that
the clustering of features improves the accuracy of document
classification
Filter unwanted messages from walls and blocking nonlegitimate user in osnIJSRD
This document proposes a system to filter unwanted messages from walls and block non-legitimate users in online social networks. It uses machine learning for content-based filtering of messages. Short text is classified and filtering rules are provided to block certain content. Blacklists are also used to prevent some users from posting messages temporarily. The proposed system aims to provide privacy and control over the content visible on users' walls.
New prediction method for data spreading in social networks based on machine ...TELKOMNIKA JOURNAL
Information diffusion prediction is the study of the path of dissemination of news, information, or topics in a structured data such as a graph. Research in this area is focused on two goals, tracing the information diffusion path and finding the members that determine future the next path. The major problem of traditional approaches in this area is the use of simple probabilistic methods rather than intelligent methods. Recent years have seen growing interest in the use of machine learning algorithms in this field. Recently, deep learning, which is a branch of machine learning, has been increasingly used in the field of information diffusion prediction. This paper presents a machine learning method based on the graph neural network algorithm, which involves the selection of inactive vertices for activation based on the neighboring vertices that are active in a given scientific topic. Basically, in this method, information diffusion paths are predicted through the activation of inactive vertices byactive vertices. The method is tested on three scientific bibliography datasets: The Digital Bibliography and Library Project (DBLP), Pubmed, and Cora. The method attempts to answer the question that who will be the publisher of thenext article in a specific field of science. The comparison of the proposed method with other methods shows 10% and 5% improved precision in DBL Pand Pubmed datasets, respectively.
IRJET-Semantic Based Document Clustering Using Lexical ChainsIRJET Journal
This document discusses a semantic-based document clustering approach using lexical chains. It proposes using WordNet to perform word sense disambiguation on documents to extract core semantic features represented as lexical chains. Lexical chains identify semantically related words in a text based on relations like synonyms and hypernyms. Documents are then clustered based on the lexical chains extracted. The approach aims to overcome issues in traditional clustering like synonyms and polysemy by incorporating semantic information from WordNet ontology. It is argued that identifying themes based on disambiguated semantic features extracted via lexical chains can improve text clustering performance compared to bag-of-words models. An evaluation of the approach showed better results when using a threshold of 50% for lexical chain selection.
Rule based messege filtering and blacklist management for online social networkeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Exploiting Wikipedia and Twitter for Text Mining ApplicationsIRJET Journal
This document discusses exploiting Wikipedia and Twitter for text mining applications. It explores using Wikipedia's category-article structure for text classification, subjectivity analysis, and keyword extraction. It evaluates classifying tweets as relevant/irrelevant to entities or brands and classifying tweets into topical dimensions like workplace or innovation. Features used include relatedness scores between tweet text and Wikipedia categories, topic modeling scores, and Twitter-specific features. Experimental results show the Wikipedia framework based on its category-article structure outperforms standard text mining techniques.
The increased potential of the ontologies to reduce the human interference has wide range of applications. This paper identifies requirements for an ontology development platform to innovate artificially intelligent web. To facilitate this process, RDF and OWL have been developed as standard formats for the sharing and integration of data and knowledge. The knowledge in the form of rich conceptual schemas called ontologies. Based on the framework, an architectural paradigm is put forward in view of ontology engineering and development of ontology applications and a development portal designed to support ontology engineering, content authoring and application development with a view to maximal scalability in size and complexity of semantic knowledge and flexible reuse of ontology models and ontology application processes in a distributed and collaborative engineering environment.
Association Rule Mining Based Extraction of Semantic Relations Using Markov L...IJwest
Ontology may be a conceptualization of a website into a human understandable, however machine-readable format consisting of entities, attributes, relationships and axioms. Ontologies formalize the intentional aspects of a site, whereas the denotative part is provided by a mental object that contains assertions about instances of concepts and relations. Semantic relation it might be potential to extract the whole family-tree of a outstanding personality employing a resource like Wikipedia. In a way, relations describe the linguistics relationships among the entities involve that is beneficial for a higher understanding of human language. The relation can be identified from the result of concept hierarchy extraction. The existing ontology learning process only produces the result of concept hierarchy extraction. It does not produce the semantic relation between the concepts. Here, we have to do the process of constructing the predicates and also first order logic formula. Here, also find the inference and learning weights using Markov Logic Network. To improve the relation of every input and also improve the relation between the contents we have to propose the concept of ARSRE. This method can find the frequent items between concepts and converting the extensibility of existing lightweight ontologies to formal one. The experimental results can produce the good extraction of semantic relations compared to state-of-art method.
Sentimental classification analysis of polarity multi-view textual data using...IJECEIAES
The data and information available in most community environments is complex in nature. Sentimental data resources may possibly consist of textual data collected from multiple information sources with different representations and usually handled by different analytical models. These types of data resource characteristics can form multi-view polarity textual data. However, knowledge creation from this type of sentimental textual data requires considerable analytical efforts and capabilities. In particular, data mining practices can provide exceptional results in handling textual data formats. Besides, in the case of the textual data exists as multi-view or unstructured data formats, the hybrid and integrated analysis efforts of text data mining algorithms are vital to get helpful results. The objective of this research is to enhance the knowledge discovery from sentimental multi-view textual data which can be considered as unstructured data format to classify the polarity information documents in the form of two different categories or types of useful information. A proposed framework with integrated data mining algorithms has been discussed in this paper, which is achieved through the application of X-means algorithm for clustering and HotSpot algorithm of association rules. The analysis results have shown improved accuracies of classifying the sentimental multi-view textual data into two categories through the application of the proposed framework on online polarity user-reviews dataset upon a given topics.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
DOW JONES INDUSTRIAL AVERAGE Time series Data Analysis: Analysis and Results ...Ajay Bidyarthy
This document analyzes time series data of the Dow Jones Industrial Average from 2005 to 2012. Various statistical analyses are performed, including calculating descriptive statistics, plotting scatter plots and histograms, and performing autocorrelation and partial autocorrelation analyses. Autoregressive (AR) models (AR(1) and AR(2)) are fitted to the log-return data and found to be stationary. Bootstrap confidence intervals are estimated for the mean and standard deviation of the original and AR models' data.
Understanding Reality.- The Allisone Paradigm
This presentation deals with Seven Principles of Reality to help you understand the nature of reality.
These Principles are:
1 Fractals
2. Sacred Geometry
3. Cymatics
4. Swarm Intelligence
5. Holography
6. Biocentrism
7. The Trinity
Another Huge Paradigm Shift
In 1610, Galileo changed the Astrological Paradigm. Again, in 1905, Albert Einstein changed the Scientific Paradigm. Now, our current paradigm is about to make another profound change. This new paradigm is already here – it is called the Allisone Paradigm™.
For thousands of years, humanity’s lack of understanding of the fundamental principles of reality has diminished the quality of life on our planet.
We have the ability to change our lives for the better -- by “creating” the things we want in our lives instead of creating the things we don't want.
As the Allisone Paradigm™ becomes a widely accepted belief system, peace will begin to break out on the earth – because each of us will become aware that others are actually One with us – and we cannot harm (or help) another without harming (or benefiting) ourselves.
This Oneness principle is not just a sweet religious concept – it is pure physical reality; To believe otherwise means that we have become ensnared in the illusions that we have been taught all of our lives.
Wonderful scientific and spiritual insights will come to you as you begin to comprehend Oneness principles.
David Allisone -- is a Petroleum Engineer and Geologist with an MBA. He is also halfway to a Masters Degree in Psychology. David was an USAF Intelligence Officer during the Vietnam War / Cold War and he has worked as an engineer and computer specialist with five Fortune-100 Companies. He was the Chairman of Houston’s oldest Internet Company and he is the founder of the petroleum industry’s oldest project listing service, which connects energy projects with funding sources. After spending years as an agnostic, David has regained an appreciation for Christianity, as well as other religions of the world and, accordingly, he felt inspired to revive the ancient non-denominational order of the Children of the Law of One. He has been happily married for over 50 years and he has four sons, thirteen grandchildren, and he has lived in The Woodlands, Texas for the past 38 years. While David actively participates in several spiritual traditions, he follows but one religion – called TRUTH.
Phone: 281-962-0400
Email: davidallisone@gmail.com
www.childrenofthelawofone.org
Where all religions meet, there is One God.
Ideliance is a precursor of so-called Semantic Web. FIrst version was developped in 1993. It allows individuals and groups to organize their personal or collective / corporate knowledge as a semantic network. This document is a presentation of the product written in year 2000. Ideliance has been marketed for large companies like Air LIquide, France Telecom, PSA, CEA, EDF, GDF, Danone, Merckk. It has been used in Military Intelligence applications. It has been designed by Sylvie Le Bars (Arkandis.com) Jean Rohmer, and implemented mainly by Stéphane Jean and Denis Poisson.
A system to filter unwanted messages from theMadan Golla
This document presents a system to filter unwanted messages from social network users' walls. It consists of three main components: filtering rules, thresholds for applying the rules which are customized for each user, and a blacklist mechanism. The filtering rules allow users to control what types of messages are allowed on their walls based on attributes of the message creator and their relationship to the user. The system aims to provide flexible and transparent filtering of messages while minimizing mistakes.
1) The document discusses challenges with achieving interoperability between ultra large scale systems due to heterogeneity in platforms, data, and semantics.
2) It proposes a three-layered model for interoperability using web service technologies and semantic web approaches to address these challenges.
3) Key aspects of interoperability discussed include different levels (e.g. syntactic, semantic), use of ontologies to provide common understandings and resolve conflicts, and semantic web service approaches like OWL-S that semantically annotate service descriptions.
Semantic Massage Addressing based on Social Cloud Actor's InterestsCSCJournals
Wireless communication with Mobile Terminals has become popular tools for collecting and sending information and data. With mobile communication comes the Short Message Service (SMS) technology which is an ideal way to stay connected with anyone, anywhere anytime to help maintain business relationships with customers. Sending individual SMS messages to long list of mobile numbers can be very time consuming, and face problems of wireless communications such as variable and asymmetric bandwidth, geographical mobility and high usage costs and face the rigidity of lists. This paper proposes a technique that assures sending the message to semantically specified group of recipients. A recipient group is automatically identified based on personal information (interests, work place, publications, social relationships, etc.) and behavior based on a populated ontology created by integrating the publicly available FOAF (Friend-of-a-Friend) documents. We demonstrate that our simple technique can first, ensure extracting groups effectively according to the descriptive attributes and second send SMS effectively and can help combat unintentional spam and preserve the privacy of mobile numbers and even individual identities. The technique provides fast, effective, and dynamic solution to save time in constructing lists and sending group messages which can be applied both on personal level or in business.
This document summarizes a research paper that proposes and compares fuzzy and Naive Bayes models for detecting obfuscated plagiarism in Marathi language texts. It first provides background on plagiarism detection and describes different types of plagiarism, including obfuscated plagiarism. It then presents the fuzzy semantic similarity model, which uses fuzzy logic rules and semantic relatedness between words to calculate similarity scores between texts. Next, it describes the Naive Bayes model for plagiarism detection using Bayes' theorem. The paper compares the performance of the fuzzy and Naive Bayes models on precision, recall, F-measure and granularity. It finds that the Naive Bayes model provides more accurate detection of obfuscated plagiar
A Review: Text Classification on Social Media DataIOSR Journals
This document provides a review of different classifiers used for text classification on social media data. It discusses how social media data is often unstructured and contains users' opinions and sentiments. Various machine learning algorithms can be used to classify this social media text data, extracting meaningful information. The document focuses on describing Naive Bayes classifiers, which are commonly used for text classification tasks. It explains how Naive Bayes classifiers work by calculating the posterior probability that a document belongs to a certain class, based on applying Bayes' theorem with an independence assumption between features.
Context Driven Technique for Document ClassificationIDES Editor
In this paper we present an innovative hybrid Text
Classification (TC) system that bridges the gap between
statistical and context based techniques. Our algorithm
harnesses contextual information at two stages. First it extracts
a cohesive set of keywords for each category by using lexical
references, implicit context as derived from LSA and wordvicinity
driven semantics. And secondly, each document is
represented by a set of context rich features whose values are
derived by considering both lexical cohesion as well as the extent
of coverage of salient concepts via lexical chaining. After
keywords are extracted, a subset of the input documents is
apportioned as training set. Its members are assigned categories
based on their keyword representation. These labeled
documents are used to train binary SVM classifiers, one for
each category. The remaining documents are supplied to the
trained classifiers in the form of their context-enhanced feature
vectors. Each document is finally ascribed its appropriate
category by an SVM classifier.
Classification-based Retrieval Methods to Enhance Information Discovery on th...IJMIT JOURNAL
The widespread adoption of the World-Wide Web (the Web) has created challenges both for society as a whole and for the technology used to build and maintain the Web. The ongoing struggle of information retrieval systems is to wade through this vast pile of data and satisfy users by presenting them with information that most adequately it’s their needs. On a societal level, the Web is expanding faster than we can comprehend its implications or develop rules for its use. The ubiquitous use of the Web has raised important social concerns in the areas of privacy, censorship, and access to information. On a technical level, the novelty of the Web and the pace of its growth have created challenges not only in the development of new applications that realize the power of the Web, but also in the technology needed to scale applications to accommodate the resulting large data sets and heavy loads. This thesis presents searching algorithms and hierarchical classification techniques for increasing a search service's understanding of web queries. Existing search services rely solely on a query's occurrence in the document collection to locate relevant documents. They typically do not perform any task or topic-based analysis of queries using other available resources, and do not leverage changes in user query patterns over time. Provided within are a set of techniques and metrics for performing temporal analysis on query logs. Our log analyses are shown to be reasonable and informative, and can be used to detect changing trends and patterns in the query stream, thus providing valuable data to a search service.
QUERY EXPANSION WITH ENRICHED USER PROFILES FOR PERSONALIZED SEARCH UTILIZING...Prasadu Peddi
The document proposes two novel techniques for personalized query expansion using folksonomy data. It introduces a model that constructs enriched user profiles by integrating word embeddings with topic models from user annotations and an external corpus. The first technique selects expansion terms using topical weights-enhanced word embeddings. The second calculates topical relevance between the query and user profile terms. An evaluation shows the approaches outperform existing non-personalized and personalized query expansion methods.
Scaling Down Dimensions and Feature Extraction in Document Repository Classif...ijdmtaiir
-In this study a comprehensive evaluation of two
supervised feature selection methods for dimensionality
reduction is performed - Latent Semantic Indexing (LSI) and
Principal Component Analysis (PCA). This is gauged against
unsupervised techniques like fuzzy feature clustering using
hard fuzzy C-means (FCM) . The main objective of the study is
to estimate the relative efficiency of two supervised techniques
against unsupervised fuzzy techniques while reducing the
feature space. It is found that clustering using FCM leads to
better accuracy in classifying documents in the face of
evolutionary algorithms like LSI and PCA. Results show that
the clustering of features improves the accuracy of document
classification
Filter unwanted messages from walls and blocking nonlegitimate user in osnIJSRD
This document proposes a system to filter unwanted messages from walls and block non-legitimate users in online social networks. It uses machine learning for content-based filtering of messages. Short text is classified and filtering rules are provided to block certain content. Blacklists are also used to prevent some users from posting messages temporarily. The proposed system aims to provide privacy and control over the content visible on users' walls.
New prediction method for data spreading in social networks based on machine ...TELKOMNIKA JOURNAL
Information diffusion prediction is the study of the path of dissemination of news, information, or topics in a structured data such as a graph. Research in this area is focused on two goals, tracing the information diffusion path and finding the members that determine future the next path. The major problem of traditional approaches in this area is the use of simple probabilistic methods rather than intelligent methods. Recent years have seen growing interest in the use of machine learning algorithms in this field. Recently, deep learning, which is a branch of machine learning, has been increasingly used in the field of information diffusion prediction. This paper presents a machine learning method based on the graph neural network algorithm, which involves the selection of inactive vertices for activation based on the neighboring vertices that are active in a given scientific topic. Basically, in this method, information diffusion paths are predicted through the activation of inactive vertices byactive vertices. The method is tested on three scientific bibliography datasets: The Digital Bibliography and Library Project (DBLP), Pubmed, and Cora. The method attempts to answer the question that who will be the publisher of thenext article in a specific field of science. The comparison of the proposed method with other methods shows 10% and 5% improved precision in DBL Pand Pubmed datasets, respectively.
IRJET-Semantic Based Document Clustering Using Lexical ChainsIRJET Journal
This document discusses a semantic-based document clustering approach using lexical chains. It proposes using WordNet to perform word sense disambiguation on documents to extract core semantic features represented as lexical chains. Lexical chains identify semantically related words in a text based on relations like synonyms and hypernyms. Documents are then clustered based on the lexical chains extracted. The approach aims to overcome issues in traditional clustering like synonyms and polysemy by incorporating semantic information from WordNet ontology. It is argued that identifying themes based on disambiguated semantic features extracted via lexical chains can improve text clustering performance compared to bag-of-words models. An evaluation of the approach showed better results when using a threshold of 50% for lexical chain selection.
Rule based messege filtering and blacklist management for online social networkeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Exploiting Wikipedia and Twitter for Text Mining ApplicationsIRJET Journal
This document discusses exploiting Wikipedia and Twitter for text mining applications. It explores using Wikipedia's category-article structure for text classification, subjectivity analysis, and keyword extraction. It evaluates classifying tweets as relevant/irrelevant to entities or brands and classifying tweets into topical dimensions like workplace or innovation. Features used include relatedness scores between tweet text and Wikipedia categories, topic modeling scores, and Twitter-specific features. Experimental results show the Wikipedia framework based on its category-article structure outperforms standard text mining techniques.
The increased potential of the ontologies to reduce the human interference has wide range of applications. This paper identifies requirements for an ontology development platform to innovate artificially intelligent web. To facilitate this process, RDF and OWL have been developed as standard formats for the sharing and integration of data and knowledge. The knowledge in the form of rich conceptual schemas called ontologies. Based on the framework, an architectural paradigm is put forward in view of ontology engineering and development of ontology applications and a development portal designed to support ontology engineering, content authoring and application development with a view to maximal scalability in size and complexity of semantic knowledge and flexible reuse of ontology models and ontology application processes in a distributed and collaborative engineering environment.
Association Rule Mining Based Extraction of Semantic Relations Using Markov L...IJwest
Ontology may be a conceptualization of a website into a human understandable, however machine-readable format consisting of entities, attributes, relationships and axioms. Ontologies formalize the intentional aspects of a site, whereas the denotative part is provided by a mental object that contains assertions about instances of concepts and relations. Semantic relation it might be potential to extract the whole family-tree of a outstanding personality employing a resource like Wikipedia. In a way, relations describe the linguistics relationships among the entities involve that is beneficial for a higher understanding of human language. The relation can be identified from the result of concept hierarchy extraction. The existing ontology learning process only produces the result of concept hierarchy extraction. It does not produce the semantic relation between the concepts. Here, we have to do the process of constructing the predicates and also first order logic formula. Here, also find the inference and learning weights using Markov Logic Network. To improve the relation of every input and also improve the relation between the contents we have to propose the concept of ARSRE. This method can find the frequent items between concepts and converting the extensibility of existing lightweight ontologies to formal one. The experimental results can produce the good extraction of semantic relations compared to state-of-art method.
Sentimental classification analysis of polarity multi-view textual data using...IJECEIAES
The data and information available in most community environments is complex in nature. Sentimental data resources may possibly consist of textual data collected from multiple information sources with different representations and usually handled by different analytical models. These types of data resource characteristics can form multi-view polarity textual data. However, knowledge creation from this type of sentimental textual data requires considerable analytical efforts and capabilities. In particular, data mining practices can provide exceptional results in handling textual data formats. Besides, in the case of the textual data exists as multi-view or unstructured data formats, the hybrid and integrated analysis efforts of text data mining algorithms are vital to get helpful results. The objective of this research is to enhance the knowledge discovery from sentimental multi-view textual data which can be considered as unstructured data format to classify the polarity information documents in the form of two different categories or types of useful information. A proposed framework with integrated data mining algorithms has been discussed in this paper, which is achieved through the application of X-means algorithm for clustering and HotSpot algorithm of association rules. The analysis results have shown improved accuracies of classifying the sentimental multi-view textual data into two categories through the application of the proposed framework on online polarity user-reviews dataset upon a given topics.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
DOW JONES INDUSTRIAL AVERAGE Time series Data Analysis: Analysis and Results ...Ajay Bidyarthy
This document analyzes time series data of the Dow Jones Industrial Average from 2005 to 2012. Various statistical analyses are performed, including calculating descriptive statistics, plotting scatter plots and histograms, and performing autocorrelation and partial autocorrelation analyses. Autoregressive (AR) models (AR(1) and AR(2)) are fitted to the log-return data and found to be stationary. Bootstrap confidence intervals are estimated for the mean and standard deviation of the original and AR models' data.
Understanding Reality.- The Allisone Paradigm
This presentation deals with Seven Principles of Reality to help you understand the nature of reality.
These Principles are:
1 Fractals
2. Sacred Geometry
3. Cymatics
4. Swarm Intelligence
5. Holography
6. Biocentrism
7. The Trinity
Another Huge Paradigm Shift
In 1610, Galileo changed the Astrological Paradigm. Again, in 1905, Albert Einstein changed the Scientific Paradigm. Now, our current paradigm is about to make another profound change. This new paradigm is already here – it is called the Allisone Paradigm™.
For thousands of years, humanity’s lack of understanding of the fundamental principles of reality has diminished the quality of life on our planet.
We have the ability to change our lives for the better -- by “creating” the things we want in our lives instead of creating the things we don't want.
As the Allisone Paradigm™ becomes a widely accepted belief system, peace will begin to break out on the earth – because each of us will become aware that others are actually One with us – and we cannot harm (or help) another without harming (or benefiting) ourselves.
This Oneness principle is not just a sweet religious concept – it is pure physical reality; To believe otherwise means that we have become ensnared in the illusions that we have been taught all of our lives.
Wonderful scientific and spiritual insights will come to you as you begin to comprehend Oneness principles.
David Allisone -- is a Petroleum Engineer and Geologist with an MBA. He is also halfway to a Masters Degree in Psychology. David was an USAF Intelligence Officer during the Vietnam War / Cold War and he has worked as an engineer and computer specialist with five Fortune-100 Companies. He was the Chairman of Houston’s oldest Internet Company and he is the founder of the petroleum industry’s oldest project listing service, which connects energy projects with funding sources. After spending years as an agnostic, David has regained an appreciation for Christianity, as well as other religions of the world and, accordingly, he felt inspired to revive the ancient non-denominational order of the Children of the Law of One. He has been happily married for over 50 years and he has four sons, thirteen grandchildren, and he has lived in The Woodlands, Texas for the past 38 years. While David actively participates in several spiritual traditions, he follows but one religion – called TRUTH.
Phone: 281-962-0400
Email: davidallisone@gmail.com
www.childrenofthelawofone.org
Where all religions meet, there is One God.
Ideliance is a precursor of so-called Semantic Web. FIrst version was developped in 1993. It allows individuals and groups to organize their personal or collective / corporate knowledge as a semantic network. This document is a presentation of the product written in year 2000. Ideliance has been marketed for large companies like Air LIquide, France Telecom, PSA, CEA, EDF, GDF, Danone, Merckk. It has been used in Military Intelligence applications. It has been designed by Sylvie Le Bars (Arkandis.com) Jean Rohmer, and implemented mainly by Stéphane Jean and Denis Poisson.
This document discusses fuzzy rules and fuzzy reasoning. It covers the extension principle, fuzzy relations, fuzzy if-then rules, the compositional rule of inference, and fuzzy reasoning using single and multiple rules with single and multiple antecedents. Methods like max-min and max-product composition are presented for combining fuzzy relations. Linguistic variables and terms that take linguistic values like "old" are also introduced.
The document is a social media handbook for the United States Army. It provides guidance on using social media for soldiers and army personnel. It emphasizes maintaining operational security and not sharing sensitive information online. The handbook also outlines standards for army leaders on social media and provides checklists for establishing an official social media presence and handling crisis communications through social media.
Artificial Intelligence Past Present and FutureJean Rohmer
A presentation from IFIP Congress 2004, where I give my vision of the evolution of AI, reasons of AI winter, and belief that it is the monly way to improve Information Processing in the future
L'informatique n'est pas l'amie des donnéesJean Rohmer
Voici la présentation que j'ai faire au colloque GREC-O "Les systèmes complexes face au tsunami exponentiel du numérique".
Pour moi, une donnée est une phrase entière, un énoncé.
J'y explique que l'ordinateur "matériel" n'a pas été fait pour traiter les données, les langages de programmation non plus.
Et que cela handicape beaucoup les utilisations de l'informatique.
New Challenges in Learning Classifier Systems: Mining Rarities and Evolving F...Albert Orriols-Puig
The document discusses new challenges in learning classifier systems (LCS) when dealing with domains containing rare classes. It proposes using a design decomposition approach to analyze how LCS address rare classes. Specifically, it examines how the extended classifier system (XCS) handles rare classes. It identifies five critical elements of LCS that are important for detecting small niches associated with rare classes: 1) estimating classifier parameters correctly, 2) providing representatives of rare niches during initialization, 3) generating and growing representatives of rare niches, 4) adjusting the genetic algorithm application rate, and 5) ensuring representatives of rare niches dominate their niches. The document focuses on analyzing the first element of estimating classifier parameters for XCS when dealing with domains
The document discusses particle swarm optimization (PSO), which is a population-based optimization technique where multiple candidate solutions called particles fly through the problem search space looking for the optimal position. Each particle adjusts its position based on its own experience and the experience of neighboring particles. The procedure for implementing PSO involves initializing particles with random positions and velocities, evaluating each particle, updating particles' velocities and positions based on personal and global best experiences, and repeating until a stopping criterion is met. The document also discusses modifications to basic PSO such as limiting maximum velocity, adding an inertia weight, using a constriction factor, features of PSO, and strategies for selecting PSO parameters.
This document explains how to build a deductive inference engine for rule-based systems, business rules. It leads to a useful architecure for Complex Event Processing and Data streams
A brief introduction on the principles of particle swarm optimizaton by Rajorshi Mukherjee. This presentation has been compiled from various sources (not my own work) and proper references have been made in the bibliography section for further reading. This presentation was made as a presentation for submission for our college subject Soft Computing.
Intelligence Artificielle: résolution de problèmes en Prolog ou Prolog pour l...Jean Rohmer
Ce papier explique en détail et de manière pédagogique comment résoudre des problèmes en intelligence artificielle à l'aide du langage Prolog. Les classiques du loup, chèvre et chou, et de la tour de Hanoï sont expliqués en détail. On décrit comment appliquer l'approche "general problem solver" en Prolog
Introduction and architecture of expert systempremdeshmane
An expert system is an interactive computer program that uses knowledge acquired from experts to solve complex problems in a specific domain. It consists of an inference engine that applies rules and logic to the facts contained within a knowledge base in order to provide recommendations or advice to users. The first expert system was called DENDRAL and was developed in the 1970s at Stanford University to identify unknown organic molecules. Expert systems are used in applications like diagnosis, financial planning, configuration, and more to perform tasks previously requiring human expertise. They have benefits like increased productivity and quality, reduced costs and errors, and the ability to capture scarce human knowledge. However, they also have limitations such as difficulty acquiring and representing human expertise and an inability to operate outside their
This document discusses particle swarm optimization (PSO), which is an optimization technique inspired by swarm intelligence. It summarizes that PSO was developed in 1995 and can be applied to various search and optimization problems. PSO works by having a swarm of particles that communicate locally to find the best solution within a search space, balancing exploration and exploitation.
The document discusses Particle Swarm Optimization (PSO), which is an optimization technique inspired by swarm intelligence and the social behavior of bird flocking. PSO initializes a population of random solutions and searches for optima by updating generations of candidate solutions. Each candidate, or particle, updates its position based on its own experience and the experience of neighboring highly-ranked particles. The algorithm is simple to implement and converges quickly to produce approximate solutions to difficult optimization problems.
An expert system is a knowledge-based information system that uses knowledge from a specific domain to provide information to users like a human expert. Expert systems are useful when human experts are unavailable, inconsistent, or unable to clearly explain decisions. They can be applied when a problem lacks a clear algorithmic solution, is hazardous, has a scarcity of human experts, or requires standardization. Some examples of early expert systems include LITHIAN which advised archaeologists and DENDRAL which identified chemical structures. Expert systems have advantages like enhanced decision quality, reduced consulting costs, and ability to solve complex problems, but developing and maintaining them can be difficult and expensive.
The document discusses expert systems, which are designed to solve real problems in a particular domain that normally require human expertise. Developing an expert system involves extracting knowledge from domain experts. The key components of an expert system are the knowledge base, inference engine, explanation facility, knowledge acquisition facility, and user interface. Expert systems use knowledge rather than data to solve problems and can explain their reasoning. They have limitations such as being difficult to maintain and only applicable to narrow problems.
Lecture5 Expert Systems And Artificial IntelligenceKodok Ngorex
Expert systems aim to emulate human expertise by storing knowledge provided by human experts. They utilize various artificial intelligence techniques like rule-based reasoning, pattern recognition, and case-based reasoning to solve complex problems. An expert system consists of a user interface, knowledge base containing domain-specific knowledge, and an inference engine that applies logic and reasoning to the knowledge base. While expert systems can increase availability of expertise, there are limitations in coding human common sense and adapting to new problems.
Content Based Message Filtering For OSNS Using Machine Learning ClassifierIJMER
The document proposes a content-based message filtering system for online social networks (OSNs) using machine learning classifiers. It aims to filter unwanted messages from OSN user walls. The system uses a machine learning classifier to categorize messages and implements customizable filtering rules. It also includes a blacklist mechanism to block users who frequently post unwanted content. The architecture is divided into three layers: a social network manager layer, a content filtering layer using classifiers, and a graphical user interface layer. Filtering rules allow restricting messages based on sender attributes and relationships. Blacklist rules determine which users to block based on the percentage of their messages that violate rules.
This document describes a system called Filtered Wall (FW) that aims to filter unwanted messages from users' walls on online social networks (OSNs). The system uses machine learning techniques like radial basis function networks to classify short text messages as neutral or non-neutral. Non-neutral messages are further classified into categories. The system also provides flexible rules that allow users to specify which content should not be displayed on their walls based on criteria like user relationships, profiles, and user-defined blacklists. When a user posts a message, the system extracts metadata using text classification and enforces the user's filtering rules to determine if the message will be published or filtered.
This document discusses machine learning techniques for filtering unwanted messages in online social networks. It proposes a content-based filtering system that allows users to control the messages posted on their walls by filtering out unwanted messages. The system uses a machine learning-based classifier to automatically categorize short text messages based on their content. It also includes a blacklist feature to block specific users from posting if they consistently share unwanted messages. The goal is to give users better control over their social media experience by reducing noise and unwanted content on their walls.
The document proposes a system to automatically filter unwanted messages from online social network user walls based on message content and the relationship between the message creator and recipient. It utilizes machine learning text classification techniques to categorize messages and provides flexible rules that allow users to customize filtering criteria for their walls. The system was found to effectively filter political and vulgar messages while allowing for personalized control over wall content.
An automatic filtering task in osn using content based approachIAEME Publication
This document summarizes an academic paper on developing an automatic filtering system for online social networks using content-based approaches. It describes a three-tier architecture for the filtering system, with the lowest layer managing social networks, a middle layer performing message categorization and blacklisting, and a top layer providing a graphical user interface. The system works by intercepting messages, extracting metadata using machine learning classification, applying filtering and blacklisting rules, and publishing approved messages while filtering unwanted ones based on content and creator. It aims to allow users more control over messages on their walls by blocking offensive, political, or other undesirable content in an automatic way.
A novel method for generating an elearning ontologyIJDKP
The Semantic Web provides a common framework that allows data to be shared and reused across
applications, enterprises, and community boundaries. The existing web applications need to express
semantics that can be extracted from users' navigation and content, in order to fulfill users' needs. Elearning
has specific requirements that can be satisfied through the extraction of semantics from learning
management systems (LMS) that use relational databases (RDB) as backend. In this paper, we propose
transformation rules for building owl ontology from the RDB of the open source LMS Moodle. It allows
transforming all possible cases in RDBs into ontological constructs. The proposed rules are enriched by
analyzing stored data to detect disjointness and totalness constraints in hierarchies, and calculating the
participation level of tables in n-ary relations. In addition, our technique is generic; hence it can be applied
to any RDB.
Filter unwanted messages from walls and blocking nonlegitimate user in osnIJSRD
Today’s life is totally based on Internet. Now a days people cannot imagine life without Internet. Information and communication technology plays vital role in today’s online networked society. In today’s life, we are very close to the online social networks. Online social networks are used for posting and sharing information across various social networking sites. But user’s privacy is not maintained by online social networks. For maintaining users sensitive information’s privacy online social networks provides little or no support. For filtering unwanted messages we propose a system using machine learning (ML). Using machine learning in soft classifier content based filtering performed. In proposed system filtering rules (FR’s) are provided for content independent filtering.. Blacklists are used for more flexibility by which filtering choices are increased. Proposed system provides security to the Online Social Networks.
Here are the key points about using content-based filtering techniques:
- Content-based filtering relies on analyzing the content or description of items to recommend items similar to what the user has liked in the past. It looks for patterns and regularities in item attributes/descriptions to distinguish highly rated items.
- The item content/descriptions are analyzed automatically by extracting information from sources like web pages, or entered manually from product databases.
- It focuses on objective attributes about items that can be extracted algorithmically, like text analysis of documents.
- However, personal preferences and what makes an item appealing are often subjective qualities not easily extracted algorithmically, like writing style or taste.
- So while content-based filtering can
Filter unwanted messages from walls and blocking non legitimate users in osnIAEME Publication
1. The document presents a system to filter unwanted messages from user walls in online social networks. It aims to give users more control over the content that appears on their walls.
2. A machine learning classifier is used to automatically label messages by category. Users can then specify filtering rules to block certain categories or keywords from appearing.
3. The system also implements a blacklist to temporarily or permanently block users who frequently post unwanted content, as determined by filtering rules and a threshold.
International Journal of Engineering Research and Development (IJERD)IJERD Editor
This document presents a novel approach for clustering textual information in emails using text data mining techniques. It discusses using k-means clustering and a vector space model to group similar emails based on word patterns and frequencies. The methodology involves preprocessing emails, applying a Porter stemmer, calculating term frequencies, and using k-means to form clusters. Clusters will contain emails with similar content, allowing users to more easily process emails based on priority. This clustering approach could reduce the time users spend filtering through emails one by one.
Building a recommendation system based on the job offers extracted from the w...IJECEIAES
Recruitment, or job search, is increasingly used throughout the world by a large population of users through various channels, such as websites, platforms, and professional networks. Given the large volume of information related to job descriptions and user profiles, it is complicated to appropriately match a user's profile with a job description, and vice versa. The job search approach has drawbacks since the job seeker needs to search a job offers in each recruitment platform, manage their accounts, and apply for the relevant job vacancies, which wastes considerable time and effort. The contribution of this research work is the construction of a recommendation system based on the job offers extracted from the web and on the e-portfolios of job seekers. After the extraction of the data, natural language processing is applied to structured data and is ready for filtering and analysis. The proposed system is a content-based system, it measures the degree of correspondence between the attributes of the e-portfolio with those of each job offer of the same list of competence specialties using the Euclidean distance, the result is classified with a decreasing way to display the most relevant to the least relevant job offers
This document discusses web document clustering using a hybrid approach in data mining. It begins with an abstract describing the huge amount of data on the internet and need to organize web documents into clusters. It then discusses requirements for document clustering like scalability, noise tolerance, and ability to present concise cluster summaries. Different existing document clustering approaches are described, including text-based and link-based approaches. The proposed approach uses a concept-based mining model along with hierarchical agglomerative clustering and link-based algorithms to cluster web documents based on both their content and hyperlinks. This hybrid approach aims to provide more relevant clustered documents to users than previous methods.
A Framework for Content Preparation to Support Open-Corpus Adaptive HypermediaKillian Levacher
This document proposes a framework for preparing open-corpus content to support adaptive hypermedia systems. The framework includes three components: 1) structural analysis to segment web pages and remove redundant information, 2) statistical analysis to extract concepts using techniques like hidden Markov models, and 3) intelligent slicing to fulfill specific information requests from adaptive systems by retrieving and tailoring open-corpus content. The goal is to leverage existing open web content for adaptive systems by automatically preparing and enriching content with metadata in a format agnostic to any specific system.
IRJET- Concept Extraction from Ambiguous Text Document using K-MeansIRJET Journal
This document discusses using a K-means clustering algorithm to extract concepts from ambiguous text documents. It involves preprocessing the text by tokenizing, removing stop words, and stemming words. The words are then represented as vectors and dimensionality reduction using PCA is applied. Finally, K-means clustering is used to group similar words into clusters to identify the overall concepts in the document without reading the entire text. The aim is to help users understand the key topics in a document in a time-efficient manner without having to read the full text.
Great model a model for the automatic generation of semantic relations betwee...ijcsity
The
large
a
v
ailable
am
ou
n
t
of
non
-
structured
texts
that
b
e
-
long
to
differe
n
t
domains
su
c
h
as
healthcare
(e.g.
medical
records),
justice
(e.g.
l
a
ws,
declarations),
insurance
(e.g.
declarations),
etc. increases
the
effort
required
for
the
analysis
of
information
in
a
decision making
pro
-
cess.
Differe
n
t
pr
o
jects
and t
o
ols
h
av
e
pro
p
osed
strategies
to
reduce
this
complexi
t
y
b
y
classifying,
summarizing
or
annotating
the
texts.
P
artic
-
ularl
y
,
text
summary
strategies
h
av
e
pr
ov
en
to
b
e
v
ery
useful
to
pr
o
vide
a
compact
view
of
an
original
text.
H
ow
e
v
er,
the
a
v
ailable
strategies
to
generate
these
summaries
do
not
fit
v
ery
w
ell
within
the
domains
that
require
ta
k
e
i
n
to
consideration
the
tem
p
oral
dimension
of
the
text
(e.g.
a
rece
n
t
piece
of
text
in
a
medical
record
is
more
im
p
orta
n
t
than
a
pre
-
vious
one)
and
the
profile
of
the
p
erson
who
requires
the
summary
(e.g
the
medical
s
p
ecialization).
T
o
co
p
e with
these
limitations
this
pa
p
er
prese
n
ts
”GRe
A
T”
a
m
o
del
for
automatic
summary
generation
that
re
-
lies
on
natural
language
pr
o
cessing
and
text
mining
te
c
hniques
to
extract
the
most
rele
v
a
n
t
information
from
narrati
v
e
texts
and
disc
o
v
er
new
in
-
formation
from
the
detection
of
related
information. GRe
A
T
M
o
del
w
as impleme
n
ted
on
sof
tw
are
to
b
e
v
alidated
in
a
health
institution
where
it
has
sh
o
wn
to
b
e
v
ery
useful
to displ
a
y
a
preview
of
the
information
a
b
ou
t
medical
health
records
and
disc
o
v
er
new
facts
and
h
y
p
otheses
within
the
information.
Se
v
eral
tests
w
ere
executed
su
c
h
as
F
unctional
-
i
t
y
,
Usabili
t
y
and
P
erformance
regarding
to
the
impleme
n
ted
sof
t
w
are.
In
addition,
precision
and
recall
measures
w
ere
applied
on
the
results
ob
-
tained
through
the
impleme
n
ted
t
o
ol,
as
w
ell
as
on
the
loss
of
information
obtained
b
y
pr
o
viding
a
text
more
shorter than
the
original
An in-depth review on News Classification through NLPIRJET Journal
This document provides an in-depth literature review of news classification through natural language processing (NLP). It discusses several existing approaches to news classification, including models that use convolutional neural networks (CNNs), graph-based approaches, and attention mechanisms. The document also notes that current search engines often return too many irrelevant results, so classification could help layer search results. It concludes that while many techniques have been developed, inconsistencies remain in effectively classifying news, so further research on combining NLP, feature extraction, and fuzzy logic is needed.
The socialization of the web has undertaken a new dimension after the emergence of the Online
Social Networks (OSN) concept. The fact that each Internet user becomes a potential content
creator entails managing a big amount of data. This paper explores the most popular
professional OSN: LinkedIn. A scraping technique was implemented to get around 5 Million
public profiles. The application of natural language processing techniques (NLP) to classify the
educational background and to cluster the professional background of the collected profiles led
us to provide some insights about this OSN’s users and to evaluate the relationships between
educational degrees and professional careers.
Scraping and Clustering Techniques for the Characterization of Linkedin Profilescsandit
The socialization of the web has undertaken a new dimension after the emergence of the Online
Social Networks (OSN) concept. The fact that each Internet user becomes a potential content
creator entails managing a big amount of data. This paper explores the most popular
professional OSN: LinkedIn. A scraping technique was implemented to get around 5 Million
public profiles. The application of natural language processing techniques (NLP) to classify the
educational background and to cluster the professional background of the collected profiles led
us to provide some insights about this OSN’s users and to evaluate the relationships between
educational degrees and professional careers.
World Wide Web is a huge repository of information and there is a tremendous increase in the volume of
information daily. The number of users are also increasing day by day. To reduce users browsing time lot
of research is taken place. Web Usage Mining is a type of web mining in which mining techniques are
applied in log data to extract the behaviour of users. Clustering plays an important role in a broad range
of applications like Web analysis, CRM, marketing, medical diagnostics, computational biology, and many
others. Clustering is the grouping of similar instances or objects. The key factor for clustering is some sort
of measure that can determine whether two objects are similar or dissimilar. In this paper a novel
clustering method to partition user sessions into accurate clusters is discussed. The accuracy and various
performance measures of the proposed algorithm shows that the proposed method is a better method for
web log mining.
This document provides a survey of semantic web personalization techniques. It begins by defining semantic web personalization and its advantages over traditional web personalization. It then classifies semantic web personalization approaches into several categories, including ontology-based, context-based, and hybrid recommendation systems. For each category, it provides examples of approaches and compares their methods and steps for personalization. The goal of the survey is to analyze and compare different techniques used for personalization in the semantic web.
Similar to Iaetsd hierarchical fuzzy rule based classification (20)
iaetsd Survey on cooperative relay based data transmissionIaetsd Iaetsd
The document discusses cooperative relay based data transmission and proposes a system to select the most energy efficient relay node for a source node to transmit data through. It analyzes different cooperative relaying techniques like amplify-and-forward, decode-and-forward, and compress-and-forward. The proposed system aims to minimize the source node's cost for cooperation by selecting the relay node that provides the highest energy efficiency. This allows high data transmission over long distances with improved energy efficiency compared to direct transmission without a relay.
iaetsd Software defined am transmitter using vhdlIaetsd Iaetsd
This document discusses the design and implementation of an amplitude modulation (AM) software defined radio transmitter using an FPGA. It begins with an abstract describing the goals of the project. It then provides an overview of the system design, including discussion of the individual components like the microphone, analog to digital converter, digital to analog converter, carrier frequency generator, and antenna. It describes how these components will be implemented on the FPGA, including using behavioral modeling with VHDL. It also discusses designing filters and modulation/demodulation circuits. The overall summary is that this document outlines the goals and high-level system design for creating an AM transmitter using an FPGA that can transmit an audio signal by digitally modulating a carrier frequency.
iaetsd Health monitoring system with wireless alarmIaetsd Iaetsd
The document describes a health monitoring system with wireless alarm that detects a patient's heart rate and temperature. It consists of a sensor unit worn on the wrist that monitors vital signs and transmits data wirelessly to an alarm and display unit. This allows caregivers to be alerted quickly if a patient's condition changes, such as if their heart rate is too high or low. The system uses a microcontroller to process sensor readings from a pulse oximetry sensor and transmit data via RF to the receiving unit, which contains another microcontroller connected to an RF receiver and buzzer alarm. If an abnormal heart rate is detected, the system triggers an alarm to notify caregivers.
iaetsd Equalizing channel and power based on cognitive radio system over mult...Iaetsd Iaetsd
This document summarizes a research paper about equalizing power and channel allocation in a cognitive radio system using multiuser OFDM. It discusses how frequency spectrum is becoming scarce due to increased wireless usage, and how cognitive radio can help improve spectrum utilization by allowing unlicensed secondary users to access licensed bands opportunistically when primary users are not using them. The paper presents a system model for a cognitive radio network with one primary user and multiple secondary user pairs. It formulates the problem of allocating subcarriers and power to the secondary users while avoiding interference to the primary user.
iaetsd Economic analysis and re design of driver’s car seatIaetsd Iaetsd
The document discusses redesigning car seats to improve comfort. It notes that car seat design must balance comfort, safety, and health. Static comfort relates to the seat's form and support, while dynamic comfort considers vibration levels. The study reexamines existing car seat designs and proposes a novel design with improvements in form, features, usability, and comfort. A survey was also conducted to define important comfort factors like pain prevention to help guide future seat designs.
iaetsd Design of slotted microstrip patch antenna for wlan applicationIaetsd Iaetsd
This document describes the design and simulation of a slotted microstrip patch antenna for wireless local area network (WLAN) applications operating at 2.4 GHz. The antenna was designed on an FR-4 substrate with a dielectric constant of 4.2 and thickness of 1.6 mm. Simulation in HFSS showed the antenna has a voltage standing wave ratio of 1.88 at the resonant frequency, with omnidirectional radiation patterns. The compact size and simple design make this slotted patch antenna suitable for use in embedded wireless systems.
REVIEW PAPER- ON ENHANCEMENT OF HEAT TRANSFER USING RIBSIaetsd Iaetsd
This document reviews research on enhancing heat transfer using ribs mounted inside ducts. Various studies investigated ribs of different shapes, pitches, heights and angles. Continuous ribs, transverse ribs, angled ribs, and other rib configurations were examined. Most studies found that ribs increased turbulence and heat transfer compared to smooth ducts. Some key findings included V-shaped ribs providing better performance than other shapes, and certain rib pitches and angles performing better depending on parameters like Reynolds number. In general, ribs were found to effectively enhance heat transfer through boundary layer disruption and increased turbulence compared to smooth ducts.
A HYBRID AC/DC SOLAR POWERED STANDALONE SYSTEM WITHOUT INVERTER BASED ON LOAD...Iaetsd Iaetsd
This document discusses two methods for generating power from solar panels for a home without using inverters or batteries.
Method 1 proposes a hybrid AC/DC home grid system that shifts harmonic intensive loads to the DC side to reduce power conversion losses and isolates harmonic content. Solar power is fed to the home through a DC-DC converter, MPPT, and inverter to power AC loads, with a separate DC connection for DC loads.
Method 2 generates AC power directly from an array of solar cells connected in an alternating anti-parallel configuration, eliminating power losses from an inverter. Compatibility with residential loads is analyzed. This novel technique could remove the need for batteries and reduce overall cost.
The performance of
This document describes the fabrication of a dual power bike that can operate using either an internal combustion engine or electric motor. The goal is to improve fuel efficiency and reduce pollution by allowing electric-only operation in the city. The bike combines a petrol engine with a battery and electric motor, resulting in twice the fuel economy of a conventional bike. It works by using the electric motor powered by the battery for low-power city driving, and switching to the petrol engine for higher speeds or power needs. This hybrid system aims to lower costs and pollution compared to other vehicles.
This document discusses Blue Brain technology and the goal of creating an artificial brain using silicon chips. It aims to upload the contents of a natural human brain into a virtual brain. This would allow human intelligence, memories, and personalities to potentially persist after death through the virtual brain. The document outlines how nanobots could scan a human brain at a cellular level and transfer that information to a supercomputer to recreate the brain's structure and function virtually. It compares key aspects of natural and virtual brains, such as how inputs, interpretation, outputs, memory, and processing would theoretically work for a virtual brain modeled after the human brain.
iirdem The Livable Planet – A Revolutionary Concept through Innovative Street...Iaetsd Iaetsd
The document proposes an innovative street lighting and surveillance system using Internet of Things (IoT) and Li-Fi technologies. The system uses LED street lights equipped with sensors and cameras that can monitor traffic, detect crimes, and provide emergency assistance. Data from the lights would be transmitted using Li-Fi and stored in the cloud for analysis. This integrated system could save energy, reduce costs, and improve safety, traffic management, and emergency response capabilities in cities.
The document proposes a Surveillance Aided Robotic Bird (SARB) to improve on existing surveillance systems. SARB would be designed like a bird and equipped with cameras, including night vision, to monitor areas remotely. It would be powered by carbon nanotubes, allowing for wireless charging and extended flight time. SARB could track intruders under the control of image processing and fly between fixed points for charging. This would provide a more natural, mobile and energy efficient form of surveillance compared to static cameras.
iirdem Growing India Time Monopoly – The Key to Initiate Long Term Rapid GrowthIaetsd Iaetsd
This document discusses how small and medium enterprises can achieve long term rapid growth. It focuses on the concept of "time monopoly" which has two components - a competitive advantage over competitors in a niche market, and a time advantage where it takes competitors longer to catch up.
The literature review discusses different sources of competitive advantage according to Porter, including variety-based positioning, needs-based positioning, and access-based positioning. It also discusses the importance of fit between different business activities for achieving a competitive advantage.
The paper proposes five propositions for rapid growth. These include that all areas can enable or hinder growth; areas can be transformed from hindering to enabling growth; businesses need scalability; and time monopoly,
iirdem Design of Efficient Solar Energy Collector using MPPT AlgorithmIaetsd Iaetsd
This document discusses the design of an efficient solar energy collector using a Maximum Power Point Tracking (MPPT) algorithm. It aims to maximize solar energy output through the use of lenses to concentrate sunlight onto solar panels and an MPPT algorithm to track the optimal power point. The methodology involves designing a DC-DC boost converter, lens-based solar cell, and a microcontroller with driver circuit. Simulations and hardware implementation will analyze the solar array, boost converter, and verify the system collects more energy than a fixed panel system.
iirdem CRASH IMPACT ATTENUATOR (CIA) FOR AUTOMOBILES WITH THE ADVOCATION OF M...Iaetsd Iaetsd
This document describes a proposed crash impact attenuation system for automobiles that uses mechatronic systems. The system includes an accident prediction system using ultrasound sensors to monitor vehicle surroundings and detect potential collisions. It also includes a crash absorption system with components like a pneumatic cylinder attached to the vehicle chassis that can push and pull a shock energy absorber upon detection of an imminent crash by the microcontroller. This proposed system aims to reduce crash impacts and potentially save lives by fully absorbing crash forces through controlled actuation of the absorber components.
iirdem ADVANCING OF POWER MANAGEMENT IN HOME WITH SMART GRID TECHNOLOGY AND S...Iaetsd Iaetsd
1) The document describes a smart home energy management system that uses wireless sensor networks and ZigBee technology to monitor and control home appliances in real-time. Electrical parameters like voltage, current, and power consumption are measured.
2) The system allows flexible control of appliances based on consumer needs. Appliances can be monitored and controlled remotely or automatically based on power consumption thresholds.
3) Key features of the system include using a TRIAC circuit to control appliances without needing a microcontroller, and providing flexible control options to users for switching devices on/off according to their preferences. This allows improving consumer comfort while optimizing energy use.
iaetsd Shared authority based privacy preserving protocolIaetsd Iaetsd
This document proposes a Shared Authority based Privacy preserving Authentication protocol (SAPA) for handling privacy issues in cloud storage. SAPA achieves shared access authority through an anonymous access request matching mechanism. It applies attribute-based access control to allow users to reliably access their own data fields. It also uses proxy re-encryption to provide temporary authorized data sharing among multiple users. The goal is to preserve user privacy during data access and sharing in the cloud.
iaetsd Secured multiple keyword ranked search over encrypted databasesIaetsd Iaetsd
This document proposes a Robust Key-Aggregate Cryptosystem (RKAC) that allows flexible and efficient assignment of decryption rights for encrypted data stored in cloud storage. The RKAC produces constant-sized ciphertexts such that a constant-sized aggregate decryption key can decrypt any subset of ciphertexts. This allows the data owner to share access to selected encrypted files by sending a single small aggregate key to authorized users, without decrypting the files themselves or distributing individual keys. The RKAC is described as providing a secure and flexible method for sharing encrypted data stored in the cloud.
iaetsd Robots in oil and gas refineriesIaetsd Iaetsd
This document discusses attribute-based encryption in cloud computing with outsourced revocation. It proposes a pseudonym generation scheme for identity-based encryption and outsourced revocation in cloud computing. The scheme offloads most key generation operations to a Key Update Cloud Service Provider during key issuing and updating, leaving only simple operations for the Private Key Generator and users. It aims to reduce computation overhead at the Private Key Generator while using an untrusted cloud service provider.
iaetsd Modeling of solar steam engine system using parabolicIaetsd Iaetsd
The document describes the modeling and testing of a solar-steam engine system using a parabolic concentrator. The system focuses solar radiation onto a boiler to generate steam, which is then used to power an oscillating steam engine coupled to a generator to produce electricity. The parabolic dish has a diameter of 0.625m and focuses sunlight onto a 1L boiler. Testing showed the system could produce 9V with no load and 5.3V under load, demonstrating its potential for rural electrification applications.
Electric vehicle and photovoltaic advanced roles in enhancing the financial p...IJECEIAES
Climate change's impact on the planet forced the United Nations and governments to promote green energies and electric transportation. The deployments of photovoltaic (PV) and electric vehicle (EV) systems gained stronger momentum due to their numerous advantages over fossil fuel types. The advantages go beyond sustainability to reach financial support and stability. The work in this paper introduces the hybrid system between PV and EV to support industrial and commercial plants. This paper covers the theoretical framework of the proposed hybrid system including the required equation to complete the cost analysis when PV and EV are present. In addition, the proposed design diagram which sets the priorities and requirements of the system is presented. The proposed approach allows setup to advance their power stability, especially during power outages. The presented information supports researchers and plant owners to complete the necessary analysis while promoting the deployment of clean energy. The result of a case study that represents a dairy milk farmer supports the theoretical works and highlights its advanced benefits to existing plants. The short return on investment of the proposed approach supports the paper's novelty approach for the sustainable electrical system. In addition, the proposed system allows for an isolated power setup without the need for a transmission line which enhances the safety of the electrical network
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...shadow0702a
This document serves as a comprehensive step-by-step guide on how to effectively use PyCharm for remote debugging of the Windows Subsystem for Linux (WSL) on a local Windows machine. It meticulously outlines several critical steps in the process, starting with the crucial task of enabling permissions, followed by the installation and configuration of WSL.
The guide then proceeds to explain how to set up the SSH service within the WSL environment, an integral part of the process. Alongside this, it also provides detailed instructions on how to modify the inbound rules of the Windows firewall to facilitate the process, ensuring that there are no connectivity issues that could potentially hinder the debugging process.
The document further emphasizes on the importance of checking the connection between the Windows and WSL environments, providing instructions on how to ensure that the connection is optimal and ready for remote debugging.
It also offers an in-depth guide on how to configure the WSL interpreter and files within the PyCharm environment. This is essential for ensuring that the debugging process is set up correctly and that the program can be run effectively within the WSL terminal.
Additionally, the document provides guidance on how to set up breakpoints for debugging, a fundamental aspect of the debugging process which allows the developer to stop the execution of their code at certain points and inspect their program at those stages.
Finally, the document concludes by providing a link to a reference blog. This blog offers additional information and guidance on configuring the remote Python interpreter in PyCharm, providing the reader with a well-rounded understanding of the process.
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
artificial intelligence and data science contents.pptxGauravCar
What is artificial intelligence? Artificial intelligence is the ability of a computer or computer-controlled robot to perform tasks that are commonly associated with the intellectual processes characteristic of humans, such as the ability to reason.
› ...
Artificial intelligence (AI) | Definitio
Batteries -Introduction – Types of Batteries – discharging and charging of battery - characteristics of battery –battery rating- various tests on battery- – Primary battery: silver button cell- Secondary battery :Ni-Cd battery-modern battery: lithium ion battery-maintenance of batteries-choices of batteries for electric vehicle applications.
Fuel Cells: Introduction- importance and classification of fuel cells - description, principle, components, applications of fuel cells: H2-O2 fuel cell, alkaline fuel cell, molten carbonate fuel cell and direct methanol fuel cells.
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
Rainfall intensity duration frequency curve statistical analysis and modeling...bijceesjournal
Using data from 41 years in Patna’ India’ the study’s goal is to analyze the trends of how often it rains on a weekly, seasonal, and annual basis (1981−2020). First, utilizing the intensity-duration-frequency (IDF) curve and the relationship by statistically analyzing rainfall’ the historical rainfall data set for Patna’ India’ during a 41 year period (1981−2020), was evaluated for its quality. Changes in the hydrologic cycle as a result of increased greenhouse gas emissions are expected to induce variations in the intensity, length, and frequency of precipitation events. One strategy to lessen vulnerability is to quantify probable changes and adapt to them. Techniques such as log-normal, normal, and Gumbel are used (EV-I). Distributions were created with durations of 1, 2, 3, 6, and 24 h and return times of 2, 5, 10, 25, and 100 years. There were also mathematical correlations discovered between rainfall and recurrence interval.
Findings: Based on findings, the Gumbel approach produced the highest intensity values, whereas the other approaches produced values that were close to each other. The data indicates that 461.9 mm of rain fell during the monsoon season’s 301st week. However, it was found that the 29th week had the greatest average rainfall, 92.6 mm. With 952.6 mm on average, the monsoon season saw the highest rainfall. Calculations revealed that the yearly rainfall averaged 1171.1 mm. Using Weibull’s method, the study was subsequently expanded to examine rainfall distribution at different recurrence intervals of 2, 5, 10, and 25 years. Rainfall and recurrence interval mathematical correlations were also developed. Further regression analysis revealed that short wave irrigation, wind direction, wind speed, pressure, relative humidity, and temperature all had a substantial influence on rainfall.
Originality and value: The results of the rainfall IDF curves can provide useful information to policymakers in making appropriate decisions in managing and minimizing floods in the study area.
AI assisted telemedicine KIOSK for Rural India.pptx
Iaetsd hierarchical fuzzy rule based classification
1. Hierarchical fuzzy rule based classification
systems with genetic rule
selection To Filter Unwanted Messages
shaik masthan baba1
, shaik aslam2
, K.Naresh babu3
1
Computer science and engineering.
2
Computer science and engineering
3
Computer science and engineering
1
masthan201@gmail.com,2
aslambasha592@gmail.com ,3
naresh.kosuri@gmail.com
Abstract: Social networking sites that
facilitate communication of information
between users allow users to post messages
as an important function. Unnecessary posts
could spam a user’s wall, which is the page
where posts are displayed, thus disabling
the user from viewing relevant messages.
The aim of this paper is to improve the
performance of fuzzy rule based
classification systems on imbalanced
domains, increasing the granularity of the
fuzzy partitions on the boundary areas
between the classes, in order to obtain a
better separability. We propose the use of a
hierarchical fuzzy rule based classification
system using the neural network learning
model to filter out unwanted messages
from Online Social Networking (OSN)
user walls, which is based on the refinement
of a simple linguistic fuzzy model by means
of the extension of the structure of the
knowledge base in a hierarchical way and
the use of a genetic rule selection process in
order to get a compact and accurate model.
Keywords:On-line Social Networks,
Classification,Fuzzy rule based classification
systems,Imbalanced data-sets , Genetic rule
selection
1.INTRODUCTION
Online Social Networks (OSNs) are today
one of the most popular interactive medium
to communicate, share and disseminate a
considerable amount of human life
information. Daily and continuous
communications imply the exchange of
several types of content, including free text,
image, audio and video data. The huge and
dynamic character of these data creates
the premise for the employment of web
content mining strategies aimed to
automatically discover useful information
dormant within the data. They are
instrumental to provide an active support in
complex and sophisticated tasks involved
in OSN management, such as for instance
access control or information filtering.
Information filtering has been greatly
explored for what concerns textual
documents and more recently, web content
[1-3].Information filtering can therefore be
used to give users the ability to
automatically control the messages written
on their own walls, by filtering out
unwanted messages. Indeed, today OSNs
provide very little support to prevent
unwanted messages on user walls [4-6]. No
content-based preferences are supported
and therefore it is not possible to prevent
undesired messages, Providing this service
is not only a matter of using previously
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
92
2. defined web content mining techniques for a
different application, rather it requires to
design ad hoc classification strategies. The
aim of the present work is therefore to
propose and experimentally evaluate an
automated system, called Filtered Wall
(FW), able to filter unwanted messages
from OSN user walls [7-9]. We exploit
Machine Learning (ML) text categorization
techniques to automatically assign with
each short text message a set of categories
based on its content. The major efforts in
building a robust short text classifier (STC)
are concentrated in the extraction and
selection of a set of characterizing and
discriminate features. The solutions
investigated in which we inherit the learning
model and the elicitation procedure for
generating pre-classified data..In particular,
we base the overall short text classification
strategy on Radial Basis Function Networks
(RBFN) for their proven capabilities in
acting as soft classifiers, in managing noisy
data and intrinsically vague classes We
insert the neural model within a hierarchical
two level classification strategy [8-10]. In
the first level,the RBFNcategorizes short
messages as Neutral and Nonneutral;in the
second stage, No neutral messages are
classified producing gradual estimates of
appropriateness to each of the considered
category. Besides classification facilities,
the system provides a powerful rule
layer exploiting a flexible language to
specify Filtering Rules (FRs), by which
users can state what contents should not be
displayed on their walls. FRs can support a
variety of different filtering criteria that can
be combined and customized according to
the user needs. In addition, the system
provides the support for user-defined
Blacklists (BLs), that is, lists of users that
are temporarily prevented to post any kind
of messages on a user wall [11-13]. The
experiments we have carried out show
the effectiveness of the developed
filtering techniques the proposal of a
system to automatically filter unwanted
messages from OSN user walls on the basis
of both message content and the message
creator relationships and characteristics.
2. LITERATURE REVIEW
The main contribution of this paper is the
design of a system providing customizable
content-based message filtering for OSNs,
based on ML techniques. As we have
pointed out in the introduction, to the best of
our knowledge we are the first proposing
such kind of application for OSNs.
However, our work has relationships both
with the state of the art in content-based
filtering, as well as with the field of policy-
based personalization for OSNs and, more in
general, web contents. Therefore, in what
follows, we survey the literature in both
these fields.
2.1 Content-based filtering:
Information filtering systems are designed
to classify a stream of dynamically
generated information dispatched
asynchronously by an information producer
and present to the user those information
that are likely to satisfy his/her requirements
[3]. In content-based filtering each user is
assumed to operate independently. As a
result, a content-based filtering system
selects information items based on the
correlation between the content of the items
and the user preferences as opposed to a
collaborative filtering system that chooses
items based on the correlation between
people with similar preferences. Documents
processed in content-based filtering are
mostly textual in nature and this makes
content-based filtering close to text
classification. The activity of filtering can be
modeled, in fact, as a case of single label,
binary classification, partitioning incoming
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
93
3. documents into relevant and non relevant
categories [4]. More complex filtering
systems include multi-label text
categorization automatically labeling
messages into partial thematic categories.
Content-based filtering is mainly based on
the use of the ML paradigm according to
which a classifier is automatically induced
by learning from a set of pre-classified
examples. A remarkable variety of related
work has recently appeared, which differ for
the adopted feature extraction methods,
model learning, and collection of samples
[5], [6], [7], [8],[9]. The feature extraction
procedure maps text into a compact
representation of its content and is uniformly
applied to training and generalization
phases. The application of content-based
filtering on messages posted on OSN user
walls poses additional challenges given the
short length of these messages other than the
wide range of topics that can be discussed.
Short text classification has received up to
now few attention in the scientific
community. Recent work highlights
difficulties in defining robust features,
essentially due to the fact that the
description of the short text is concise, with
many misspellings, non standard terms and
noise. Focusing on the OSN domain, interest
in access control and privacy protection is
quite recent. As far as privacy is concerned,
current work is mainly focusing on privacy-
preserving data mining techniques, that is,
protecting information related to the
network, i.e., relationships/nodes, while
performing social network analysis [5].
Works more related to our proposals are
those in the field of access control. In this
field, many different access control models
and related mechanisms have been proposed
so far (e.g., [6,2,10]), which mainly differ on
the expressivity of the access control policy
language and on the way access control is
enforced (e.g., centralized vs. decentralized).
Most of these models express access control
requirements in terms of relationships that
the requestor should have with the resource
owner. We use a similar idea to identify the
users to which a filtering rule applies.
However, the overall goal of our proposal is
completely different, since we mainly deal
with filtering of unwanted contents rather
than with access control. As such, one of the
key ingredients of our system is the
availability of a description for the message
contents to be exploited by the filtering
mechanism as well as by the language to
express filtering rules. In contrast, no one of
the access control models previously cited
exploit the content of the resources to
enforce access control. We believe that this
is a fundamental difference. Moreover, the
notion of black- lists and their management
are not considered by any of these access
control models
2.2 Policy-based personalization of OSN
contents
Recently, there have been some proposals
exploiting classification mechanisms for
personalizing access in OSNs. For instance,
in [11] a classification method has been
proposed to categorize short text messages
in order to avoid overwhelming users of
microblogging services by raw data. The
system described in [11] focuses on Twitter2
and associates a set of categories with each
tweet describing its content. The user can
then view only certain types of tweets based
on his/her interests. In contrast, Golbeck and
Kuter [12] propose an application, called
FilmTrust, that exploits OSN trust
relationships and provenance information to
personalize access to the website. However,
such systems do not provide a filtering
policy layer by which the user can exploit
the result of the classification process to
decide how and to which extent filtering out
unwanted information. In contrast, our
filtering policy language allows the setting
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
94
4. of FRs according to a variety of criteria, that
do not consider only the results of the
classification process but also the
relationships of the wall owner with other
OSN users as well as information on the
user profile. Moreover, our system is
complemented by a flexible mechanism for
BL management that provides a further
opportunity of customization to the filtering
procedure. The only social networking
service we are aware of providing filtering
abilities to its users is MyWOT, a social
networking service which gives its
subscribers the ability to: 1) rate resources
with respect to four criteria: trustworthiness,
vendor reliability, privacy, and child safety;
2) specify preferences determining whether
the browser should block access to a given
resource, or should simply return a warning
message on the basis of the specified rating.
Despite the existence of some similarities,
the approach adopted by MyWOT is quite
different from ours. In particular, it supports
filtering criteria which are far less flexible
than the ones of Filtered Wall since they are
only based on the four above-mentioned
criteria. Moreover, no automatic
classification mechanism is provided to the
end user. Our work is also inspired by the
many access control models and related
policy languages and enforcement
mechanisms that have been proposed so far
for OSNs, since filtering shares several
similarities with access control. Actually,
content filtering can be considered as an
extension of access control, since it can be
used both to protect objects from
unauthorized subjects, and subjects from
inappropriate objects. In the field of OSNs,
the majority of access control models
proposed so far enforce topology-based
access control, according to which access
control requirements are expressed in terms
of relationships that the requester should
have with the resource owner. We use a
similar idea to identify the users to which a
FR applies. However, our filtering policy
language extends the languages proposed for
access control policy specification in OSNs
to cope with the extended requirements of
the filtering domain. Indeed, since we are
dealing with filtering of unwanted contents
rather than with access control, one of the
key ingredients of our system is the
availability of a description for the message
contents to be exploited by the filtering
mechanism. In contrast, no one of the access
control models previously cited exploit the
content of the resources to enforce access
control. Moreover, the notion of BLs and
their management are not considered by any
of the above-mentioned access control
models.
3. ANALYSIS OF PROBLEM:
The use of effective and appropriate
methods in facilitating projects enhances its
effectiveness and efficiency. The method
will be applied in system analysis and
design method where an existing system is
studied to proffer better options to solving
existing problems. Indeed, today OSNs
provide very little support to prevent
unwanted messages on user walls. For
example, Facebook allows users to state
who is allowed to insert messages in their
walls (i.e., friends, friends of friends, or
defined groups of friends). However, no
content-based preferences are supported and
therefore it is not possible to prevent
undesired messages, such as political or
vulgar ones, no matter of the user who posts
them. However, no content-based
preferences are supported and therefore it is
not possible to prevent undesired messages,
no matter of the user who posts them.
Providing this service is not only a matter of
using previously defined web content
mining techniques for a different
application, rather it requires to design ad
hoc classification strategies. This is because
wall messages are constituted by short text
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
95
5. for which traditional classification methods
have serious limitations since short texts do
not provide sufficient word occurrences.
4.IMBALANCED DATA-SETS IN
CLASSIFICATION
In this section, we will first introduce the
problem of imbalanced data-sets. Then, we
will describe the preprocessing technique
that we have applied in order to deal with
the imbalanced data-sets: the SMOTE
algorithm [7]. Finally, we will present the
evaluation metrics for this kind of
classification problem.
4.1. The problem of imbalanced data-sets
Learning from imbalanced data is an
important topic that has recently appeared in
the machine learning community.When
treating with imbalanced data-sets, one or
more classes might be represented by a large
number of examples whereas the others are
represented by only a few.We focus on the
binary-class imbalanced data-sets, where
there is only one positive and one negative
class. We consider the positive class as the
one with the lowest number of examples and
the negative class the one with the highest
number of examples. Furthermore, in this
work we use the IR, defined as the ratio of
the number of instances of the majority class
and the minority class, to organize the
different data-sets according to their IR.The
problem of imbalanced data-sets is
extremely significant because it is implicit in
most real world applications, such as fraud
detection [16], text classification, risk
management or medical applications.In
classification, this problem (also named the
‘‘class imbalance problem”) will cause a
bias on the training of classifiers
and will result in the lower sensitivity of
detecting the minority class examples. For
this reason, a large number of approaches
have been previously proposed to deal with
the class imbalance problem. These
approaches can be categorized into two
groups: the internal approaches that create
new algorithms or modify existing ones to
take the class imbalance problem into
consideration [3] and external approaches
that preprocess the data in order to diminish
the effect cause by their class imbalance
[4,15].The internal approaches have the
disadvantage of being algorithm specific,
whereas external approaches are
independentof the classifier used and are, for
this reason, more versatile. Furthermore, in
our previous work on this topic [18] we
analyzed the cooperation of some
preprocessing methods with FRBCSs,
showing a good behaviour for the over-
sampling methods,specially in the case of
the SMOTE methodology.According to this,
we will employ in this paper the SMOTE
algorithm in order to deal with the problem
of imbalanced data-sets. This method is
detailed in the next subsection.
4.2. Preprocessing imbalanced data-sets.
The SMOTE algorithm As mentioned
before, applying a preprocessing step in
order to balance the class distribution is a
positive solution to the imbalance data-set
problem [4]. Specifically, in this work we
have chosen an over-sampling method
which is a reference in this area: the
SMOTE algorithm [7].In this approach the
minority class is over-sampled by taking
each minority class sample and introducing
synthetic examples along the line segments
joining any/all of the k minority class
nearest neighbours. Depending upon the
amount of oversampling required,
neighbours from the k-nearest neighbours
are randomly chosen. This process is
illustrated in Fig. 1,where xi is the selected
point, xi1 to xi4 are some selected nearest
neighbours and r1 to r4 the synthetic data
points created by the randomized
interpolation. The implementation employed
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
96
6. in this work uses only one nearest neighbour
using the euclidean distance, and balance
both classes to the 50%
distribution.Synthetic samples are generated
in the following way: take the difference
between the feature vector (sample) under
consideration and its nearest neighbour.
Multiply this difference by a random
number between 0 and 1, and add it to the
feature vector under consideration. This
causes the selection of a random point along
the line segment between two specific
features.This approach effectively forces the
decision region of the minority class to
become more general. An example is
detailed in Fig. 2.In short, its main idea is to
form new minority class examples by
interpolating between several minority class
examples that lie together. Thus, the
overfitting problem is avoided and causes
the decision boundaries for the minority
class to spread further into the majority class
space.
4.3. Evaluation in imbalanced domains
The measures of the quality of classification
are built from a confusion matrix (shown in
Table 1) which records correctly and
incorrectly recognized examples for each
class.The most used empirical measure,
accuracy (1), does not distinguish between
the number of correct labels of different
classes, which in the framework of
imbalanced problems may lead to erroneous
conclusions. For example a classifier that
obtains an accuracy of 90% in a data-set
with an IR value of 9, might not be accurate
if it does not cover correctly any minority
class instance.
Because of this, instead of using accuracy,
more correct metrics are considered. Two
common measures, sensitivity and
specificity (2,3), approximate the probability
of the positive (negative) label being true. In
other words, they assess the effectiveness
of the algorithm on a single class.
The metric used in this work is the
geometric mean of the true rates [3], which
can be defined as
Fig.1.An illustration on how to create the synthetic data points in
the SMOTE algorithm.
Fig.2.Example of SMOTE application
This metric attempts to maximize the
accuracy of each one of the two classes with
a good balance. It is a performance metric
that links both objectives.
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
97
7. Table.1. Confusion matrix for a two-class problem.
Class Positive
prediction
Negative
Prediction
Positive class True
positive(TP)
False
Negative(FN)
Negative class False
positive(FP)
True
Negative(TN)
5. Hierarchical rule base genetic rule
selection process
In the previous section we have mentioned
that an excessive number of rules may not
produce a good performance and it makes
difficult to understand the model behaviour.
We may find different types of rules in a
large fuzzy rule set: irrelevant rules, which
do not contain significant information;
redundant rules, whose actions are covered
by other rules; erroneous rules, which are
wrong defined and distort the performance
of the FRBCS; and conflicting rules, which
perturb the performance of the FRBCS when
they coexist with others.In this work, we
consider the CHC genetic model [14] in
order to make the rule selection process,
since it has achieved good results for binary
selection problems [6]. In the following, the
main characteristics of this genetic approach
are presented.
1. Coding scheme and initial gene pool: It is
based on a binary coded GA where each
gene indicates whether a rule is selected or
not (alleles ‘1’ or ‘0’, respectively).
Considering that N rules are contained in the
preliminary/candidate rule set, the
chromosome C = (c1, . . . ,cN) represents a
subset of rules composing the final HRB,
such that:
with Ri being the corresponding ith rule in
the candidate rule set and HRB being the
final hierarchical rule base.The initial
pool is obtained with an individual having
all genes with value ‘1’ and the remaining
individuals generated at random in {0, 1}, so
that the initial HRB is taking into account in
the genetic selection process.
2. Chromosome evaluation: The fitness
function must be in accordance with the
framework of imbalanced data-sets. Thus,
we will use, as presented in Section 2.3, the
geometric mean of the true rates, defined in
(4) as:
3. Crossover operator: The half uniform
crossover scheme (HUX) is employed. In
this approach, the two parents are combined
to produce two new offspring. The
individual bits in the string are compared
between the two parents and exactly half of
the non-matching bits are swapped. Thus the
Hamming distance (the number of differing
bits) is first calculated.This number is
divided by two. The resulting number is how
many of the bits that do not match between
the two parents will be swapped.
4. Restarting approach: To get away from
local optima, this algorithm uses a restart
approach. In this case, the best chromosome
is maintained and the remaining are
generated at random in {1,0}. The restart
procedure is applied when a threshold value
is reached, which means that all the
individuals coexisting in the population are
very similar.
5. Evolutionary model: The CHC genetic
model makes use of a ‘‘Population-based
Selection” approach. N parents and their
corresponding offspring are combined to
select the best N individuals to take part of
the next population. The CHC approach
makes use of an incest prevention
mechanism and a restarting process to
provoke diversity in the population,instead
of the well-known mutation operator.
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
98
8. This incest prevention mechanism will be
considered in order to apply the HUX
operator, i.e., two parents are crossed if their
hamming distance divided by 2 is higher
than a predetermined threshold, L. The
threshold value is initialized as:
L = (#Genes/4.0). Following the original
CHC scheme, L is decremented by one
when the population does not change in
one generation. The algorithm restarts when
L is below zero. We will stop the genetic
process if more than 3 restarts are performed
without including any new chromosome in
the population.
6. CONCLUSION
In this paper, we have proposed an HFRBCS
approach for classification with imbalanced
data-sets. Our aim was to employ a
hierarchical model to obtain a good balance
among different granularity levels. A fine
granularity is applied in the boundary areas,
and a thick granularity may be applied in the
rest of the classification space providing a
good generalization. Thus,this approach
enhances the classification performance in
the overlapping areas between the minority
and majority classes.Furthermore, we have
made use of the SMOTE algorithm in order
to balance the training data before the rule
learning generation phase. This
preprocessing step enables the obtention of
better fuzzy rules than using the original
data-sets and therefore, we improve the
global performance of the fuzzy model to
filter out unwanted messages from Online
Social Networking (OSN).
7.REFERENCES
[1] R. Alcalá, J. Alcalá-Fdez, F. Herrera, J.
Otero, Genetic learning of accurate and
compact fuzzy rule based systems based on
the 2-tuples linguistic representation,
International Journal of Approximate
Reasoning 44 (2007) 4564.
[2] A. Asuncion, D. Newman, 2007. UCI
machine learning repository. University of
California, Irvine, School of Information
and Computer Sciences. URL:
<http://www.ics.uci.edu/~mlearn/MLReposi
tory.html>.
[3] R. Barandela, J.S. Sánchez, V. García, E.
Rangel, Strategies for learning in class
imbalance problems, Pattern Recognition 36
(3) (2003) 849–851.
[4] G.E.A.P.A. Batista, R.C. Prati, M.C.
Monard, A study of the behaviour of several
methods for balancing machine learning
training data, SIGKDD Explorations 6 (1)
(2004) 20–29.
[5] P. Campadelli, E. Casiraghi, G.
Valentini, Support vector machines for
candidate nodules classification, Letters on
Neurocomputing 68 (2005) 281–288.
[6] J.R. Cano, F. Herrera, M. Lozano, Using
evolutionary algorithms as instance selection
for data reduction in kdd: an experimental
study, IEEE Transactions on Evolutionary
Computation 7 (6) (2003) 561–575.
[7] N.V. Chawla, K.W. Bowyer, L.O. Hall,
W.P. Kegelmeyer, Smote: synthetic
minority over-sampling technique, Journal
of Artificial Intelligent Research 16 (2002)
321–357.
[8] N.V. Chawla, N. Japkowicz, A. Kolcz,
Editorial: special issue on learning from
imbalanced data-sets, SIGKDD Explorations
6 (1) (2004) 1–6.
[9] Z. Chi, H. Yan, T. Pham, Fuzzy
algorithms with applications to image
processing and pattern recognition, World
Scientific, 1996.
[10] J.-N. Choi, S.-K. Oh, W. Pedrycz,
Structural and parametric design of fuzzy
inference systems using hierarchical fair
competition-based parallel genetic
algorithms and information granulation,
International Journal of Approximate
Reasoning 49 (3) (2008) 631–648.
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
99
9. [11] O. Cordón, M.J. del Jesus, F. Herrera,
A proposal on reasoning methods in fuzzy
rule-based classification systems,
International Journal of Approximate
Reasoning 20 (1) (1999) 21–45.
[12] O. Cordón, F. Herrera, I. Zwir,
Linguistic modeling by hierarchical systems
of linguistic rules, IEEE Transactions on
Fuzzy Systems 10 (1) (2002) 2–20.
[13] J. Demšar, Statistical comparisons of
classifiers over multiple data-sets, Journal of
Machine Learning Research 7 (2006) 1–30.
[14] L.J. Eshelman, 1991. Foundations of
Genetic Algorithms. Morgan Kaufman, Ch.
The CHC Adaptive Search Algorithm: How
to have Safe Search when Engaging in
Nontraditional Genetic Recombination, pp.
265–283.
[15] A. Estabrooks, T. Jo, N. Japkowicz, A
multiple resampling method for learning
from imbalanced data-sets, Computational
Intelligence 20 (1) (2004) 18–36.
[16] T. Fawcett, F.J. Provost, Adaptive fraud
detection, Data Mining and Knowledge
Discovery 1 (3) (1997) 291–316.
[17] A. Fernández, S. García, M.J. del Jesus,
F. Herrera, An analysis of the rule weights
and fuzzy reasoning methods for linguistic
rule based classification systems applied to
problems with highly imbalanced data-sets,
in: International Workshop on Fuzzy Logic
and Applications (WILF07), Lecture Notes
on Computer Science, vol. 4578, Springer-
Verlag, 2007, pp. 170–179.
[18] A. Fernández, S. García, M.J. del Jesus,
F. Herrera, A study of the behaviour of
linguistic fuzzy rule based classification
systems in the framework of imbalanced
data-sets, Fuzzy Sets and Systems 159 (18)
(2008) 2378–2398.
[19] M. Friedman, The use of ranks to avoid
the assumption of normality implicit in the
analysis of variance, Journal of the
American Statistical Association 32 (1937)
675–701.
[20] S. García, D. Molina, M. Lozano, F.
Herrera, A study on the use of non-
parametric tests for analyzing the
evolutionary algorithms’ behaviour: a case
study on the CEC’2005 special session on
real parameter optimization. Journal of
Heuristics, in press, doi: 10.1007/s10732-
008-9080-4.
INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ENGINEERING RESEARCH, ICDER - 2014
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT www.iaetsd.in
100