The quick access to information on social media networks as well as its exponential rise also made it
difficult to distinguish among fake information or real information. The fast dissemination by way of
sharing has enhanced its falsification exponentially. It is also important for the credibility of social media
networks to avoid the spread of fake information. So it is emerging research challenge to automatically
check for misstatement of information through its source, content, or publisher and prevent the
unauthenticated sources from spreading rumours. This paper demonstrates an artificial intelligence based
approach for the identification of the false statements made by social network entities. Two variants of
Deep neural networks are being applied to evalues datasets and analyse for fake news presence. The
implementation setup produced maximum extent 99% classification accuracy, when dataset is tested for
binary (true or false) labeling with multiple epochs.
FAKE NEWS DETECTION WITH SEMANTIC FEATURES AND TEXT MININGijnlc
Nearly 70% of people are concerned about the propagation of fake news. This paper aims to detect fake news in online articles through the use of semantic features and various machine learning techniques. In this research, we investigated recurrent neural networks vs. the naive bayes classifier and random forest classifiers using five groups of linguistic features. Evaluated with real or fake dataset from kaggle.com, the best performing model achieved an accuracy of 95.66% using bigram features with the random forest classifier. The fact that bigrams outperform unigrams, trigrams, and quadgrams show that word pairs as opposed to single words or phrases best indicate the authenticity of news.
Recently, fake news has been incurring many problems to our society. As a result, many researchers have been working on identifying fake news. Most of the fake news detection systems utilize the linguistic feature of the news. However, they have difficulty in sensing highly ambiguous fake news which can be detected only after identifying meaning and latest related information. In this paper, to resolve this problem, we shall present a new Korean fake news detection system using fact DB which is built and updated by human's direct judgement after collecting obvious facts. Our system receives a proposition, and search the semantically related articles from Fact DB in order to verify whether the given proposition is true or not by comparing the proposition with the related articles in fact DB. To achieve this, we utilize a deep learning model, Bidirectional Multi Perspective Matching for Natural Language Sentence BiMPM , which has demonstrated a good performance for the sentence matching task. However, BiMPM has some limitations in that the longer the length of the input sentence is, the lower its performance is, and it has difficulty in making an accurate judgement when an unlearned word or relation between words appear. In order to overcome the limitations, we shall propose a new matching technique which exploits article abstraction as well as entity matching set in addition to BiMPM. In our experiment, we shall show that our system improves the whole performance for fake news detection. Prasanth. K | Praveen. N | Vijay. S | Auxilia Osvin Nancy. V ""Fake News Detection using Machine Learning"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-2 , February 2020,
URL: https://www.ijtsrd.com/papers/ijtsrd30014.pdf
Paper Url : https://www.ijtsrd.com/engineering/information-technology/30014/fake-news-detection-using-machine-learning/prasanth-k
FAKE NEWS DETECTION WITH SEMANTIC FEATURES AND TEXT MININGijnlc
Nearly 70% of people are concerned about the propagation of fake news. This paper aims to detect fake news in online articles through the use of semantic features and various machine learning techniques. In this research, we investigated recurrent neural networks vs. the naive bayes classifier and random forest classifiers using five groups of linguistic features. Evaluated with real or fake dataset from kaggle.com, the best performing model achieved an accuracy of 95.66% using bigram features with the random forest classifier. The fact that bigrams outperform unigrams, trigrams, and quadgrams show that word pairs as opposed to single words or phrases best indicate the authenticity of news.
Recently, fake news has been incurring many problems to our society. As a result, many researchers have been working on identifying fake news. Most of the fake news detection systems utilize the linguistic feature of the news. However, they have difficulty in sensing highly ambiguous fake news which can be detected only after identifying meaning and latest related information. In this paper, to resolve this problem, we shall present a new Korean fake news detection system using fact DB which is built and updated by human's direct judgement after collecting obvious facts. Our system receives a proposition, and search the semantically related articles from Fact DB in order to verify whether the given proposition is true or not by comparing the proposition with the related articles in fact DB. To achieve this, we utilize a deep learning model, Bidirectional Multi Perspective Matching for Natural Language Sentence BiMPM , which has demonstrated a good performance for the sentence matching task. However, BiMPM has some limitations in that the longer the length of the input sentence is, the lower its performance is, and it has difficulty in making an accurate judgement when an unlearned word or relation between words appear. In order to overcome the limitations, we shall propose a new matching technique which exploits article abstraction as well as entity matching set in addition to BiMPM. In our experiment, we shall show that our system improves the whole performance for fake news detection. Prasanth. K | Praveen. N | Vijay. S | Auxilia Osvin Nancy. V ""Fake News Detection using Machine Learning"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-2 , February 2020,
URL: https://www.ijtsrd.com/papers/ijtsrd30014.pdf
Paper Url : https://www.ijtsrd.com/engineering/information-technology/30014/fake-news-detection-using-machine-learning/prasanth-k
Research of usability of Mashup Tools done for Kent County Council as part of the Pic and Mix Pilot (2009), opening up Kent related datasets for all to use and exploit.
DETECTION OF FAKE ACCOUNTS IN INSTAGRAM USING MACHINE LEARNINGijcsit
With the advent of the Internet and social media, while hundreds of people have benefitted from the vast sources of information available, there has been an enormous increase in the rise of cyber-crimes, particularly targeted towards women. According to a 2019 report in the [4] Economics Times, India has witnessed a 457% rise in cybercrime in the five year span between 2011 and 2016. Most speculate that this is due to impact of social media such as Facebook, Instagram and Twitter on our daily lives. While these definitely help in creating a sound social network, creation of user accounts in these sites usually needs just an email-id. A real life person can create multiple fake IDs and hence impostors can easily be made. Unlike the real world scenario where multiple rules and regulations are imposed to identify oneself in a unique manner (for example while issuing one’s passport or driver’s license), in the virtual world of social media, admission does not require any such checks. In this paper, we study the different accounts of Instagram, in particular and try to assess an account as fake or real using Machine Learning techniques namely Logistic Regression and Random Forest Algorithm.
Hate speech has been an ongoing problem on the Internet for many years. Besides, social media, especially Facebook, and Twitter have given it a global stage where those hate speeches can spread far more rapidly. Every social media platform needs to implement an effective hate speech detection system to remove offensive content in real-time. There are various approaches to identify hate speech, such as Rule-Based, Machine Learning based, deep learning based and Hybrid approach. Since this is a review paper, we explained the valuable works of various authors who have invested their valuable time in studying to identifying hate speech using various approaches.
CATEGORIZING 2019-N-COV TWITTER HASHTAG DATA BY CLUSTERINGijaia
Unsupervised machine learning techniques such as clustering are widely gaining use with the recent increase in social communication platforms like Twitter and Facebook. Clustering enables the finding of patterns in these unstructured datasets. We collected tweets matching hashtags linked to COVID-19 from a Kaggle dataset. We compared the performance of nine clustering algorithms using this dataset. We evaluated the generalizability of these algorithms using a supervised learning model. Finally, using a selected unsupervised learning algorithm we categorized the clusters. The top five categories are Safety,
Crime, Products, Countries and Health. This can prove helpful for bodies using large amount of Twitter data needing to quickly find key points in the data before going into further classification.
Poster presentation in 3rd big data conclave at vit chennai on 20th april 2017Rohit Desai
Title:Leveraging Big Data and Social Sensors For Predicting Epidemic Disease Outbreak
Event & Venue:3rd Big Data Conclave at VIT Chennai on 20th & 21st April 2017
Project Conducted Under Guide:Dr.Sweetlin Hemalatha Professor at VIT Chennai
On Semantics and Deep Learning for Event Detection in Crisis SituationsCOMRADES project
In this paper, we introduce Dual-CNN, a semantically-enhanced deep learning model to target the problem of event detection in crisis situations from
social media data. A layer of semantics is added to a traditional Convolutional Neural Network (CNN) model to capture the contextual information that is generally scarce in short, ill-formed social media messages. Our results show that
our methods are able to successfully identify the existence of events, and event types (hurricane, floods, etc.) accurately (> 79% F-measure), but the performance of the model significantly drops (61% F-measure) when identifying fine-grained event-related information (affected individuals, damaged infrastructures, etc.).
These results are competitive with more traditional Machine Learning models, such as SVM.
http://oro.open.ac.uk/49639/1/event_detection.pdf
ANALYSIS OF TOPIC MODELING WITH UNPOOLED AND POOLED TWEETS AND EXPLORATION OF...IJCSEA Journal
In this digital era, social media is an important tool for information dissemination. Twitter is a popular social media platform. Social media analytics helps make informed decisions based on people's needs and opinions. This information, when properly perceived provides valuable insights into different domains, such as public policymaking, marketing, sales, and healthcare. Topic modeling is an unsupervised algorithm to discover a hidden pattern in text documents. In this study, we explore the Latent Dirichlet Allocation (LDA) topic model algorithm. We collected tweets with hashtags related to corona virus related discussions. This study compares regular LDA and LDA based on collapsed Gibbs sampling (LDAMallet) algorithms. The experiments use different data processing steps including trigrams, without trigrams, hashtags, and without hashtags. This study provides a comprehensive analysis of LDA for short text messages using un-pooled and pooled tweets. The results suggest that a pooling scheme using hashtags helps improve the topic inference results with a better coherence score.
In the age of social media communication, it is easy to
modulate the minds of users and also instigate violent
actions being taken by them in some cases. There is a need
to have a system that can analyze the threat level of tweets
from influential users and rank their Twitter handles so
that dangerous tweets can be avoided going public on
Twitter before fact-checking which can hurt the sentiments
of people and can take the shape of violence. The study
aims to analyse and rank twitter users according to their
influential power and extremism of their tweets to help
prevent major protests and violent events. We scraped top
trending topics and fetched tweets using those hashtags.
We propose a custom ranking algorithm which considers
source based and content based features along with a
knowledge graph which generates the score and rank the
twitter users according to the scores. Our aim with this
study is to identify and rank extremist twitter users with
regards to their impact and influence. We use a technique
that takes into consideration both source based and
content-based features of tweets to generate the ranking of
the extremist twitter users having a high impact factor
Sentiment Analysis is the process of finding the sentiments from different classes of words.
Generally speaking, sentiment analysis aims to determine the attitude of a speaker or a writer with
respect to some topic or the overall contextual polarity of a document. The attitude may be his or
her judgment or evaluation, affective state, or the intended emotional communication. In this case,
‘tweets’! Given a micro-blogging platform where official, verified tweets are available to us, we
need to identify the sentiments of those tweets. A model must be constructed where the sentiments
are scored, for each product individually and then they are compared with, diagrammatically,
portraying users’ feedback from the producers stand point.
There are many websites that offer a comparison between various products or services based on
certain features of the article such as its predominant traits, price, and its welcome in the market and
so on. However not many provide a juxtaposing of commodities with user review as the focal point.
Those few that do work with Naïve Bayes Machine Learning Algorithms, that poses a disadvantage
as it mandatorily assumes that the features, in our project, words, are independent of each other.
This is a comparatively inefficient method of performing Sentiment Analysis on bulk text, for
official purposes, since sentences will not give the meaning they are supposed to convey, if each
word is considered a separate entity. Maximum Entropy Classifier overcomes this draw back by
limiting the assumptions it makes of the input data feed, which is what we use in the proposed
system.
Hoax analyzer for Indonesian news using RNNs with fasttext and glove embeddingsjournalBEEI
Misinformation has become an innocuous yet potentially harmful problem ever since the development of internet. Numbers of efforts are done to prevent the consumption of misinformation, including the use of artificial intelligence (AI), mainly natural language processing (NLP). Unfortunately, most of natural language processing use English as its linguistic approach since English is a high resource language. On the contrary, Indonesia language is considered a low resource language thus the amount of effort to diminish consumption of misinformation is low compared to English-based natural language processing. This experiment is intended to compare fastText and GloVe embeddings for four deep neural networks (DNN) models: long short-term memory (LSTM), bidirectional long short-term memory (BI-LSTM), gated recurrent unit (GRU) and bidirectional gated recurrent unit (BI-GRU) in terms of metrics score when classifying news between three classes: fake, valid, and satire. The latter results show that fastText embedding is better than GloVe embedding in supervised text classification, along with BI-GRU + fastText yielding the best result.
Political prediction analysis using text mining and deep learningVishwambhar Deshpande
We have proposed a system to determine current sentiment on twitter using Twit-
ter API for open access which includes opinions from dierent content structures like
latest news, audits, articles and social media posts. and Deep Learning method to
study Historic Data for predicting future results. we utilized Naive Bayes and dictio-
nary based algorithms to predict the sentiment on Live Twitter Data.
How could a product or service is reasonably evaluated by anyone in the shortest time? A million dollar
question but it is having a simple answer: Sentiment analysis. Sentiment analysis is consumers review on
products and services which helps both the producers and consumers (stakeholders) to take effective and
efficient decision within a shortest period of time. Producers can have better knowledge of their products
and services through the sentiment analysis (ex. positive and negative comments or consumers likes and
dislikes) which will help them to know their products status (ex. product limitations or market status).
Consumers can have better knowledge of their interested products and services through the sentiment
analysis (ex. positive and negative comments or consumers likes and dislikes) which will help them to know
their deserving products status (ex. product limitations or market status). For more specification of the
sentiment values, fuzzy logic could be introduced. Therefore, sentiment analysis with the help of fuzzy logic
(deals with reasoning and gives closer views to the exact sentiment values) will help the producers or
consumers or any interested person for taking the effective decision according to their product or service
interest.
Fake News Detection on Social Media using Machine Learningclassic tpr
For some years, mostly since the rise of social media, fake news has become a society
problem, in some occasions spreading more and faster than the true information. Hence it is
very important to detect and reduce the involvement of fake news in social platforms. This
project comes up with the applications of NLP (Natural Language Processing) techniques for
detecting the 'fake news', that is, misleading news stories that come from the non-reputable
sources. Natural language processing (NLP) refers to the branch of computer science- and
more specifically, the branch of artificial intelligence or AI-concerned with giving computers
the ability to understand text and spoken words in much the same way human beings can.
NLP combines computational linguistics-rule based modelling of human language- with
statistical, machine learning and deep learning models. Together these technologies enable
computers to process human language in the form of text or voice data and to 'understand' its
full meaning, complete with the speaker or writer's intent and sentiment. Only by building a
model based on a count vectorizer (using word tallies) or a (Term Frequency Inverse
Document Frequency) tfidf matrix, can only get you so far. But these models do not consider
the important qualities like word ordering and context. It is very possible that two articles that
are similar in their word count will be completely different in their meaning. The data science
community has responded by taking actions against the problem.
Research of usability of Mashup Tools done for Kent County Council as part of the Pic and Mix Pilot (2009), opening up Kent related datasets for all to use and exploit.
DETECTION OF FAKE ACCOUNTS IN INSTAGRAM USING MACHINE LEARNINGijcsit
With the advent of the Internet and social media, while hundreds of people have benefitted from the vast sources of information available, there has been an enormous increase in the rise of cyber-crimes, particularly targeted towards women. According to a 2019 report in the [4] Economics Times, India has witnessed a 457% rise in cybercrime in the five year span between 2011 and 2016. Most speculate that this is due to impact of social media such as Facebook, Instagram and Twitter on our daily lives. While these definitely help in creating a sound social network, creation of user accounts in these sites usually needs just an email-id. A real life person can create multiple fake IDs and hence impostors can easily be made. Unlike the real world scenario where multiple rules and regulations are imposed to identify oneself in a unique manner (for example while issuing one’s passport or driver’s license), in the virtual world of social media, admission does not require any such checks. In this paper, we study the different accounts of Instagram, in particular and try to assess an account as fake or real using Machine Learning techniques namely Logistic Regression and Random Forest Algorithm.
Hate speech has been an ongoing problem on the Internet for many years. Besides, social media, especially Facebook, and Twitter have given it a global stage where those hate speeches can spread far more rapidly. Every social media platform needs to implement an effective hate speech detection system to remove offensive content in real-time. There are various approaches to identify hate speech, such as Rule-Based, Machine Learning based, deep learning based and Hybrid approach. Since this is a review paper, we explained the valuable works of various authors who have invested their valuable time in studying to identifying hate speech using various approaches.
CATEGORIZING 2019-N-COV TWITTER HASHTAG DATA BY CLUSTERINGijaia
Unsupervised machine learning techniques such as clustering are widely gaining use with the recent increase in social communication platforms like Twitter and Facebook. Clustering enables the finding of patterns in these unstructured datasets. We collected tweets matching hashtags linked to COVID-19 from a Kaggle dataset. We compared the performance of nine clustering algorithms using this dataset. We evaluated the generalizability of these algorithms using a supervised learning model. Finally, using a selected unsupervised learning algorithm we categorized the clusters. The top five categories are Safety,
Crime, Products, Countries and Health. This can prove helpful for bodies using large amount of Twitter data needing to quickly find key points in the data before going into further classification.
Poster presentation in 3rd big data conclave at vit chennai on 20th april 2017Rohit Desai
Title:Leveraging Big Data and Social Sensors For Predicting Epidemic Disease Outbreak
Event & Venue:3rd Big Data Conclave at VIT Chennai on 20th & 21st April 2017
Project Conducted Under Guide:Dr.Sweetlin Hemalatha Professor at VIT Chennai
On Semantics and Deep Learning for Event Detection in Crisis SituationsCOMRADES project
In this paper, we introduce Dual-CNN, a semantically-enhanced deep learning model to target the problem of event detection in crisis situations from
social media data. A layer of semantics is added to a traditional Convolutional Neural Network (CNN) model to capture the contextual information that is generally scarce in short, ill-formed social media messages. Our results show that
our methods are able to successfully identify the existence of events, and event types (hurricane, floods, etc.) accurately (> 79% F-measure), but the performance of the model significantly drops (61% F-measure) when identifying fine-grained event-related information (affected individuals, damaged infrastructures, etc.).
These results are competitive with more traditional Machine Learning models, such as SVM.
http://oro.open.ac.uk/49639/1/event_detection.pdf
ANALYSIS OF TOPIC MODELING WITH UNPOOLED AND POOLED TWEETS AND EXPLORATION OF...IJCSEA Journal
In this digital era, social media is an important tool for information dissemination. Twitter is a popular social media platform. Social media analytics helps make informed decisions based on people's needs and opinions. This information, when properly perceived provides valuable insights into different domains, such as public policymaking, marketing, sales, and healthcare. Topic modeling is an unsupervised algorithm to discover a hidden pattern in text documents. In this study, we explore the Latent Dirichlet Allocation (LDA) topic model algorithm. We collected tweets with hashtags related to corona virus related discussions. This study compares regular LDA and LDA based on collapsed Gibbs sampling (LDAMallet) algorithms. The experiments use different data processing steps including trigrams, without trigrams, hashtags, and without hashtags. This study provides a comprehensive analysis of LDA for short text messages using un-pooled and pooled tweets. The results suggest that a pooling scheme using hashtags helps improve the topic inference results with a better coherence score.
In the age of social media communication, it is easy to
modulate the minds of users and also instigate violent
actions being taken by them in some cases. There is a need
to have a system that can analyze the threat level of tweets
from influential users and rank their Twitter handles so
that dangerous tweets can be avoided going public on
Twitter before fact-checking which can hurt the sentiments
of people and can take the shape of violence. The study
aims to analyse and rank twitter users according to their
influential power and extremism of their tweets to help
prevent major protests and violent events. We scraped top
trending topics and fetched tweets using those hashtags.
We propose a custom ranking algorithm which considers
source based and content based features along with a
knowledge graph which generates the score and rank the
twitter users according to the scores. Our aim with this
study is to identify and rank extremist twitter users with
regards to their impact and influence. We use a technique
that takes into consideration both source based and
content-based features of tweets to generate the ranking of
the extremist twitter users having a high impact factor
Sentiment Analysis is the process of finding the sentiments from different classes of words.
Generally speaking, sentiment analysis aims to determine the attitude of a speaker or a writer with
respect to some topic or the overall contextual polarity of a document. The attitude may be his or
her judgment or evaluation, affective state, or the intended emotional communication. In this case,
‘tweets’! Given a micro-blogging platform where official, verified tweets are available to us, we
need to identify the sentiments of those tweets. A model must be constructed where the sentiments
are scored, for each product individually and then they are compared with, diagrammatically,
portraying users’ feedback from the producers stand point.
There are many websites that offer a comparison between various products or services based on
certain features of the article such as its predominant traits, price, and its welcome in the market and
so on. However not many provide a juxtaposing of commodities with user review as the focal point.
Those few that do work with Naïve Bayes Machine Learning Algorithms, that poses a disadvantage
as it mandatorily assumes that the features, in our project, words, are independent of each other.
This is a comparatively inefficient method of performing Sentiment Analysis on bulk text, for
official purposes, since sentences will not give the meaning they are supposed to convey, if each
word is considered a separate entity. Maximum Entropy Classifier overcomes this draw back by
limiting the assumptions it makes of the input data feed, which is what we use in the proposed
system.
Hoax analyzer for Indonesian news using RNNs with fasttext and glove embeddingsjournalBEEI
Misinformation has become an innocuous yet potentially harmful problem ever since the development of internet. Numbers of efforts are done to prevent the consumption of misinformation, including the use of artificial intelligence (AI), mainly natural language processing (NLP). Unfortunately, most of natural language processing use English as its linguistic approach since English is a high resource language. On the contrary, Indonesia language is considered a low resource language thus the amount of effort to diminish consumption of misinformation is low compared to English-based natural language processing. This experiment is intended to compare fastText and GloVe embeddings for four deep neural networks (DNN) models: long short-term memory (LSTM), bidirectional long short-term memory (BI-LSTM), gated recurrent unit (GRU) and bidirectional gated recurrent unit (BI-GRU) in terms of metrics score when classifying news between three classes: fake, valid, and satire. The latter results show that fastText embedding is better than GloVe embedding in supervised text classification, along with BI-GRU + fastText yielding the best result.
Political prediction analysis using text mining and deep learningVishwambhar Deshpande
We have proposed a system to determine current sentiment on twitter using Twit-
ter API for open access which includes opinions from dierent content structures like
latest news, audits, articles and social media posts. and Deep Learning method to
study Historic Data for predicting future results. we utilized Naive Bayes and dictio-
nary based algorithms to predict the sentiment on Live Twitter Data.
How could a product or service is reasonably evaluated by anyone in the shortest time? A million dollar
question but it is having a simple answer: Sentiment analysis. Sentiment analysis is consumers review on
products and services which helps both the producers and consumers (stakeholders) to take effective and
efficient decision within a shortest period of time. Producers can have better knowledge of their products
and services through the sentiment analysis (ex. positive and negative comments or consumers likes and
dislikes) which will help them to know their products status (ex. product limitations or market status).
Consumers can have better knowledge of their interested products and services through the sentiment
analysis (ex. positive and negative comments or consumers likes and dislikes) which will help them to know
their deserving products status (ex. product limitations or market status). For more specification of the
sentiment values, fuzzy logic could be introduced. Therefore, sentiment analysis with the help of fuzzy logic
(deals with reasoning and gives closer views to the exact sentiment values) will help the producers or
consumers or any interested person for taking the effective decision according to their product or service
interest.
Fake News Detection on Social Media using Machine Learningclassic tpr
For some years, mostly since the rise of social media, fake news has become a society
problem, in some occasions spreading more and faster than the true information. Hence it is
very important to detect and reduce the involvement of fake news in social platforms. This
project comes up with the applications of NLP (Natural Language Processing) techniques for
detecting the 'fake news', that is, misleading news stories that come from the non-reputable
sources. Natural language processing (NLP) refers to the branch of computer science- and
more specifically, the branch of artificial intelligence or AI-concerned with giving computers
the ability to understand text and spoken words in much the same way human beings can.
NLP combines computational linguistics-rule based modelling of human language- with
statistical, machine learning and deep learning models. Together these technologies enable
computers to process human language in the form of text or voice data and to 'understand' its
full meaning, complete with the speaker or writer's intent and sentiment. Only by building a
model based on a count vectorizer (using word tallies) or a (Term Frequency Inverse
Document Frequency) tfidf matrix, can only get you so far. But these models do not consider
the important qualities like word ordering and context. It is very possible that two articles that
are similar in their word count will be completely different in their meaning. The data science
community has responded by taking actions against the problem.
During the last decade, the social media has been regarded as a rich dominant source of information and news. Its unsupervised nature leads to the emergence and spread of fake news. Fake news detection has gained a great importance posing many challenges to the research community. One of the main challenges is the detection accuracy which is highly affected by the chosen and extracted features and the used classification algorithm. In this paper, we propose a context-based solution that relies on account features and random forest classifier to detect fake news. It achieves the precision of 99.8%. The system accuracy has been compared to other commonly used classifiers such as decision tree classifier, Gaussian Naïve Bayes and neural network which give precision of 98.4%, 92.6%, and 62.7% respectively. The experiments’ accuracy results show the possibility of distinguishing fake news and giving credibility scores for social media news with a relatively high performance.
A Machine Learning Ensemble Model for the Detection of Cyberbullyinggerogepatton
The pervasive use of social media platforms, such as Facebook, Instagram, and X, has significantly amplified
our electronic interconnectedness. Moreover, these platforms are now easily accessible from any location at
any given time. However, the increased popularity of social media has also led to cyberbullying.It is imperative
to address the need for finding, monitoring, and mitigating cyberbullying posts on social media platforms.
Motivated by this necessity, we present this paper to contribute to developing an automated system for
detecting binary labels of aggressive tweets.Our study has demonstrated remarkable performance compared to
previous experiments on the same dataset. We employed the stacking ensemble machine learning method,
utilizing four various feature extraction techniques to optimize performance within the stacking ensemble
learning framework. Combining five machine learning algorithms,Decision Trees, Random Forest, Linear
Support Vector Classification, Logistic Regression, and K-Nearest Neighbors into an ensemble method, we
achieved superior results compared to traditional machine learning classifier models. The stacking classifier
achieved a high accuracy rate of 94.00%, outperforming traditional machine learning models and surpassing
the results of prior experiments that utilized the same dataset. The outcomes of our experiments showcased an
accuracy rate of 0.94% in detection tweets as aggressive or non-aggressive.
A MACHINE LEARNING ENSEMBLE MODEL FOR THE DETECTION OF CYBERBULLYINGijaia
The pervasive use of social media platforms, such as Facebook, Instagram, and X, has significantly amplified
our electronic interconnectedness. Moreover, these platforms are now easily accessible from any location at
any given time. However, the increased popularity of social media has also led to cyberbullying.It is imperative
to address the need for finding, monitoring, and mitigating cyberbullying posts on social media platforms.
Motivated by this necessity, we present this paper to contribute to developing an automated system for
detecting binary labels of aggressive tweets.Our study has demonstrated remarkable performance compared to
previous experiments on the same dataset. We employed the stacking ensemble machine learning method,
utilizing four various feature extraction techniques to optimize performance within the stacking ensemble
learning framework. Combining five machine learning algorithms,Decision Trees, Random Forest, Linear
Support Vector Classification, Logistic Regression, and K-Nearest Neighbors into an ensemble method, we
achieved superior results compared to traditional machine learning classifier models. The stacking classifier
achieved a high accuracy rate of 94.00%, outperforming traditional machine learning models and surpassing
the results of prior experiments that utilized the same dataset. The outcomes of our experiments showcased an
accuracy rate of 0.94% in detection tweets as aggressive or non-aggressive.
A Machine Learning Ensemble Model for the Detection of Cyberbullyinggerogepatton
The pervasive use of social media platforms, such as Facebook, Instagram, and X, has significantly amplified
our electronic interconnectedness. Moreover, these platforms are now easily accessible from any location at
any given time. However, the increased popularity of social media has also led to cyberbullying.It is imperative
to address the need for finding, monitoring, and mitigating cyberbullying posts on social media platforms.
Motivated by this necessity, we present this paper to contribute to developing an automated system for
detecting binary labels of aggressive tweets.Our study has demonstrated remarkable performance compared to
previous experiments on the same dataset. We employed the stacking ensemble machine learning method,
utilizing four various feature extraction techniques to optimize performance within the stacking ensemble
learning framework. Combining five machine learning algorithms,Decision Trees, Random Forest, Linear
Support Vector Classification, Logistic Regression, and K-Nearest Neighbors into an ensemble method, we
achieved superior results compared to traditional machine learning classifier models. The stacking classifier
achieved a high accuracy rate of 94.00%, outperforming traditional machine learning models and surpassing
the results of prior experiments that utilized the same dataset. The outcomes of our experiments showcased an
accuracy rate of 0.94% in detection tweets as aggressive or non-aggressive.
Combating propaganda texts using transfer learningIAESIJAI
Recently, it has been observed that people are shifting away from traditional news media sources towards trusting social networks to gather news information. Social networks have become the primary news source, although the validity and reliability of the information provided are uncertain. Memes are crucial content types that are very popular among young people and play a vital role in social media. It spreads quickly and continues to spread rapidly among people in a peer-to-peer manner rather than a prescriptive. Unfortunately, promoters and propagandists have adopted memes to indirectly manipulate public opinion and influence their attitudes using psychological and rhetorical techniques. This type of content could lead to unpleasant consequences in communities. This paper introduces an ensemble model system that resolves one of the most recent natural language processing research topics; propaganda techniques detection in texts extracted from memes. The paper also explores state-of-the-art pre-trained language models. The proposed model also uses different optimization techniques, such as data augmentation and model ensemble. It has been evaluated using a reference dataset from SemEval-2021 task 6. Our system outperforms the baseline and state-of-the-art results by achieving an F1-micro score of 0.604% on the test set.
ANALYZING AND IDENTIFYING FAKE NEWS USING ARTIFICIAL INTELLIGENCEIAEME Publication
The main reason behind the spread of fake news is because of many fake and hyperpartisan sites present on the Internet. These fake sites try to manipulate the truth which creates misunderstanding in society. Therefore, it is important to detect fake news and try to make people aware of the truth. This paper gives an insight into how to detect fake news using Machine Learning and Deep Learning Techniques. On observing our data, we have categorized our data into five attributes namely Title, Text, Subject, Date, and Labels. In order to develop an efficient fake news detection system, the feature along with its degree of impact on the system must be taken into consideration. This paper attempts at providing a detailed analysis of detecting fake news using various models such as LSTM, ANN, Naïve Bayes, SVM, Logistic Regression, XGBoost, and Bert.
How does fakenews spread understanding pathways of disinformation spread thro...Araz Taeihagh
What are the pathways for spreading disinformation on social media platforms? This article addresses this question by collecting, categorising, and situating an extensive body of research on how application programming interfaces (APIs) provided by social media platforms facilitate the spread of disinformation. We first examine the landscape of official social media APIs, then perform quantitative research on the open-source code repositories GitHub and GitLab to understand the usage patterns of these APIs. By inspecting the code repositories, we classify developers' usage of the APIs as official and unofficial, and further develop a four-stage framework characterising pathways for spreading disinformation on social media platforms. We further highlight how the stages in the framework were activated during the 2016 US Presidential Elections, before providing policy recommendations for issues relating to access to APIs, algorithmic content, advertisements, and suggest rapid response to coordinate campaigns, development of collaborative, and participatory approaches as well as government stewardship in the regulation of social media platforms.
Applications of data science in social media.pptxlyudmilabaruah
Data science has become one of the fastest growing fields across industries. Data science plays a crucial role in social media. From personalized content recommendations to proactive spam detection, data science fuels a transformative era, shaping a more engaging, secure, and individualized social media landscape. The slides in this presentation therefore aim to help you in understanding the applications of data science in social media.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Automobile Management System Project Report.pdfKamal Acharya
The proposed project is developed to manage the automobile in the automobile dealer company. The main module in this project is login, automobile management, customer management, sales, complaints and reports. The first module is the login. The automobile showroom owner should login to the project for usage. The username and password are verified and if it is correct, next form opens. If the username and password are not correct, it shows the error message.
When a customer search for a automobile, if the automobile is available, they will be taken to a page that shows the details of the automobile including automobile name, automobile ID, quantity, price etc. “Automobile Management System” is useful for maintaining automobiles, customers effectively and hence helps for establishing good relation between customer and automobile organization. It contains various customized modules for effectively maintaining automobiles and stock information accurately and safely.
When the automobile is sold to the customer, stock will be reduced automatically. When a new purchase is made, stock will be increased automatically. While selecting automobiles for sale, the proposed software will automatically check for total number of available stock of that particular item, if the total stock of that particular item is less than 5, software will notify the user to purchase the particular item.
Also when the user tries to sale items which are not in stock, the system will prompt the user that the stock is not enough. Customers of this system can search for a automobile; can purchase a automobile easily by selecting fast. On the other hand the stock of automobiles can be maintained perfectly by the automobile shop manager overcoming the drawbacks of existing system.
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSEDuvanRamosGarzon1
AIRCRAFT GENERAL
The Single Aisle is the most advanced family aircraft in service today, with fly-by-wire flight controls.
The A318, A319, A320 and A321 are twin-engine subsonic medium range aircraft.
The family offers a choice of engines
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
Saudi Arabia stands as a titan in the global energy landscape, renowned for its abundant oil and gas resources. It's the largest exporter of petroleum and holds some of the world's most significant reserves. Let's delve into the top 10 oil and gas projects shaping Saudi Arabia's energy future in 2024.
Top 10 Oil and Gas Projects in Saudi Arabia 2024.pdf
A RELIABLE ARTIFICIAL INTELLIGENCE MODEL FOR FALSE NEWS DETECTION MADE BY PUBLIC FIGURES
1. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
DOI: 10.5121/caij.2021.8301 1
A RELIABLE ARTIFICIAL INTELLIGENCE MODEL
FOR FALSE NEWS DETECTION MADE BY
PUBLIC FIGURES
Prashant Kumar Shrivastava1
, Mayank Sharma2
, Megha Kamble3
, Vaibhav Gore4
1, 3,4
Computer Science & Engineering, Lakshmi Narain College of Technology
Excellence,
1,3,4
LNCTE Bhopal ,
2
ITM University ,Gwalior, India
1,2,3,4
Computer Science & Engineering
ABSTRACT
The quick access to information on social media networks as well as its exponential rise also made it
difficult to distinguish among fake information or real information. The fast dissemination by way of
sharing has enhanced its falsification exponentially. It is also important for the credibility of social media
networks to avoid the spread of fake information. So it is emerging research challenge to automatically
check for misstatement of information through its source, content, or publisher and prevent the
unauthenticated sources from spreading rumours. This paper demonstrates an artificial intelligence based
approach for the identification of the false statements made by social network entities. Two variants of
Deep neural networks are being applied to evalues datasets and analyse for fake news presence. The
implementation setup produced maximum extent 99% classification accuracy, when dataset is tested for
binary (true or false) labeling with multiple epochs.
KEYWORDS
fake news; artificial intelligence; deep learning: RNN, RAMP studio.
1. INTRODUCTION
The concept "fake news (FNs)" was much less unheard of and not popular in a few decades, but it
emerged as a major monster in this modern age of social media. Fake news, bubbles in
information, manipulation of news, or loss of interest in media are growing problems in our
culture. However, a clear understanding of FNs as well as its sources is important to begin to
resolve this issue. It is only in this context that one should look at the various methods & fields of
machine learning (ML), natural language processing (NLP), & artificial intelligence ( AI), which
could help us face the situation.
The calculation or even the proper definition of fake news could very easily become a subjective
issue instead of an objective metric. In its purest nature, fake news is entirely composed,
manipulated to imitate trustworthy media, and draws full exposure.
Motivation
The popular issue of fake news is very hard to resolve in today's digital world when thousands of
information sharing platforms will distribute FNs or incorrect data. A big issue in AI is that of
2. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
2
developing artificial bots to produce and distribute fake news. [1]. problem is tragic because
many people think they read something on the Internet, even those who are amateur or new to
modern technologies will easily be tricked. A related issue is a spam or fraudulent emails or
communications that may lead to theft. It is also reasonable enough to recognize this problem to
solve crime rates, political uneasiness, grief & combat the attempts at distributing false news.
The definition of fake news consists of two parts: authenticity & intent. Authenticity ensures that
false news material may be confirmed as such and that conspiracy is not found in fake news and,
in most situations, it is impossible to prove true or false. The second part, intent, means that the
false information is written to mislead the reader.
There are also tens of millions of languages spoken and each language has its vocabulary, script
& syntax. The NLP is an AI branch that involves strategies for using text, creating models, and
forecasting. The nature of outline text is of different forms and it is difficult to use simply for
different linguistic features or forms, like sarcasm, metaphors, etc.This paper aims to build a
framework or model that can utilize details in previous news stories to predict whether or not a
news report is false.
1.1. Fake News Types
The authors[2] have provided comprehensive literature and the types of fake news can be
summarized as follows, based on literaure.
1. Visual-based: These fake news items are much more useful as material, including
fraudulent images, medical videos, or both [3].
2. User-based: Fake news stories produce this form of content, which threatens those age
groups, genders, culture, political ideology.
3. Knowledge-based: These forms of communications explain some unresolved problems
by researchers (so-called) or make users believe it is true.
4. Style-based: Journalists who pretend to replicate the style of other accredited journalists
are writing blogs.
5. Stance-based: It is also a depiction of valid statements so that its meaning and intent are
changed.
3. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
3
Figure 1: Different approaches to fake news detection.
Looking at Figure 1, it can be seen that unsupervised & supervised models of learning are
primarily based on textual news content. Note that ML models are usually a tradeoff of precision
or recall so that a model that is very effective in detected counterfeit news will have a high false-
positive rate relative to a model that does not likely show a low false-positive rate. However
automated censorship related topics can not be discussed here due to ethical issues.
2. LITERATURE REVIEW
Various researchers have tried several ways to overcome the problem, to test which approach
works, and to achieve desired outcomes. A variety of studies have discussed fake news
identification methods, including role extraction or model construction, from a data mining
viewpoint. Feature extraction approaches (both features of news material and features of the
social context) combined with metric assessment using accuracy, reminder, or F1 ratings have
been shown to bring informed results. Other parameters such as bot-spam, click-bait, and news
source also impact the forecast [4].
These were approached by data mining & NLP, but researchers have become increasingly
involved in heavy NN-oriented methods with more & more research & development in AI. A
story illustrates a method for "capturing," "picking" or "integrating." A model of RNNs is created
for the identification of fake data. They also used the RNN to compile the temporal pattern of
user behaviors around a single article/text. All of this material is used and incorporated into a
false news classification model bu the authors in [5].
Another paper [6]deals with LSTM or CNN linguistically-infused NN for classifying Twitter
messages. The GloVe Library of Pretrained Vectors was used to introduce the linguistic
component [6].
This review provides a model for recognizing forged twitter news in the light of automatic forged
Twitter news detection, through finding out how to anticipate reliable estimates We have later
conducted a separate comparison of the efficiency of classification performance of dataset among
five well-known ML algos like SVM, NB Process, LR & RNN. Our experimental findings have
shown that the SVM & NB classification is above the other algos. [7].
To solve this issue, this paper explores the approach of NLP & ML. Using bag-of-words, n-
grams, count vectorizers, TF-IDF, or data from five classifiers have been equipped to investigate
4. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
4
which of these data sets of labeled news statements works well. The accuracy, the reminder or f1
scores allow us to find the most effective model. [8].
The authors have described in [9] the language cue approaches with ML, word bags, rhetorical
structure or discourse analysis, network analysis approaches, or SVM classifiers. These text-
based models do not or do not boost existing approaches.
In [10], authors appreciated or analyzed the principles, methods, or algos used for categorizing
fake or fabricated social network news items, authors & subjects. the paper highlighted the study
challenge by exposed features of fake news & numerous similarities between news articles,
writers & topics. authors of the paper address the FakeDetector model of auto-found news. It
creates a deep, pervasive network model based on text and helps the author and subject to learn
about the presentation of news articles simultaneously.
3. RESEARCH METHODOLOGY
Problem Formulation
Social media is a double-edged sword for news consumption. On the one hand, the low cost,
quick access & fast circulation lead users in the social media to browse and receive content. It
makes, on the other hand, a wide variety of " FNs " with purposely misleading facts, i.e. bad
quality news. The widespread of fake news is very harmful to individuals & society.
Consequently, fake social media news detection has recently grown into an emerging study that
attracts considerable interest. Fake social media news detection presents unique features or
challenges which make existing news algos ineffective or not applicable. Firstly, FNs is written
purposely to mislead readers with FNs, which makes the detection of that too difficult
& inexplicably dependent on news content; therefore, supplemental data, like user social
commitments on social media, must be included in our evaluation. The second challenge, as
social commitments of users with FNs, produces big, incomplete, unstructured, and noisy data is
to manipulate this auxiliary material.
Data Pre-Processing
It should be pre-processed before the use of AI algos on the results. First of all, only statements
themselves were decided for purposes of classification. That means none of the provided
metadata is used for grading. This metadata could improve classification algo in the future. The
following measures have been used for pre-processing:
Split the statements into individual tokens (words).
All numbers are deleted.
Removal of all marks of punctuation.
Delete all other characters, not alpha.
The rest of the tokens should apply a steaming process. In linguistic morphology, the
approach used to avoid (or lemmatize) recruitment information is to reduce inflected or
derived terms in the terms stem, base, or root shape – typically in writing. This encourages
them to use words that are identical to the same ones (for example "write").
Stop word removal. Word stops are included in basically a kind of text. These terms are
popular and do not impact their importance, so it is beneficial to get rid of them [6].
Substitute terms with tf-idf values. In information collection tf – idf is a figural metric that
represents the meaning of a word for a record in a collection or corpus and is a short term
for frequency-inverse document frequency. [7].
5. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
5
1. Accuracy classification based on six available categories
2. Accuracy in binary classification: This metric is as if 2 possible categories were used in the
statement – true (according to the last three categories described above) or false (based on
the first three above mentioned categories).
Proposed Approach
Several AI algos were used for statement classification during the current study, of which DNN
provided the best results. With some parameter tuning, this work selected the modified NN as
the approach proposed that is superior to DNN. This is implemented with available standard
libraries of a scikit-learn (a programming language library of Python). Two separate measures
have been evaluated for all algos:
Accuracy classification based on 6 obtainable category.
Binary accuracy of the classification: This metric is regarded as being only available in 2
possible categories, real (based on the last three categories mentioned above).
Recurrent Neural Network
RNN is NN that is specialized in data processing x(t)= x(1), . . . , x(τ) with an index time step of t
between 1 and μm. It is also easier to use RNNs for tasks containing sequential inputs like speech
and language. RNN is called repeating since it achieves the same purpose for each part of a
sequence, based on the previous calculations.
Architecture: Let's go over an important RNN network briefly.
Fig 2: Architecture of RNN
The left side of the diagram showed an RNN notation as well as an RNN was unrolled (or
unfolded) into a whole network on the right-hand side. We say we write down the system for the
whole sequence. by unrolling. If the sequence we care about is a sentence of three terms, for
example, the network will be deployed into a network of three layers, one layer for each word.
Input: x(t) shall be taken as network input at step t.
Hidden state: h(t) is hidden state at time t, function as network memory. h(t) is present
input or hidden state for the previous time stage: h(t) = f(U x(t) + W h(t−1)).
Weights: The RNN has contributions to hidden connections, a U weight matrix parameterized,
the W weight matrix parameterized, or V weight matrix parameterized, to hidden recurring
connections, which are distributed over time (U, V, W).
6. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
6
Output: o(t) illustrates the output of the network.
4. RESULTS ILLUSTRATIONS
A RAMP studio, a famour machine learning workflow research team [3] has compiled the data
set used for training & testing. It includes brief statements by revered persons available on social
media. For statement 6 primary labels were available. The same dataset is taken from web
repositoty and binary and multi class deep neural networks are applied with parameter tuning for
optimized results.
Fig 3: Binary DNN
Fig 4: Multi DNN
Figures 3 & 4 show the results of the simulation tool showing multiple parameters of accuracy in the case
of DNN of two different classes.
Fig 5: Binary RNN
7. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
7
Fig 6: Multiclass RNN
Figures 5 and 6 visualize the result of RNN in both binary and multi-class analysis. From all the
figures it has been noticed that RNN has given the best accuracy in all the parameters of accurate
measurements.
Fig 7: Comparison graph of Binary class of DNN and RNN
Fig 8: Comparison graph of multi-class of DNN and RNN
In figures 7 & 8, the outcomes of the summarized classifications are shown. The present
NN provides the best outcomes in the accuracy of both multiclass and binary classes.
5. CONCLUSION
The fake news challenge is risky or quickly spreads like a wildfire since information can be
readily reached in different ways. In this paper, we have compared two NN models for checking
the verification of data extracted from the RAMP studio. From the results predicted it clearly
75
80
85
90
95
100
105
Acc Rec Prec F1-Meas
DNN
RNN
75
80
85
90
95
100
DNN
RNN
8. Computer Applications: An International Journal (CAIJ), Vol.8, No.3, August 2021
8
shows how RNN has given the best accuracy on the same dataset as was used in the previous
work. Although fake news or messages using different ML methods can be identified
successfully in many previous papers. Deep neural network works better in non linear
voluminous data set. Hence RNN is found to be the best method in the classification of the
news.In the future, we need to improve the results of Multiclass DNN and RNN as RNN can give
an accuracy of 86% in a multiclass dataset. A new technique is hence needed to invent for this
purpose.
REFERENCES
[1] Shao, C., Ciampaglia, G. L., Varol, O., Flammini, A., &Menczer, F. (2017). The spread of fake news
by social bots. arXiv preprint arXiv:1707.07592, 96-104.
[2] Hunt, E. (2016). What is fake news? How to spot it and what you can do to stop it. The Guardian, 17
[3] Golbeck, Jennifer, et al. “A Large Labeled Corpus for Online Harassment Research” Proceeding of
the 2017 ACM on Web Science Conference. ACM,2017
[4] Shu, K., Sliva, A., Wang, S., Tang, J., & Liu, H. (2017). Fake news detection on social media: A data
mining perspective. ACM SIGKDD Explorations Newsletter, 19(1), 22-36.
[5] Ruchansky, N., Seo, S., & Liu, Y. (2017, November). CSI: A hybrid deep model for fake news
detection. In Proceedings of the 2017 ACM on Conference on Information and Knowledge
Management (pp. 797-806). ACM.
[6] Volkova, S., Shaffer, K., Jang, J. Y., &Hodas, N. (2017, July). Separating facts from fiction:
Linguistic models to classify suspiciously and trusted news posts on twitter. In Proceedings of the
55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (pp.
647-653).
[7] Abdullah-All-Tanvir, Mahir, E. M., Akhter, S., & Huq, M. R. (2019). Detecting Fake News using
Machine Learning and Deep Learning Algorithms. 2019 7th International Conference on Smart
Computing & Communications (ICSCC).
[8] Vasu Agarwala, H.Parveen Sultanaa, Srijan Malhotra, Amitrajit Sarkar “Analysis of Classifiers for
Fake News Detection” INTERNATIONAL CONFERENCE ON RECENT TRENDS IN
ADVANCED COMPUTING 2019, ICRTAC 2019
[9] Conroy, N. J., Rubin, V. L., & Chen, Y. (2015, November). Automatic deception detection: Methods
for finding fake news. In Proceedings of the 78th ASIS&T Annual Meeting: Information Science
with Impact: Research in and for the Community (p. 82). American Society for Information Science.
[10] Zhang, J., Cui, L., Fu, Y., & Gouza, F. B. (2018). Fake news detection with the deep diffusive
network model. arXiv preprint arXiv:1805.08751.