In this Slide Share, Professor Aditya Bagchi explained Interplay of Trust and Risk in Social Media Communication.
We will discuss how cybercrime through social media is affecting us and what we can do to counter that.
This document discusses using Twitter data for sentiment analysis and influence tracking. It describes how Twitter data was collected using its APIs and preprocessed by removing links, usernames and stopwords. N-grams and part-of-speech tags were then extracted as features from the tweets. The tweets were classified into positive, negative, neutral or irrelevant categories. Sentiment analysis was performed at the entity level to determine sentiment towards specific topics mentioned in tweets, like products. Influence was tracked using algorithms that rank users based on retweets, followers and mentions.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
The architecture social media and online newspaper credibility measurement fo...TELKOMNIKA JOURNAL
This document proposes an architecture for detecting fake news on social media and online newspapers. It involves three main steps: 1) keyword extraction from user-input text, 2) news scoring to measure credibility of information from online news sources based on time, website and message credibility factors, and 3) social media scoring to further measure credibility from social media based on similar factors, if news scoring is insufficient. The goal is to produce a likelihood score of the information being fake news or fact. Previous related work on fake news detection using knowledge-based, context-based and style-based approaches is also discussed.
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
An Automated Model to Detect Fake Profiles and botnets in Online Social Netwo...IOSR Journals
This document discusses an automated model for detecting fake profiles and botnets in online social networks. It begins with background on the prevalence of fake accounts, which can compromise user privacy and security. Next, it reviews related work on using data hiding techniques like steganography and watermarking to embed information in profile pictures in order to identify suspicious accounts. The proposed model aims to automatically detect fake profiles and botnets to replace current manual methods that are costly and labor-intensive.
How to Make People Click on a Dangerous Link Despite their Security Awareness mark-smith
It is possible to make virtually any person click on a link, as any person will be curious about something, or interested in some topic, or find the message plausible because they know the sender, or because it fits their expectations (context).
Data mining applied about polygamy using sentiment analysis on Twitters in In...journalBEEI
Polygamy remains one of the interesting key topics in various society, especially Indonesia. Polygamy is the act of marrying multiple spouses that means having more one wife at the same time, is common case in worldwide. In this problem, the most of Muslim in Indonesia adopt Islamic law that let men to do the polygamy with certain requirements. In worldwide communities, this has been a prominent feature and has become the subject of numerous books, heated debates, journal articles, discussion papers, and so on. Furthermore, in Indonesia the polygamous marriage has recently become a heated topic. Despite controversy, Indonesia’s law, of recent, allows people for doing polygamy in certain conditions. Because the polygamy is often debated, this study focuses on assessment of Indonesian’s perceptions through sentiment analysis and would determine people’s perception about polygamy issue from 500 tweets on Twitter. To conclude, it elucidates that the polygamy in Indonesia is a normal thing and few assume the case is as negative thing.
This document discusses using Twitter data for sentiment analysis and influence tracking. It describes how Twitter data was collected using its APIs and preprocessed by removing links, usernames and stopwords. N-grams and part-of-speech tags were then extracted as features from the tweets. The tweets were classified into positive, negative, neutral or irrelevant categories. Sentiment analysis was performed at the entity level to determine sentiment towards specific topics mentioned in tweets, like products. Influence was tracked using algorithms that rank users based on retweets, followers and mentions.
International Journal of Engineering Research and DevelopmentIJERD Editor
Electrical, Electronics and Computer Engineering,
Information Engineering and Technology,
Mechanical, Industrial and Manufacturing Engineering,
Automation and Mechatronics Engineering,
Material and Chemical Engineering,
Civil and Architecture Engineering,
Biotechnology and Bio Engineering,
Environmental Engineering,
Petroleum and Mining Engineering,
Marine and Agriculture engineering,
Aerospace Engineering.
The architecture social media and online newspaper credibility measurement fo...TELKOMNIKA JOURNAL
This document proposes an architecture for detecting fake news on social media and online newspapers. It involves three main steps: 1) keyword extraction from user-input text, 2) news scoring to measure credibility of information from online news sources based on time, website and message credibility factors, and 3) social media scoring to further measure credibility from social media based on similar factors, if news scoring is insufficient. The goal is to produce a likelihood score of the information being fake news or fact. Previous related work on fake news detection using knowledge-based, context-based and style-based approaches is also discussed.
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
An Automated Model to Detect Fake Profiles and botnets in Online Social Netwo...IOSR Journals
This document discusses an automated model for detecting fake profiles and botnets in online social networks. It begins with background on the prevalence of fake accounts, which can compromise user privacy and security. Next, it reviews related work on using data hiding techniques like steganography and watermarking to embed information in profile pictures in order to identify suspicious accounts. The proposed model aims to automatically detect fake profiles and botnets to replace current manual methods that are costly and labor-intensive.
How to Make People Click on a Dangerous Link Despite their Security Awareness mark-smith
It is possible to make virtually any person click on a link, as any person will be curious about something, or interested in some topic, or find the message plausible because they know the sender, or because it fits their expectations (context).
Data mining applied about polygamy using sentiment analysis on Twitters in In...journalBEEI
Polygamy remains one of the interesting key topics in various society, especially Indonesia. Polygamy is the act of marrying multiple spouses that means having more one wife at the same time, is common case in worldwide. In this problem, the most of Muslim in Indonesia adopt Islamic law that let men to do the polygamy with certain requirements. In worldwide communities, this has been a prominent feature and has become the subject of numerous books, heated debates, journal articles, discussion papers, and so on. Furthermore, in Indonesia the polygamous marriage has recently become a heated topic. Despite controversy, Indonesia’s law, of recent, allows people for doing polygamy in certain conditions. Because the polygamy is often debated, this study focuses on assessment of Indonesian’s perceptions through sentiment analysis and would determine people’s perception about polygamy issue from 500 tweets on Twitter. To conclude, it elucidates that the polygamy in Indonesia is a normal thing and few assume the case is as negative thing.
Teaching Johnny Not to Fall for Phish, for ISSA 2010 on May 2010Jason Hong
This document discusses research into detecting and preventing phishing attacks. It begins by outlining the negative impacts of phishing from consumer, corporate, and societal perspectives. It then describes how phishing attacks are becoming more sophisticated through spear phishing, social engineering, and the use of malware. The document outlines various techniques developed by the author and his colleagues to educate users and automatically detect phishing attacks, including embedded phishing training software called PhishGuru, educational microgames like Anti-Phishing Phil, analyzing search engine results to detect fake websites, and using machine learning to label new phishing sites based on similarities to known ones. User studies demonstrated that embedded, game-based training can significantly reduce users falling for phishing
Identifying ghost users using social media metadata - University College LondonGreg Kawere
You are your Metadata: Identification and Obfuscation of Social Media Users using Metadata Information a joint research project of the Alan Turing Institute and University College in London
IRJET- Fake Profile Identification using Machine LearningIRJET Journal
This document proposes a framework for identifying fake profiles on social media sites using machine learning. It involves selecting profile attributes from a dataset, training a random forest classifier model on 80% of the labeled real and fake profiles, and testing it on the remaining 20%. The random forest classifier is able to accurately classify profiles as real or fake with an efficiency of 95% based on attributes like number of friends, followers, and status updates. This automatic fake profile detection method could help social media sites manage the large volume of profiles that can't be manually reviewed.
Cyber bullying Detection based on Semantic-Enhanced Marginalized Denoising Au...dbpublications
As a side effect of increasingly popular social media, cyberbullying has emerged as a serious problem afflicting children, adolescents and young adults. Machine learning techniques make automatic detection of bullying messages in social media possible, and this could help to construct a healthy and safe social media environment. In this meaningful research area, one critical issue is robust and discriminative numerical representation learning of text messages. In this paper, we propose a new representation learning method to tackle this problem. Our method named Semantic-Enhanced Marginalized Denoising Auto-Encoder (smSDA) is developed via semantic extension of the popular deep learning model stacked denoising autoencoder. The semantic extension consists of semantic dropout noise and sparsity constraints, where the semantic dropout noise is designed based on domain knowledge and the word embedding technique. Our proposed method is able to exploit the hidden feature structure of bullying information and learn a robust and discriminative representation of text. Comprehensive experiments on two public cyberbullying corpora (Twitter and MySpace) are conducted, and the results show that our proposed approaches outperform other baseline text representation learning methods..
IRJET - Real-Time Cyberbullying Analysis on Social Media using Machine Learni...IRJET Journal
This document presents a system for real-time analysis of cyberbullying on social media using machine learning and text mining. The system aims to detect abusive conversations and censor harmful words to protect victims. It uses an artificial neural network machine learning algorithm to analyze words that could psychologically affect individuals. The system identifies abusive words in posts and comments and replaces them with censored content. This aims to prevent innocent users from being exposed to depressing or criminal activities online. The document discusses the system architecture, including tools for sentiment analysis, monitoring discussions, identifying abusive words, and updating a word database. Diagrams show the data flow, use cases, and interactions between system components.
The advancement of Information Technology has hastened the ability to disseminate information across the globe. In particular, the recent trends in ‘Social Networking’ have led to a spark in personally sensitive information being published on the World Wide Web. While such socially active websites are creative tools for expressing one’s personality it also entails serious privacy concerns. Thus, Social Networking websites could be termed a double edged sword. It is important for the law to keep abreast of these developments in technology. The purpose of this paper is to demonstrate the limits of extending existing laws to battle privacy intrusions in the Internet especially in the context of social networking. It is suggested that privacy specific legislation is the most appropriate means of protecting online privacy. In doing so it is important to maintain a balance between the competing right of expression, the failure of which may hinder the reaping of benefits offered by Internet technology
This document summarizes a research paper that proposes using a logistic regression classifier trained with stochastic gradient descent to predict Twitter users' personalities from their tweets. It begins with an abstract of the paper and an introduction on personality prediction from social media. It then provides more detail on the anatomy of the research, including defining personality prediction from Twitter, its applications, and the general process of using machine learning for the task. Next, it reviews several previous studies on personality prediction from Twitter and social networks, noting their approaches, findings and limitations. It identifies remaining research gaps, such as the need for improved linguistic analysis of tweets and more robust/scalable predictive models. Finally, it proposes using a logistic regression classifier as the personality prediction model to address
DETECTION OF FAKE ACCOUNTS IN INSTAGRAM USING MACHINE LEARNINGijcsit
With the advent of the Internet and social media, while hundreds of people have benefitted from the vast sources of information available, there has been an enormous increase in the rise of cyber-crimes, particularly targeted towards women. According to a 2019 report in the [4] Economics Times, India has witnessed a 457% rise in cybercrime in the five year span between 2011 and 2016. Most speculate that this is due to impact of social media such as Facebook, Instagram and Twitter on our daily lives. While these definitely help in creating a sound social network, creation of user accounts in these sites usually needs just an email-id. A real life person can create multiple fake IDs and hence impostors can easily be made. Unlike the real world scenario where multiple rules and regulations are imposed to identify oneself in a unique manner (for example while issuing one’s passport or driver’s license), in the virtual world of social media, admission does not require any such checks. In this paper, we study the different accounts of Instagram, in particular and try to assess an account as fake or real using Machine Learning techniques namely Logistic Regression and Random Forest Algorithm.
This document discusses methods for measuring privacy in online social networks. It first introduces the concept of the Privacy Index (PIDX), which quantifies a user's privacy exposure based on their visible attributes. It then describes calculating a user's Privacy Quotient (PQ) using a naive approach, which considers the sensitivity and visibility of shared profile items. Finally, it proposes a new model called Privacy Armor that would measure privacy leaks in unstructured data like posts and messages using an Item Response Theory model to calculate sensitivity, visibility, and privacy quotient. The goal is to warn users about unintended sharing of private information online.
Automatic detection of online abuse and analysis of problematic users in wiki...Melissa Moody
For their 2019 capstone project, DSI Master of Science in Data Science students Charu Rawat, Arnab Sarkar, and Sameer Singh proposed a framework to understand and detect such abuse in the English Wikipedia community.
Rawat, Sarkar, and Singh received the award for Best Paper in the Data Science for Society category at the 2019 Systems & Information Design Symposium (SIEDS). In "Automatic Detection of Online Abuse and Analysis of Problematic Users in Wikipedia," the team presented an analysis of user misconduct in Wikipedia and a system for the automated early detection of inappropriate behavior.
IRJET- Tweet Segmentation and its Application to Named Entity RecognitionIRJET Journal
This document summarizes a research paper on tweet segmentation and its application to named entity recognition. It proposes a novel framework called Hybrid Segmentation that splits tweets into meaningful segments to preserve semantic context for downstream natural language processing applications like named entity recognition. HybridSeg finds optimal tweet segmentations by maximizing the stickiness score of segments, which considers global context based on English phrases and local context based on linguistic features or term dependencies within a batch of tweets. Experiments show significantly improved segmentation quality when learning both global and local contexts compared to global context alone. The segmented tweets can then be used for high accuracy named entity recognition through part-of-speech tagging.
The document describes a project to detect fake news using machine learning models. It discusses how the project classified news websites as real or fake using a combination of bag-of-words, word embeddings and feature descriptions with 87.39% accuracy. Some ways to improve the model are also provided, such as using more features in the word embeddings. Real-world applications of fake news detection include verifying news on social media during elections and detecting fake job postings.
MDS 2011 Paper: An Unsupervised Approach to Discovering and Disambiguating So...Carlton Northern
This document presents an unsupervised approach to discover and disambiguate social media profiles for a large group of individuals, such as employees or university students. The approach uses a combination of search engine queries, semantic web queries, and directly polling social media sites to discover potential profiles. It then applies heuristics involving keyword matching, community structure analysis, and extracting semantic and profile features to disambiguate the true profiles from false positives. The approach was tested on a set of 2016 university computer science student logins, achieving a precision of 0.863 and F-measure of 0.654 at discovering their real social media profiles from a ground truth data set.
Social network has become so popular with overwhelming high rate of growth, due to this popularity the online social networks is facing the issues of spamming, which has leads to unsubstantial economic loss to this menace of spam and spammers activities. It has leads to uncontrollable dissemination of viruses and malwares, promotional ads, phishing, and scams. spam activities has enter a new dangerous dimension, the spammers have step up their games and tactics online social networks, it consumes large amounts of network bandwidth leading to less revenue and significant economic loss to both private and public sectors. From the previous scholars work on spammer classification taxonomy, various machine learning techniques have been extensively used to detect spam activities and spammers in online social networks. There are various classifier that are learn over content-based features extracted from the user's interactions and profiles to label them as spam/spammers or legitimate. But recently, new network structural bench mark features have been proposed for spammer detection task, but their importance using structural bench mark learning methods has not been extensively evaluated yet. In this research work, we evaluate the the metric performance of some structural bench mark learning methods using scientific and strategic approach based attributes extracted from an interaction network for the task of spammer detection in online social network.
This document provides an overview of a digital literacy and citizenship course taught by Mrs. Joyce Lourenço Pereira. The course uses a research-based curriculum developed by Common Sense Media to teach students skills and ethical decision making related to digital media use. Over the course of the year, students will explore topics like internet safety, digital footprint, cyberbullying, information literacy, and copyright through lessons, assessments, and multimedia projects using tools like Edmodo, Gmail, iPads, and various web apps. The goal is for students to safely and responsibly navigate the digital world.
This document discusses a study of privacy and information sharing on Facebook. The researchers surveyed Facebook users at a US university to understand their privacy concerns, usage of Facebook, and awareness of how public their profiles were. They found that while users expressed privacy concerns, they were not very concerned about privacy on Facebook. Some users were also unaware how exposed their profiles and information were. Exposure to information about Facebook's public nature sometimes changed users' behavior regarding what they shared.
This document provides an overview of special topics in social media services, including social network analysis, link mining, text mining, web mining, and opinion mining in social media. It discusses key concepts such as social network extraction, identifying prominent actors, and characteristics of collaboration networks. The document also provides examples and definitions for social network analysis metrics like degree centrality, betweenness centrality, closeness centrality, and eigenvector. Finally, it introduces relevant books and topics for further reading.
IRJET- Personality Recognition using Social Media DataIRJET Journal
This document summarizes a research paper on personality recognition using social media data. The paper proposes analyzing personality traits based on the Big Five model (openness, conscientiousness, extraversion, agreeableness, neuroticism) using Facebook status updates and machine learning. Specifically, it involves collecting Facebook status data through a browser extension, storing it in a database, training a random forest regression model on an existing dataset, and using the model to predict personality trait values for additional Facebook users. The predicted traits are then visualized with radar charts to provide an overview of each user's personality profile.
Social networking allows users to interact and connect with other users through dedicated websites and applications. It involves expanding business and social contacts by connecting with people who share similar interests. Popular social networking sites emerged in the 1990s like MySpace and Facebook. While social networking allows users to stay connected with friends and meet new people, it also presents privacy, security, and oversharing risks that users must be aware of. Analyzing social networks can occur at the micro, meso, and macro levels.
Potential vulnerabilities to e-learning - MimecastJisc
Phillip Hay analyzes threats to e-learning. He found that the education sector saw a 87% increase in spam and phishing attacks between March and April 2020 as students shifted to online learning during COVID-19. His recommendations include only using approved school software on school-managed accounts, maintaining online etiquette, and educating students and staff about safe online practices like strong passwords and avoiding suspicious links. Hay warns that cyber attacks will likely continue targeting remote workers and students through increasingly sophisticated social engineering techniques.
Comprehensive Social Media Security Analysis & XKeyscore Espionage TechnologyCSCJournals
Social networks can offer many services to the users for sharing activities events and their ideas. Many attacks can happened to the social networking websites due to trust that have been given by the users. Cyber threats are discussed in this paper. We study the types of cyber threats, classify them and give some suggestions to protect social networking websites of variety of attacks. Moreover, we gave some antithreats strategies with future trends.
A HYBRID CLASSIFICATION ALGORITHM TO CLASSIFY ENGINEERING STUDENTS’ PROBLEMS ...IJDKP
The social networking sites have brought a new horizon for expressing views and opinions of individuals.
Moreover, they provide medium to students to share their sentiments including struggles and joy during the
learning process. Such informal information has a great venue for decision making. The large and growing
scale of information needs automatic classification techniques. Sentiment analysis is one of the automated
techniques to classify large data. The existing predictive sentiment analysis techniques are highly used to
classify reviews on E-commerce sites to provide business intelligence. However, they are not much useful
to draw decisions in education system since they classify the sentiments into merely three pre-set
categories: positive, negative and neutral. Moreover, classifying the students’ sentiments into positive or
negative category does not provide deeper insight into their problems and perks. In this paper, we propose
a novel Hybrid Classification Algorithm to classify engineering students’ sentiments. Unlike traditional
predictive sentiment analysis techniques, the proposed algorithm makes sentiment analysis process
descriptive. Moreover, it classifies engineering students’ perks in addition to problems into several
categories to help future students and education system in decision making.
Teaching Johnny Not to Fall for Phish, for ISSA 2010 on May 2010Jason Hong
This document discusses research into detecting and preventing phishing attacks. It begins by outlining the negative impacts of phishing from consumer, corporate, and societal perspectives. It then describes how phishing attacks are becoming more sophisticated through spear phishing, social engineering, and the use of malware. The document outlines various techniques developed by the author and his colleagues to educate users and automatically detect phishing attacks, including embedded phishing training software called PhishGuru, educational microgames like Anti-Phishing Phil, analyzing search engine results to detect fake websites, and using machine learning to label new phishing sites based on similarities to known ones. User studies demonstrated that embedded, game-based training can significantly reduce users falling for phishing
Identifying ghost users using social media metadata - University College LondonGreg Kawere
You are your Metadata: Identification and Obfuscation of Social Media Users using Metadata Information a joint research project of the Alan Turing Institute and University College in London
IRJET- Fake Profile Identification using Machine LearningIRJET Journal
This document proposes a framework for identifying fake profiles on social media sites using machine learning. It involves selecting profile attributes from a dataset, training a random forest classifier model on 80% of the labeled real and fake profiles, and testing it on the remaining 20%. The random forest classifier is able to accurately classify profiles as real or fake with an efficiency of 95% based on attributes like number of friends, followers, and status updates. This automatic fake profile detection method could help social media sites manage the large volume of profiles that can't be manually reviewed.
Cyber bullying Detection based on Semantic-Enhanced Marginalized Denoising Au...dbpublications
As a side effect of increasingly popular social media, cyberbullying has emerged as a serious problem afflicting children, adolescents and young adults. Machine learning techniques make automatic detection of bullying messages in social media possible, and this could help to construct a healthy and safe social media environment. In this meaningful research area, one critical issue is robust and discriminative numerical representation learning of text messages. In this paper, we propose a new representation learning method to tackle this problem. Our method named Semantic-Enhanced Marginalized Denoising Auto-Encoder (smSDA) is developed via semantic extension of the popular deep learning model stacked denoising autoencoder. The semantic extension consists of semantic dropout noise and sparsity constraints, where the semantic dropout noise is designed based on domain knowledge and the word embedding technique. Our proposed method is able to exploit the hidden feature structure of bullying information and learn a robust and discriminative representation of text. Comprehensive experiments on two public cyberbullying corpora (Twitter and MySpace) are conducted, and the results show that our proposed approaches outperform other baseline text representation learning methods..
IRJET - Real-Time Cyberbullying Analysis on Social Media using Machine Learni...IRJET Journal
This document presents a system for real-time analysis of cyberbullying on social media using machine learning and text mining. The system aims to detect abusive conversations and censor harmful words to protect victims. It uses an artificial neural network machine learning algorithm to analyze words that could psychologically affect individuals. The system identifies abusive words in posts and comments and replaces them with censored content. This aims to prevent innocent users from being exposed to depressing or criminal activities online. The document discusses the system architecture, including tools for sentiment analysis, monitoring discussions, identifying abusive words, and updating a word database. Diagrams show the data flow, use cases, and interactions between system components.
The advancement of Information Technology has hastened the ability to disseminate information across the globe. In particular, the recent trends in ‘Social Networking’ have led to a spark in personally sensitive information being published on the World Wide Web. While such socially active websites are creative tools for expressing one’s personality it also entails serious privacy concerns. Thus, Social Networking websites could be termed a double edged sword. It is important for the law to keep abreast of these developments in technology. The purpose of this paper is to demonstrate the limits of extending existing laws to battle privacy intrusions in the Internet especially in the context of social networking. It is suggested that privacy specific legislation is the most appropriate means of protecting online privacy. In doing so it is important to maintain a balance between the competing right of expression, the failure of which may hinder the reaping of benefits offered by Internet technology
This document summarizes a research paper that proposes using a logistic regression classifier trained with stochastic gradient descent to predict Twitter users' personalities from their tweets. It begins with an abstract of the paper and an introduction on personality prediction from social media. It then provides more detail on the anatomy of the research, including defining personality prediction from Twitter, its applications, and the general process of using machine learning for the task. Next, it reviews several previous studies on personality prediction from Twitter and social networks, noting their approaches, findings and limitations. It identifies remaining research gaps, such as the need for improved linguistic analysis of tweets and more robust/scalable predictive models. Finally, it proposes using a logistic regression classifier as the personality prediction model to address
DETECTION OF FAKE ACCOUNTS IN INSTAGRAM USING MACHINE LEARNINGijcsit
With the advent of the Internet and social media, while hundreds of people have benefitted from the vast sources of information available, there has been an enormous increase in the rise of cyber-crimes, particularly targeted towards women. According to a 2019 report in the [4] Economics Times, India has witnessed a 457% rise in cybercrime in the five year span between 2011 and 2016. Most speculate that this is due to impact of social media such as Facebook, Instagram and Twitter on our daily lives. While these definitely help in creating a sound social network, creation of user accounts in these sites usually needs just an email-id. A real life person can create multiple fake IDs and hence impostors can easily be made. Unlike the real world scenario where multiple rules and regulations are imposed to identify oneself in a unique manner (for example while issuing one’s passport or driver’s license), in the virtual world of social media, admission does not require any such checks. In this paper, we study the different accounts of Instagram, in particular and try to assess an account as fake or real using Machine Learning techniques namely Logistic Regression and Random Forest Algorithm.
This document discusses methods for measuring privacy in online social networks. It first introduces the concept of the Privacy Index (PIDX), which quantifies a user's privacy exposure based on their visible attributes. It then describes calculating a user's Privacy Quotient (PQ) using a naive approach, which considers the sensitivity and visibility of shared profile items. Finally, it proposes a new model called Privacy Armor that would measure privacy leaks in unstructured data like posts and messages using an Item Response Theory model to calculate sensitivity, visibility, and privacy quotient. The goal is to warn users about unintended sharing of private information online.
Automatic detection of online abuse and analysis of problematic users in wiki...Melissa Moody
For their 2019 capstone project, DSI Master of Science in Data Science students Charu Rawat, Arnab Sarkar, and Sameer Singh proposed a framework to understand and detect such abuse in the English Wikipedia community.
Rawat, Sarkar, and Singh received the award for Best Paper in the Data Science for Society category at the 2019 Systems & Information Design Symposium (SIEDS). In "Automatic Detection of Online Abuse and Analysis of Problematic Users in Wikipedia," the team presented an analysis of user misconduct in Wikipedia and a system for the automated early detection of inappropriate behavior.
IRJET- Tweet Segmentation and its Application to Named Entity RecognitionIRJET Journal
This document summarizes a research paper on tweet segmentation and its application to named entity recognition. It proposes a novel framework called Hybrid Segmentation that splits tweets into meaningful segments to preserve semantic context for downstream natural language processing applications like named entity recognition. HybridSeg finds optimal tweet segmentations by maximizing the stickiness score of segments, which considers global context based on English phrases and local context based on linguistic features or term dependencies within a batch of tweets. Experiments show significantly improved segmentation quality when learning both global and local contexts compared to global context alone. The segmented tweets can then be used for high accuracy named entity recognition through part-of-speech tagging.
The document describes a project to detect fake news using machine learning models. It discusses how the project classified news websites as real or fake using a combination of bag-of-words, word embeddings and feature descriptions with 87.39% accuracy. Some ways to improve the model are also provided, such as using more features in the word embeddings. Real-world applications of fake news detection include verifying news on social media during elections and detecting fake job postings.
Similar to AILABS - Lecture Series - Is AI the New Electricity? Topic:- Interplay of Trust and Risk in Social Media Communication, Presented by - Aditya Bagchi
MDS 2011 Paper: An Unsupervised Approach to Discovering and Disambiguating So...Carlton Northern
This document presents an unsupervised approach to discover and disambiguate social media profiles for a large group of individuals, such as employees or university students. The approach uses a combination of search engine queries, semantic web queries, and directly polling social media sites to discover potential profiles. It then applies heuristics involving keyword matching, community structure analysis, and extracting semantic and profile features to disambiguate the true profiles from false positives. The approach was tested on a set of 2016 university computer science student logins, achieving a precision of 0.863 and F-measure of 0.654 at discovering their real social media profiles from a ground truth data set.
Social network has become so popular with overwhelming high rate of growth, due to this popularity the online social networks is facing the issues of spamming, which has leads to unsubstantial economic loss to this menace of spam and spammers activities. It has leads to uncontrollable dissemination of viruses and malwares, promotional ads, phishing, and scams. spam activities has enter a new dangerous dimension, the spammers have step up their games and tactics online social networks, it consumes large amounts of network bandwidth leading to less revenue and significant economic loss to both private and public sectors. From the previous scholars work on spammer classification taxonomy, various machine learning techniques have been extensively used to detect spam activities and spammers in online social networks. There are various classifier that are learn over content-based features extracted from the user's interactions and profiles to label them as spam/spammers or legitimate. But recently, new network structural bench mark features have been proposed for spammer detection task, but their importance using structural bench mark learning methods has not been extensively evaluated yet. In this research work, we evaluate the the metric performance of some structural bench mark learning methods using scientific and strategic approach based attributes extracted from an interaction network for the task of spammer detection in online social network.
This document provides an overview of a digital literacy and citizenship course taught by Mrs. Joyce Lourenço Pereira. The course uses a research-based curriculum developed by Common Sense Media to teach students skills and ethical decision making related to digital media use. Over the course of the year, students will explore topics like internet safety, digital footprint, cyberbullying, information literacy, and copyright through lessons, assessments, and multimedia projects using tools like Edmodo, Gmail, iPads, and various web apps. The goal is for students to safely and responsibly navigate the digital world.
This document discusses a study of privacy and information sharing on Facebook. The researchers surveyed Facebook users at a US university to understand their privacy concerns, usage of Facebook, and awareness of how public their profiles were. They found that while users expressed privacy concerns, they were not very concerned about privacy on Facebook. Some users were also unaware how exposed their profiles and information were. Exposure to information about Facebook's public nature sometimes changed users' behavior regarding what they shared.
This document provides an overview of special topics in social media services, including social network analysis, link mining, text mining, web mining, and opinion mining in social media. It discusses key concepts such as social network extraction, identifying prominent actors, and characteristics of collaboration networks. The document also provides examples and definitions for social network analysis metrics like degree centrality, betweenness centrality, closeness centrality, and eigenvector. Finally, it introduces relevant books and topics for further reading.
IRJET- Personality Recognition using Social Media DataIRJET Journal
This document summarizes a research paper on personality recognition using social media data. The paper proposes analyzing personality traits based on the Big Five model (openness, conscientiousness, extraversion, agreeableness, neuroticism) using Facebook status updates and machine learning. Specifically, it involves collecting Facebook status data through a browser extension, storing it in a database, training a random forest regression model on an existing dataset, and using the model to predict personality trait values for additional Facebook users. The predicted traits are then visualized with radar charts to provide an overview of each user's personality profile.
Social networking allows users to interact and connect with other users through dedicated websites and applications. It involves expanding business and social contacts by connecting with people who share similar interests. Popular social networking sites emerged in the 1990s like MySpace and Facebook. While social networking allows users to stay connected with friends and meet new people, it also presents privacy, security, and oversharing risks that users must be aware of. Analyzing social networks can occur at the micro, meso, and macro levels.
Potential vulnerabilities to e-learning - MimecastJisc
Phillip Hay analyzes threats to e-learning. He found that the education sector saw a 87% increase in spam and phishing attacks between March and April 2020 as students shifted to online learning during COVID-19. His recommendations include only using approved school software on school-managed accounts, maintaining online etiquette, and educating students and staff about safe online practices like strong passwords and avoiding suspicious links. Hay warns that cyber attacks will likely continue targeting remote workers and students through increasingly sophisticated social engineering techniques.
Comprehensive Social Media Security Analysis & XKeyscore Espionage TechnologyCSCJournals
Social networks can offer many services to the users for sharing activities events and their ideas. Many attacks can happened to the social networking websites due to trust that have been given by the users. Cyber threats are discussed in this paper. We study the types of cyber threats, classify them and give some suggestions to protect social networking websites of variety of attacks. Moreover, we gave some antithreats strategies with future trends.
A HYBRID CLASSIFICATION ALGORITHM TO CLASSIFY ENGINEERING STUDENTS’ PROBLEMS ...IJDKP
The social networking sites have brought a new horizon for expressing views and opinions of individuals.
Moreover, they provide medium to students to share their sentiments including struggles and joy during the
learning process. Such informal information has a great venue for decision making. The large and growing
scale of information needs automatic classification techniques. Sentiment analysis is one of the automated
techniques to classify large data. The existing predictive sentiment analysis techniques are highly used to
classify reviews on E-commerce sites to provide business intelligence. However, they are not much useful
to draw decisions in education system since they classify the sentiments into merely three pre-set
categories: positive, negative and neutral. Moreover, classifying the students’ sentiments into positive or
negative category does not provide deeper insight into their problems and perks. In this paper, we propose
a novel Hybrid Classification Algorithm to classify engineering students’ sentiments. Unlike traditional
predictive sentiment analysis techniques, the proposed algorithm makes sentiment analysis process
descriptive. Moreover, it classifies engineering students’ perks in addition to problems into several
categories to help future students and education system in decision making.
This document proposes a new progressive model for digital citizenship called ABLE (Attendant, Balanced, Lifelong Learner, Empathetic). It consists of 5 components: being self-aware and other-aware online, balancing technology use in a healthy, moderate way, becoming a lifelong learner of digital skills, and showing empathy and care for others. The model is intended to teach digital citizenship skills to students through strategies like encouraging present moment awareness, moderate tech use, validating information, and developing compassion. It aims to intersect with other areas of teaching to promote well-being, self-expression and academic development through digital tools and relationships.
The document discusses how broadband access through ARE-ON has expanded to nearly all two-year colleges in Arkansas, providing educators access to the world. It outlines exploring how to use broadband through identifying what broadband is, learning about specific broadband details, finding ways to apply broadband, establishing guidelines, and planning next steps. The NETnet consortium connects 16 colleges and universities across Northeast Texas, enabling over 22,000 enrollments per semester through broadband and training over 450 faculty and staff annually while also managing a K-12 network. Broadband applications that could be adopted within a few years are discussed, such as gaming/simulations, online student support, and accessing primary sources.
Designing a Survey Study to Measure the Diversity of Digital Learners IJITE
This article describes the design of a quantitative study that aims to gather empirical data on the different
types of digital learners in a student population, inclusive of the elusive digital natives who purportedly
exist in settings laden with digital technology. The design of this study revolves on the impetus in mapping
the diversity of digital learners, followed by elucidations on the research design and methods that are to be
employed, its accompanying data analysis, ethical considerations and an elaboration of the measures that
are taken in ensuring validity and reliability.
Designing a Survey Study to Measure the Diversity of Digital LearnersIJITE
This article describes the design of a quantitative study that aims to gather empirical data on the different
types of digital learners in a student population, inclusive of the elusive digital natives who purportedly
exist in settings laden with digital technology. The design of this study revolves on the impetus in mapping
the diversity of digital learners, followed by elucidations on the research design and methods that are to be
employed, its accompanying data analysis, ethical considerations and an elaboration of the measures that
are taken in ensuring validity and reliability.
Institutional Success Via a Data-Centric Technology EcosystemIT Consultant
This document discusses Lehman College's development of a data-centric ecosystem to improve recruitment, retention, and other initiatives. It has integrated various tools and systems to expose important information, align messaging and branding, and deliver personalized content. This ecosystem allows events to be syndicated across different platforms, newsletters to automate publishing to websites, and media assets to be shared. Analytics are used to track engagement. The college is also using predictive modeling with data on enrollment, attrition rates, and other factors to develop strategies to improve student outcomes.
Detecting fake news_with_weak_social_supervisionSuresh S
This document discusses using weak social supervision for detecting fake news on social media. It defines weak supervision and describes how social media data provides a new type of weak supervision called weak social supervision. Weak social supervision can be generated from three aspects of social media - users, posts, and networks. Recent works have shown that user-based, post-based, and network-based weak social supervision can be effective for fake news detection, even with limited labeled data. Weak social supervision is a promising approach for learning tasks where labeled data is scarce.
This document discusses using text mining of online social networks to classify user personality traits. It examines using status updates from Facebook profiles as text data to extract features and classify personalities according to the Big Five model using an RBF neural network. The results showed RBF neural networks achieved higher precision than other common classifiers like SVM and Naive Bayes.
Problem statement-1-friend-affinity-finderAmitabhDas22
This document presents a problem statement to build a web/mobile application that can analyze personality traits of users and their friends from social media data in order to find commonalities between them. It involves using IBM Watson services like Personality Insights to infer the Big Five personality characteristics of individuals from text data and identify traits, interests, and preferences that they share. The application would allow authenticating with social media accounts to access profile data and visualize friend connections and classifications. It should be developed using Python and can leverage any mobile platforms.
1) Text messaging among teens has increased significantly since 2006, with more teens both owning cell phones and sending texts on a daily basis.
2) The MyInfoQuest SMS reference project allows librarians to develop core competencies in mobile communication channels by answering patron questions over text message.
3) The project is a collaborative reference service between over 60 libraries that allows patrons to send questions by text message which are then routed and answered by the participating libraries.
Similar to AILABS - Lecture Series - Is AI the New Electricity? Topic:- Interplay of Trust and Risk in Social Media Communication, Presented by - Aditya Bagchi (20)
AILABS - Lecture Series - Is AI the New Electricity? - Advances In Machine Le...AILABS Academy
Prof. Garain discusses in brief on the backgrounds of learning algorithms & major breakthroughs that have been made in the field of machine perception in the last 50 yrs. He also discusses the role of statistical algorithms like artificial neural network, support vector machines, and other concepts related to Deep Learning algorithms.
Along with the above, Prof. Garain touched upon the basics of CNN & RNN, Long Short-Term Memory Networks (LSTM) & attention network & illustrate all of these using real-life problems. Several state-of-the-art problems like image captioning, visual question answering, medical image analysis etc. were discussed to make the potential of deep learning algorithms understandable.
Prof. Utpal Garain is one of the leading minds in Kolkata in the field of Neural Networks & Artificial Intelligence. His research interest is now focused on AI research, especially exploring deep learning methods for language, image and video analysis including NLP tools, OCRs, handwriting analysis, computational forensics and the like.
AILABS - Lecture Series - Is AI the New Electricity? Topic: Intelligent proce...AILABS Academy
In This SlideShare, Dr. Sarit Bose you will learn the Intelligent Processes Propelled By Artificial intelligence.
Decoding AI:-
Deep Learning
Productive Analysis
Translation
classification and clustering
Information Extraction
Speech to text
text to speech
Image recognition
Machine Vision
AILABS - Lecture Series - Is AI the New Electricity? Topic:- Classification a...AILABS Academy
1. The document discusses classification and estimation using artificial neural networks. It provides examples of classification problems from industries like mining and banking loan approval.
2. It describes the basic components of an artificial neural network including the feedforward architecture with multiple layers of neurons and the backpropagation algorithm for learning network weights.
3. Examples are given to illustrate how neural networks can perform nonlinear classification and estimation through combinations of linear perceptron units in multiple layers with the backpropagation algorithm for training the network weights.
AILABS - Lecture Series - Is AI the New Electricity. Topic- Role of AI in Log...AILABS Academy
In this presentation, Professor Deepankar Sinha explains the use and the role of AI in Logistics. Understand the difference between Artificial Intelligence and Automation.
AILABS Lecture Series - Is AI The New Electricity. Topic - Deep Learning - Ev...AILABS Academy
Artificial Intelligence is transforming the world. Deep Learning, an integral part of this new Artificial Intelligence paradigm, is becoming one of the most sought after skills. Learn more about Deep Learning and its Evolution.
The Nano Degree in Deep learning is driving advances in artificial intelligence that are changing our world. Enroll now to build and apply your own deep neural networks to produce amazing solutions to important challenges. Once visit here at - https://ailabs.academy/deep-learning
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
AILABS - Lecture Series - Is AI the New Electricity? Topic:- Interplay of Trust and Risk in Social Media Communication, Presented by - Aditya Bagchi
1. Use and distribution limited solely to authorized personnel. (c) Copyright 2018
Lecture Series: AI is the New Electricity?
Interplay of Trust and Risk in Social Media Communication
Presented at AILABS Academy,
Kolkata on April 27th 2018
Prof. Aditya Bagchi
Emeritus Professor
School of Mathematical Sciences
Ramakrishna Mission Vivekananda Educational & Research
Institute, Belur, West Bengal, India
2. Interplay of Trust and Risk in Social Media
Communication
Aditya Bagchi
Emeritus Professor
School of Mathematical Sciences
Ramakrishna Mission Vivekananda Educational & Research
Institute
Belur, West Bengal, India
AILABS ACADEMY
Kolkata, April 27, 2018
Ramakrishna Mission Vivekananda
Educational & Research Institute
1
3. Ramakrishna Mission Vivekananda
Educational & Research Institute
2
This work is based on the project initiated at the Super
Computer Center, University of California, San Diego, USA
in January 2017. Purpose is to provide quantitative measures
for trust, risk and their interplay while communicating over
social media. Social media involve SMS, e-mail, what’s up,
social nets like Facebook, LinkedIn etc. and even mobile
phones.
Related publication:
Amarnath Gupta, Subhasis Dasgupta, Aditya Bagchi,
“PROFORMA: Proactive Forensics with Message
Analytics”, IEEE Security & Privacy, vol. 15, no. 6, pp. 33-41,
November/December 2017, doi:10.1109/MSP.2017.4251112.
Students involved in the project in India:
Pushpendu Biswas – M.Sc. CS
Anurag Banerjee - M.Sc. CS
Nandish Chattopadhyay – M.Sc. BDA
4. Ramakrishna Mission Vivekananda
Educational & Research Institute
3
We start with a real life event (Names have been changed):
• Medha, a single 35-year old resident of Kolkata, India is
communicating with Mark, resident of Berlin, Germany for
last 6 months over Facebook.
• Personal details exchanged, photos exchanged and
ultimately Mark proposed her and she accepted.
• Marriage date fixed and Mark apparently came to Delhi
and called her over Mobile Phone.
• Medha received a call from Customs Dept at New Delhi
Airport informing that Mark has been taken to custody for
bringing undeclared amount of jewelry.
• Medha being the only connection in India that Mark could
specify, should pay the bail amount in order to release him.
• After about half an hour, Medha received a SMS with bank
account detail for transferring bail amount.
• Medha did and then received no other message either from
Mark or from Customs Dept.
5. Ramakrishna Mission Vivekananda
Educational & Research Institute
4
Most prevalent type of cyber-crime is:
• Fraudulent attack on individuals using different social
media like, e-mail, social network and even mobile
phone.
• The United States Dept. of Justice calls them “Mass-
Marketing Attacks through Social Media” and divided
these fraud schemes into the following categories
(https://www.justice.gov/criminal-fraud/):
• Online-Auction and Online-Retail Schemes
• Business Opportunity or “Work-at-Home” Schemes
• Credit-Card Interest Reduction Schemes
• Inheritance Schemes
• Lottery/Prize/Sweepstakes Schemes
• Online Sales Schemes
• Bank and Financial Account Schemes
• Romance Schemes
6. Ramakrishna Mission Vivekananda
Educational & Research Institute
5
• Usually attack is made on Individuals.
• Communication with the potential victim is made for a long
time to earn his/her confidence.
• The potential victim may be communicated by an individual
or by a company which may even have a web presence.
• After building sufficient trust, the victim is asked to pay or
transfer money for some emergency.
• Sometimes, instead of directly paying any amount, the victim
may be convinced to reveal his/her bank details.
• Crime investigation in this area falls under Digital Forensics.
• Digital forensic science has traditionally been the study of
methods to recover and investigate material found in digital
devices that are examined to solve crimes involving the
computer and the internet.
• A forensic investigation is, by its very nature, a retrospective
activity -- it usually starts after the crime is committed.
7. Ramakrishna Mission Vivekananda
Educational & Research Institute
6
WE NEED PROACTIVE FORENSICS
• Needs both Crime Detection & Crime Prevention.
• A possible crime should be detected early which may be a
false alarm.
• Since crime is organized for a long time, gradual growth
of the crime possibility needs to be measured.
• Since the possible victim is also participating in organizing
the crime, appropriate warning needs to be given to the
possible victim about this gradual growth.
Available Data:
• It is possible to retrieve social media items like Facebook
posts and messages through publicly available APIs.
• With proper access privilege, both current and historical
data (semi-structured) with text content, social circle of an
individual, information entities like URLs and phone
numbers etc. can be collected and analyzed.
8. Ramakrishna Mission Vivekananda
Educational & Research Institute
7
Procedures:
• Significant advances have been made in decision-making
technologies -- including the ability to estimate trust values
of messages and individuals and evolve trust and risk
estimates over time as new events transpire.
• Involves gradual growth of trust on the possible
adversary through increased exchange of messages.
• Gradual growth of risk on the part of possible victim in
revealing sensitive information.
• Measure the interplay of trust and risk .
Subsystems:
• Profile building and matching.
• Social Context building.
• Message aggregation.
• Trust and Risk analysis.
10. Ramakrishna Mission Vivekananda
Educational & Research Institute
9
Building User Profile:
• A user profile is a collection of verified or verifiable facts, i.e.,
data records, about the user (i.e., the person being protected).
These facts are collected using the respective APIs of the social
networks accessible to the system, and are stored in a secure
personal knowledge base. Examples include birthday, current
and previous place and nature of work, academic degrees and
other skills earned etc. In a more aggressive scenario, it can
also include other publicly available facts available from the
Internet, such as the properties bought and sold, current and
prior addresses etc.
• Used for the purpose of building a personal knowledgebase
which is verifiable against any possible inconsistency.
• Can also be used for profile matching against friendship
requests.
• Different parameters may be organized as ontology structures.
11. Ramakrishna Mission Vivekananda
Educational & Research Institute
10
• The term Ontology has been borrowed from Philosophy.
• In Computer Science it is used as Semantically inter-
connected information which may even be hierarchical.
• In Web-2 or Semantic Web, most of the data is expected to
be represented as Ontologies. Examples:
has-son has-son has-son
Grandfather Father Son Grandson
==============================================
Position Name
has-responsibility Person has-name
has-qualification has-address
Qualification Location
is-instance-of
XYZ
13. Ramakrishna Mission Vivekananda
Educational & Research Institute
12
Building of Social Context:
• Social context of a user is a graph (a network); more
precisely, a property graph in which each node and edge has
a type and may have a set of attributes and values.
• Thus the ego network of a user constructed from his/her
friends and followers on Facebook (or any other social
channel) forms a social context.
• In our design, we consider the graph to be a property graph,
which means the nodes and the edges of the graph can have
a set of {attribute, value} pairs associated with them. In these
cases, the nodes of the network can be people, organizations,
places, job positions etc., and the edges represent
relationships like friend, child, colleague, work-institution
and so forth.
15. Ramakrishna Mission Vivekananda
Educational & Research Institute
14
How to build a Social Context Graph (Example: Facebook)
• Analysis with respect to Posts.
• Friends are directly related.
• A B C connected by semantic relationship
friend-of. So in Social Context graph, B has a distance of
1 and C has a distance of 2.
• Number of posts or response to posts provides the
strength of a node within a specific period.
• Posts can be positive or negative.
• Positive or negative responses may be dependent on
aspects. (One may provide +ve response on Football but
–ve response on Music).
• Overall trust may be computed as a composite measure
of +ve and –ve responses and weights assigned to
different aspects.
16. Ramakrishna Mission Vivekananda
Educational & Research Institute
15
Message Aggregation Process: The ability to automatically
detect problematic messages is a fundamental capability of a
proactive forensics system. A message, regardless of whether it
is an email, FB post, SMS message etc., has a core information
structure that consists of:
• metadata about the message including type of data, character
encoding, date and size of message.
• senders and receivers of the message.
• whether the method belongs to a thread and if so, the prior
message it refers to.
• body of the message, that contains semantic references (e.g.,
syntactic tokens like hashtags and user references, lexical
structures like URLs, lexico-semantic structures like phone
numbers and date of birth, as well as semantic entities like
the names of people, organizations and locations.
• auxiliary entities like images, audio and video files.
17. Ramakrishna Mission Vivekananda
Educational & Research Institute
16
The message aggregator translates individual message formats
from different sources into a common internal form that it
stores in a component store specialized for semistructured data
(Asterix DB - UCI). Usually JSON (JAVA Script Object
Notation) format is used.
Applications:
• Initial Trust assignment: Trust-value for a node is assigned
by combining three factors:
1. Extent of communication between the user and the node,
where greater communication (or a close family link)
implies a higher trust.
2. Link strength; the relative importance of the node with
respect to the user based on a measure of common nodes
(mutual friends).
3. Quality of message exchange between two parties.
Individual scores are combined and normalized to a [0,1] scale.
18. Ramakrishna Mission Vivekananda
Educational & Research Institute
17
Trust development between victim and adversary:
• Evaluated against multiple contexts.
• Possible contexts in FB environment –
• Interaction time span
• Number of interactions
• Interaction regularity
• Group membership
• Common interests
• Number of mutual friends etc.
19. Ramakrishna Mission Vivekananda
Educational & Research Institute
18
A profile may also be Fake. Different methods to identify fake
profiles have been developed. Result may be false +ve or –ve.
Ix represents the number of interactions my user is having with
another member x and A is the average number of interactions
among n such members. Computation of trust Tx shows that a
sudden surge of interactions over the average A will reduce the
trust value suspecting anomaly in interaction on the part of
member x.
20. Ramakrishna Mission Vivekananda
Educational & Research Institute
19
Post based friendship measure:
Standard NLP Tools like Stanford CoreNLP and Apache
OpenNLP have been used. Most relevant area is Sentiment
Analysis.
Aspect-wise Categorization:
Posts
Sports Politics Entertainment Product
Review
Humour Others
21. Ramakrishna Mission Vivekananda
Educational & Research Institute
20
If (P/N) > 𝟏 + 𝜺 𝟏 , the post will be considered positive. [ 𝜺 𝟏 ≥ 𝟎 ].
If (P/N) < (𝟏 − 𝜺 𝟐), the post will be considered negative. [ 𝟏 ≥ 𝜺 𝟐 ≥ 𝟎]
If 𝟏 + 𝜺 𝟏 ≥ (P/N) ≥ (𝟏 − 𝜺 𝟐), the post will be considered ambiguous.
23. Ramakrishna Mission Vivekananda
Educational & Research Institute
22
Trust and Risk Interplay:
• Trust in our case is associated with the possible adversary.
• Risk in our case is associated with the possible victim.
• As the risk increases or reliability decreases, trust should
decrease.
• Since the victim is involved in the scam such relationship
may not be there.
• So a method needs to be developed for defining the Trust
and Risk Interplay.
• Risk in communication is present because in the process of
increase in trust, sensitive information may be revealed by
the possible victim.
• So the potential victim should be warned in the gradual
increase in Risk.
24. Ramakrishna Mission Vivekananda
Educational & Research Institute
23
• Both Trust and Risk are defined in the domain of [0, 1].
• A dictionary of sensitive terms may be maintained.
• Each sensitive information may be associated with a cost.
• So, if reliability starts with r = 1, releasing a sensitive
information of cost c in the domain of [0, 1] will reduce the
reliability to r = (1 – c).
• Considering an exponential decay, n occurrences will
reduce reliability to rn .
• Considering trust as an exponential function again, it may
be of the nature (1-αn) for gradual rise.
• Detailed study needed for their interplay.