This document provides an overview of social media and big data analytics. It discusses key concepts like Web 2.0, social media platforms, big data characteristics involving volume, velocity, variety, veracity and value. The document also discusses how social media data can be extracted and analyzed using big data tools like Hadoop and techniques like social network analysis and sentiment analysis. It provides examples of analyzing social media data at scale to gain insights and make informed decisions.
While most organizations embrace the idea of Big data, they are yet to figure out how to solve the implications brought about by the big data explosion from social media. In this presentation we highlighted some of the key challenges that organizations face while implementing big data
While most organizations embrace the idea of Big data, they are yet to figure out how to solve the implications brought about by the big data explosion from social media. In this presentation we highlighted some of the key challenges that organizations face while implementing big data
StreamAnalytix is a software platform that enables enterprises to analyze and respond to events in real-time at Big Data scale. It is designed to rapidly build and deploy streaming analytics applications for any industry vertical, any data format, and any use case.
Big data is a term that describes the large volume of data – both structured and unstructured – that inundates a business on a day-to-day basis. But it’s not the amount of data that’s important. It’s what organizations do with the data that matters. Big data can be analyzed for insights that lead to better decisions and strategic business moves.
Machine learning with Big Data power point presentationDavid Raj Kanthi
This is an article made form the articles of IEEE published in the year 2017
The following presentation has the slides for the Title called the
Machine Learning with Big data. that following presentation which has the challenges and approaches of machine learning with big data.
The integration of the Big Data with Machine Learning has so many challenges that Big data has and what is the approach made by the machine learning mechanism for those challenges.
Big Data may well be the Next Big Thing in the IT world. The first organizations to embrace it were online and startup firms. Firms like Google, eBay, LinkedIn, and Facebook were built around big data from the beginning.
What is Big Data?
Big Data Laws
Why Big Data?
Industries using Big Data
Current process/SW in SCM
Challenges in SCM industry
How Big data can solve the problems?
Migration to Big data for an SCM industry
A glimpse into what social media is all about and how the researchers in the world are using social media. Social media is not a mere hype and not a platform to leverage word-of-the-mouth practices as is the common perception of it in Pakistan: it is much more than that and this is what this talk presented.
A Seminar Presentation on Big Data for Students.
Big data refers to a process that is used when traditional data mining and handling techniques cannot uncover the insights and meaning of the underlying data. Data that is unstructured or time sensitive or simply very large cannot be processed by relational database engines. This type of data requires a different processing approach called big data, which uses massive parallelism on readily-available hardware.
It is a brief overview of Big Data. It contains History, Applications and Characteristics on BIg Data.
It also includes some concepts on Hadoop.
It also gives the statistics of big data and impact of it all over the world.
StreamAnalytix is a software platform that enables enterprises to analyze and respond to events in real-time at Big Data scale. It is designed to rapidly build and deploy streaming analytics applications for any industry vertical, any data format, and any use case.
Big data is a term that describes the large volume of data – both structured and unstructured – that inundates a business on a day-to-day basis. But it’s not the amount of data that’s important. It’s what organizations do with the data that matters. Big data can be analyzed for insights that lead to better decisions and strategic business moves.
Machine learning with Big Data power point presentationDavid Raj Kanthi
This is an article made form the articles of IEEE published in the year 2017
The following presentation has the slides for the Title called the
Machine Learning with Big data. that following presentation which has the challenges and approaches of machine learning with big data.
The integration of the Big Data with Machine Learning has so many challenges that Big data has and what is the approach made by the machine learning mechanism for those challenges.
Big Data may well be the Next Big Thing in the IT world. The first organizations to embrace it were online and startup firms. Firms like Google, eBay, LinkedIn, and Facebook were built around big data from the beginning.
What is Big Data?
Big Data Laws
Why Big Data?
Industries using Big Data
Current process/SW in SCM
Challenges in SCM industry
How Big data can solve the problems?
Migration to Big data for an SCM industry
A glimpse into what social media is all about and how the researchers in the world are using social media. Social media is not a mere hype and not a platform to leverage word-of-the-mouth practices as is the common perception of it in Pakistan: it is much more than that and this is what this talk presented.
A Seminar Presentation on Big Data for Students.
Big data refers to a process that is used when traditional data mining and handling techniques cannot uncover the insights and meaning of the underlying data. Data that is unstructured or time sensitive or simply very large cannot be processed by relational database engines. This type of data requires a different processing approach called big data, which uses massive parallelism on readily-available hardware.
It is a brief overview of Big Data. It contains History, Applications and Characteristics on BIg Data.
It also includes some concepts on Hadoop.
It also gives the statistics of big data and impact of it all over the world.
Hadoop was born out of the need to process Big Data.Today data is being generated liked never before and it is becoming difficult to store and process this enormous volume and large variety of data, In order to cope this Big Data technology comes in.Today Hadoop software stack is go-to framework for large scale,data intensive storage and compute solution for Big Data Analytics Applications.The beauty of Hadoop is that it is designed to process large volume of data in clustered commodity computers work in parallel.Distributing the data that is too large across the nodes in clusters solves the problem of having too large data sets to be processed onto the single machine.
I have collected information for the beginners to provide an overview of big data and hadoop which will help them to understand the basics and give them a Start-Up.
Big Data and BI Tools - BI Reporting for Bay Area Startups User GroupScott Mitchell
This presentation was presented at the July 8th 2014 user group meeting for BI Reporting for Bay Area Start Ups
Content - Creation Infocepts/DWApplications
Presented by: Scott Mitchell - DWApplications
Making Internet Of Things Device Data Just Work!Memoori
Making Internet of Things Device Data Just Work! Memoori Talks to John Petze and Marc Petock from Project Haystack - www.project-haystack.org - about the importance of data interoperability and how Project Haystack provides a universal markup language to capture IoT device data semantics.
This presentation contains a broad introduction to big data and its technologies.
Big data is a term that describes the large volume of data – both structured and unstructured – that inundates a business on a day-to-day basis.
Big Data is a phrase used to mean a massive volume of both structured and unstructured data that is so large it is difficult to process using traditional database and software techniques. In most enterprise scenarios the volume of data is too big or it moves too fast or it exceeds current processing capacity.
A short presentation on big data and the technologies available for managing Big Data. and it also contains a brief description of the Apache Hadoop Framework
Real-time analytics is the use of, or the capacity to use, all available enterprise data and resources when they are needed. Big data with real-time analytics consists of 3Vs: the extreme volume of data, the wide variety of types of data and the velocity at which the data must be processed.
Similar to Social media with big data analytics (20)
A disaster is a natural or man-made hazard resulting to physical damage or destruction, loss of life, or drastic change to the natural environment
Disaster Risk Management is a broad range of activities (as opposed to disaster management) designed to prevent the loss of lives, minimize human suffering, inform the public and authorities of risk, minimize property damage and economic loss, and speed up the recovery process
The primary objective of this research is to develop a self-organizing communication model for disaster risk management. The model should be able to provide improved communication services between individuals (or groups) during disasters. The model should be able to offer reduced latency, interruptions, and failures in communication
Spark adds some abstractions and generalizations and performance optimizations to achieve much better efficiency especially in iterative workloads. Yet, spark does not concern itself with being a data file system while Hadoop has what is called HDFS.
Spark can leverage existing distributed files systems (like HDFS), a distributed data base (like HBase), traditional databases through its JDBC or ODBC adaptors, and flat files in local file systems or on a file store like S3 in Amazon cloud.
Hadoop MapReduce framework is similar to Spark in that it uses master slave-like paradigm. It has one Master node (which consists of a job tracker, name node, and RAM) and Worker Nodes (each worker node consists of a task tracker, data node, and a RAM). The task tracker in a worker node is analogues to an executor in Spark environment.
NetworkX is a Python language software package and an open-source tool for the creation, manipulation, and study of the structure, dynamics, and functions of complex networks. NetworkX can load, store and analyze networks, generate new networks, build network models, and draw networks. It is a computational network modelling tool and not a software tool development. The first public release of the library, which is all based on Python, was in April 2005.
Python is a general-purpose, and multi-paradigm dynamic object oriented programming language. Python is a simple, portable, open source, and powerful programming language.
Search engines (e.g. Google.com, Yahoo.com, and Bi
ng.com) have become the dominant model of online search. Large and small e-commerce provide built-in search capability to their visitors to examine the products they have. While most large business are able to hire the
necessary skills to build advanced search engines,
small online business still lack the ability to evaluate the results of their search engines, which means losing the opportunity to compete with larger business. The purpose of this paper is to build an open-source model that can measure the relevance of search results for online businesses
as well as the accuracy of their underlined algorithms. We used data from a Kaggle.com competition in order to show our model running on real data.
unpublished work - full study can be found at researchgate.com.
https://www.researchgate.net/publication/283723818_Scientific_Theory_of_State_and_Society_Parities_and_Disparities_between_the_Philosophical_Thoughts_of_Plato_and_Al-Farabi
-----------------------------------------------------------
While Plato was born and raised in Greece from an aristocratic high-level family few hundred years before the birth of Jesus Christ, Al-Farabi, on the hand, was born in Kazakhstan in Central Asia more than a thousand years later.
In this study, therefore, an attempt would be made to examine how each of the two great scholars imagined his own society through their respective books “The Republic” for Plato and “Opinions of the People of the Ideal City for Al-Farabi.
The Islamic nation is unique, different from other nations and believes on existing based on four basic pillars: Rabania (Law of Allah), Wahda (unity of Muslims), Wasatiyyah (balanced in thinking) and Da’wah (invitation to Allah/Islam).
Studying community structure has proven efficient in understanding social forms of interactions among people, quantitatively investigating some of the well-known social theories and providing useful recommendations to users in communities based on common interests.
Studying the community structure of Flight MH370 will help us finding patterns that emerge from that structure which can lead to demystify some of the many ambiguous aspects of that flight. The aim of this study is to analyze the mesoscopic and macroscopic features of that community using social network analysis.
Pajek, which is a program for social network analysis, is used to generate a series of social networks that represent the different network communities.
Certain modalities (such as text, graphs, tables, and images) can better present recommendations and explanations to users. The focus of this study is the visualization of explanations in recommender systems. The study falls in the area of controlling the recommendation process which gained little attention so far.
Recommender systems are software tools that supply users with suggestions for items to buy. However, it was found that many recommender systems functioned as black boxes and did not provide transparency or any information on how their internal parts work. Therefore, explanations were used to show why a specific recommendation was provided. The importance of explanations has been approved in a number of fields such as expert systems, decision support systems, intelligent tutoring systems and data explanation systems. It was found that not generating a suitable explanation might degrade the performance of recommender systems, their applicability and eventually their value for monetization. Our goal in this paper is to provide a comprehensive review on the main research fields of explanations in recommender systems along with suitable examples from literature. Open challenges in the field are also manifested. The results show that most of the work in the field focus on the set of characteristics that can be associated with explanations: transparency, validity, scrutability, trust, relevance, persuasiveness, comprehensibility, effectiveness, efficiency, satisfaction and education. All of these characteristics can increase the system's trustworthiness. Other research areas include explanation interfaces, over and underestimation and decision making.
In the emerging global economy, e-commerce has increasingly become a necessary component of business strategy and a strong impulse for economic development.
One of the most significant developments in business in recent years has been the increasing use of e-commerce. It has revolutionized many marketplaces and started opportunities never imagined before.
Businesses that are not investigating the use of e-commerce are in a great danger of finding themselves being overtaken by others who are utilizing this technology.
Iraq is characterized as one of the very few countries in the world that are still so far from the real use of inevitable services of e-commerce.
Primarily, Iraq suffers from a number of features that work as barriers to an effective use of e-commerce in life such as the lack of awareness and understanding of the benefit of e-commerce, the lack of information and communication technologies (ICT) knowledge and skills, the unstable physical network infrastructure, security and other privacy-related problems in addition to problems related to costs for the adoption of a new technology.
Senior leaders of the country must be aware of these emerging, and increasingly complex environments of e-commerce in order to compete on a global (or even on a regional) level.
The purpose of this study is two-fold. First, it seeks to investigate the hindrances to e-commerce adoption in Iraq. Second, it seeks to suggest some recommendations for successful applying of e-commerce. The work begins by examining e-commerce, its advantages, its challenges and market models of e-commerce. The remainder of this paper is structured as follows: In the next section, a typical implementation of e-commerce is given, followed by the barriers to e-commerce adoption in developing countries and then barriers to its adoption in Iraq. Later, the conclusions and recommendations are presented and the paper is finalized with limitation of study and future work.
More from Universiti Technologi Malaysia (UTM) (11)
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
Analysis insight about a Flyball dog competition team's performanceroli9797
Insight of my analysis about a Flyball dog competition team's last year performance. Find more: https://github.com/rolandnagy-ds/flyball_race_analysis/tree/main
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfEnterprise Wired
In this guide, we'll explore the key considerations and features to look for when choosing a Trusted analytics platform that meets your organization's needs and delivers actionable intelligence you can trust.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
3. Web 2.0 is
A Complex,
Organic Online
Conversation
WHAT IS WEB 2.0?
Web 2.0 is powered by:
• Social Networks
•News and
Bookmarking
•Blogs
•Microblogging
•Video/Photo-sharing
•Message Boards
•Wikis
•Virtual reality
•Social gaming
•Podcasts
•Real Simple
syndication (RSS)
•Social Media Press
Release
4. TECHNOLOGY OVERVIEW
Search: The ease of finding information through keyword search
Links: Ad-hoc guides to other relevant information
Authoring: The ability to create constantly updating content over a platform
that is shifted from being the creation of a few to being constantly updated,
interlinked work.
Tags: Categorization of content by creating tags: simple,one-word user-
determined descriptions to facilitate searching and avoid rigid, pre-made
categories
Extensions: Powerful algorithms that leverage the Web as an application
platform as well as a documentserver
Signals: The use of RSS technology to rapidly notify users of content changes
Web 2.0 websites typically include some of the following features/techniques-
SLATES
5. Social media:
is an umbrella
term that
defines the
various activities
that integrate
technology,
social
interaction, and
the construction
of words,
pictures, videos
and audio.
WEB 2.0 TECHNOLOGIES:
SOCIAL MEDIA
6. “Creation of web content, by the
people, for the people”
In Simple Language…
9. Variety of sources from where data is being
generated has also undergone a shift
The types of data being created has changed
from structured to semi-structured to
unstructured data
Structured
Data
Semi-
Structured
Data
Unstructured
Data Need to manage broad range of data types
Process analytic queries across numerous data
types
Need to extract meaningful analysis from this
data has led to several technologies to gain
traction
Examples include NoSQL databases to store
unstructured data as well as innovative
processing methods like Hadoop and massive
parallel processing (MPP)
Today 80% Of Data Existing In
Any Enterprise Is Unstructured
Data
Unstructured data from social
media has to be approached in a
non traditional manner.
UNSTRUCTURED DATA
10. Facebook
- User Likes and
Favorites
- Article/Video/Link
Shares
- Views
- Comments
- Location / Geospatial
Twitter
Tweet Characteristics
- Length
- Language Model
- Semantics
- Emoticons
- Location / Geospatial
Google / You Tube
- Blogs
- Comments
- Search Statistics
- Likes vs Dislikes
- Shares / Views /
Comments
IDENTIFYING UNSTRUCTURED DATA
SOURCES
11.
12.
13. “Big Data”
is data whose
scale, diversity,
and complexity
require new
architecture,
techniques,
algorithms, and
analytics to
manage it and
extract value and
hidden knowledge
from it…
BIG DATA IS…
BIG DATA =
15. Implication for an organization
2009 2011 2015 2020
0.8
1.9
7.9
35.0
CAGR
(2009-2020)
41.0%
Zetabytes
THE GLOBAL DATA GROWTH
16. >3,500
>40
>2,000
>200
>400
Key verticals: Healthcare,
Manufacturing, Retail, Digital
Marketing
Demand trend: High demand
of Big Data analytics
>250
Key verticals: Telecom, Retail, Banking
Demand trend: Still embryonic; most
organizations have wait and watch approach
Demand trend: Current demand
appears to be limited, however,
lack of skills may drive
outsourcing of Big Data analytics
Low awareness levels
Key verticals: Technology, Financial services,
Oil & Gas, Utilities, Manufacturing
Demand trend: European MNC’s are still in
the early stages of the adoption cycle
North
America
South America
Europe
Middle East
India
China
Japan
Key verticals: Manufacturing,
Telecom, Health & Life Sciences
Demand trend: Demand for BI
to derive operational efficiency
Key verticals: Telecom, Bioinformatics,
Retail
Demand trend: Industry is in nascent stage
with demand catching up, particularly in retail
>50
16
NORTH AMERICA & EUROPE DRIVES THE BIG DATA
OPPORTUNITY WITH OVER 85%
OF THE WORLD’S DATA
17. Tools Description
The Hadoop
Distributed
File System
(HDFS)
HDFS divides the data into smaller parts and distributes
it across the various servers/nodes
SQL Server
Integration
Service
These tools allow posts can be downloaded and loaded
into Hadoop
Apache
Flume
MapReduce
MapReduce is a process that transforms data loaded
into Hadoop into a format that can be used for analysis.
Hive
a runtime Hadoop support architecture that leverages
Structure Query Language (SQL) with the Hadoop
platform.
Jaql Jaql converts high-level queries into low-level queries
and
Zookeeper Zookeeper coordinate parallel processing across big
clusters
HBase HBase is a column-oriented database management
system that sits on top of HDFS by using a non-SQL
approach.
BIG DATA TOOLS
19. Volume
refers to the vast amounts of
data generated every second.
We are not talking Terabytes
but Zettabytes or Brontobytes.
If we take all the data
generated in the world
between the beginning of time
and 2008, the same amount of
data will soon be generated
every minute.
This makes most data sets too
large to store and analyse
using traditional database
technology.
Variety
Veracity
Value
BIG DATA: VOLUME
20. BIG DATA: VELOCITY
Variety
Veracity
Value
Velocity
refers to the speed at which
new data is generated and
the speed at which data
moves around. Just think of
social media messages
going viral in seconds.
Technology allows us now to
analyse the data while it is
being generated
(sometimes referred to as
in-memory analytics),
without ever putting it into
databases.
21. Variety
Veracity
Value
Variety
refers to the different types
of data we can now use. In
the past we only focused on
structured data that neatly
fitted into tables or
relational databases, such
as financial data. In fact,
80% of the world’s data is
unstructured (text, images,
video, voice, etc.)
BIG DATA: VARIETY
22. Variety
Veracity
Value
Veracity
refers to the messiness or
trustworthiness of the data.
With many forms of big
data quality and accuracy
are less controllable (just
think of Twitter posts with
hash tags, abbreviations,
typos and colloquial speech
as well as the reliability and
accuracy of content) but
technology now allows us to
work with this type of data.
BIG DATA: VERACITY
23. Variety
Veracity
Value
VALUE
Then there is another V to
take into account when
looking at Big Data: Value!
Having access to big data is
no good unless we can turn
it into value.
Companies are starting to
generate amazing value
from their big data.
BIG DATA: VALUE
26. Big Data is also characterized by
velocity or speed i.e. frequency of
data generation or the frequency of
data delivery
New age communication channels
such as mobile phones, emails, social
networking has increased the rate of
information flows
Examples:
Telcos adopting location based
marketing based on user location
sensed by mobile towers
Satellite images can help monitor
and analyze troop movements, a
flood plane, cloud patterns, or forest
fires
Video analysis systems could monitor
a sensitive or valuable facility,
watching for possible intruders and
alert authorities in real time
Big Data velocity enabling real
time use of data
Data
velocity
per
minute
600+
videos on
YouTube
200
million+
emails sent
2
million+
Google
search
queries
400,000+
minutes of
Skype
calling
400,000+
tweets on
Twitter
US$
300,000+
are spent
on online
shopping
700,000+
Facebook
updates
7,000+
photos on
flickr
1,500+
blog posts
3500+
ticks per
minute in
securities
trading
BIG DATA & REAL TIME USE
27. BIG DATA FOR SOCIAL MEDIA ANALYTICS
PROCESS MODEL
28. CONCEPTUAL VIEW OF FRAMEWORK FOR BIG DATA
EXTRACTION, MESSAGING AND STORE
This phase has a composite pattern that is
based on the store-and-explore and focuses on
obtaining and storing the relevant data from
sources outside our establishment.
29. CONCEPTUAL VIEW OF DISCUSSION TOPIC AND
OPINION ANALYSIS COMPONENT
This phase has a composite pattern that is based on
purposeful-and-predictive analytics to gain advanced
insight.
30. WHAT IS HADOOP?
*Hadoop is an open source
framework which is used for
storing and processing the
large scale of data sets on
large clusters of hardware.
*The specialty of Hadoop
involves in HDFS which is used
for storing data on large
commodity machines and
provides very huge bandwidth
for the cluster.
32. CONCEPTUAL VIEW OF DATA VISUALIZATION AND
DECISION-MAKING COMPONENT
This project has a composite pattern based on
actionable-analysis with the aim of taking the next best
actions that leads to take appropriate actions by
related customers.
38. Sentiment analysis…
• Analyzes people’s sentiments,
opinions, appraisals, attitudes,
evaluations, and emotions
• Towards entities such as
organizations, products,
services, individuals, topics,
issues, events, and their
attributes
• As presented online via text,
video and other means of
communication.
• These communications can fall
into three broad categories:
positive, neutral or negative.
SENTIMENT ANALYSIS
39. We can inquire about sentiment at
various linguistic levels:
O Words – objective, positive,
negative, neutral
O Clauses – “going out of my
mind”
O Sentences – possibly multiple
sentiments
O Documents
LEVEL OF ANALYSIS
41. TRUTHY: A SOCIAL MEDIA RESEARCH
PROJECT
Truthy is a research project to study how memes spread on social
media. A meme is a transmissible unit of information, such as a hashtag,
phrase, or link. This website highlights some of the research coming from
this effort and showcases some visualizations, tools, and data resources
demonstrating broader impacts of the project.