Professional Documents
Culture Documents
Social Media Bots Infographic 1649181960
Social Media Bots Infographic 1649181960
Social Media Bots use artificial intelligence, big data analytics, and other programs or databases to masquerade
as legitimate users on social media. They vary depending on their function and capability: Some are helpful, like
chat bots and automated notifications, but some can be
used to manipulate real users. When misused, Bots can
amplify disinformation and distort our perception of
what's important, polluting or even shutting down online
conversations.
Recognizing Bot behavior can help us respond
to their attacks.
The Cybersecurity and Infrastructure Security Agency (CISA) produced this graphic to highlight tactics used by disinfor-
mation campaigns that seek to disrupt critical infrastructure in the United States. CISA’s publication of information
materials about this issue are intended for public awareness, and are not intended to restrict, diminish, or demean
any person’s right to hold, express, or publish any opinion or belief, including opinions or beliefs that align with those
of a foreign government, are expressed by a foreign government-backed campaign, or dissent from the majority.
Social Media Bot capabilities have evolved from assisting with simple
online tasks to engaging in more complex behaviors imitating human
users, which bad actors use to manipulate our online interactions.
Visit CISA.gov/MDM to learn more.
Social Media Bots are increasingly integrated into many of our online activities, sometimes without us even
knowing. Bots vary in their functions and capabilities: Some help automate simple tasks, while more advanced
Bots use artificial intelligence, big data analytics, and other programs to mimic human users. Bad actors
sometimes employ Bots as part of coordinated efforts to manipulate human users.
Understanding different Bot uses can help us recognize attempts to manipulate.
Notifications Popularity
Automatically post an update Inflate follower counts and
when a trigger event occurs share posts to boost
perception of influence
Entertainment Harassment
Generate humorous content Overwhelm or ruin reputations
or aggregate news of targeted accounts to the
point of deactivation
Searches Scams
Enable key word searches Phish for personal data or
and detect dangerous activity promote a product
Information Operations
Commerce Spread propaganda to limit
Provide customer care or free speech and manipulate
schedule posts for brands democratic processes
Bad actors seeking to manipulate users on social media often employ different types of Bots as
well as trolls to spread inauthentic content:
Automated Bots run purely Semi-automated Bots allow Trolls are human users,
on programming language a user to program a set of often with obscured
executed without human parameters but require identities, who seek to
management. They can be human management, like fake create division online. Bad actors
purchased to do simple actions and accounts. These “cyborgs” are may employ Bots in coordination
to give the impression of influence. better at evading detection. with trolls.
The Cybersecurity and Infrastructure Security Agency (CISA) produced this graphic to highlight tactics used by disinfor-
mation campaigns that seek to disrupt critical infrastructure in the United States. CISA’s publication of information
materials about this issue are intended for public awareness, and are not intended to restrict, diminish, or demean
any person’s right to hold, express, or publish any opinion or belief, including opinions or beliefs that align with those
of a foreign government, are expressed by a foreign government-backed campaign, or dissent from the majority.
Social Media Bots support coordinated inauthentic behavior by bad actors
and threaten our ability to have important democratic discussions.
Visit CISA.gov/MDM to learn more.
Social Media Bots are often one part of larger inauthentic efforts through which accounts, both human-run and
automated, work in coordination to mislead people. By purchasing or setting up their own Bots, bad actors can
amplify their efforts to spread false or misleading information, shut down opposition, and elevate their own
platforms for further manipulation.
Knowing how Bots support inauthentic activity can help us mitigate their attacks.
Click/Like Farming Astroturfing
Bots inflate popularity by Bots share coordinated
liking or reposting content. content to give a false
The perception of impression that there is
influence online can genuine grassroots support
translate to actual for or opposition to an issue,
influence and distort what making it seem more
really matters. important than it is.
Hashtag Hijacking
Raids
Bots attack an audience by
Bots swarm and overwhelm
leveraging the group’s
targeted accounts with
hashtags (e.g., using spam
spam, harassing the user
or malicious links), silencing
and silencing opposing
opposing opinions and
opinions.
chilling open discussion.
The Cybersecurity and Infrastructure Security Agency (CISA) produced this graphic to highlight tactics used by disinfor-
mation campaigns that seek to disrupt critical infrastructure in the United States. CISA’s publication of information
materials about this issue are intended for public awareness, and are not intended to restrict, diminish, or demean
any person’s right to hold, express, or publish any opinion or belief, including opinions or beliefs that align with those
of a foreign government, are expressed by a foreign government-backed campaign, or dissent from the majority.
Although Social Media Bots try to imitate human users, some
characteristics may be indicators of inauthentic behavior. Recognizing
inauthentic behavior can increase resilience to manipulation.
Visit CISA.gov/MDM to learn more.
6. Coordinated
Network
How to Frequently reposts
Spot a Bot
from other suspicious
1 2 accounts or shares
similar content in
3
coordination with
1. Profile Image 4 other suspicious
May be stolen from real 5 accounts.
users, AI-generated, or a
cartoon, sometimes 7. Sharing
detectable by reverse 6 Reposts most content
image searching. from other users
rather than creating
2. Username original posts, often
Contains suspicious
sharing without stating
numbers and/or 7 an opinion.
irregular capitalization.
8. Viral Content
3. Bio Shares content that
Contains divisive
elicits an emotional
content that appeals to
response and is easily
a target group but
reposted, like memes
contains little personal
and GIFs; spams
information.
targeted hashtags; or
4. Creation Date uses emoticons and
Account was created punctuation in notable
recently or only became 8 patterns.
active recently after a
9. Erratic Behavior
period of dormancy.
Shares content about
5. Followed many unrelated topics
Accounts or changes interests
Account follows a high and behavior
number of other suddenly, such as
accounts to build a randomly posting in a
following and may be new language.
followed by an almost 9
10. Hyperactive
identical, high number Shares a large amount
of accounts (e.g., follow of content, sometimes
10
for follow). nonstop around the
clock or spiking at
certain times.
The Cybersecurity and Infrastructure Security Agency (CISA) produced this graphic to highlight tactics used by disinfor-
mation campaigns that seek to disrupt critical infrastructure in the United States. CISA’s publication of information
materials about this issue are intended for public awareness, and are not intended to restrict, diminish, or demean
any person’s right to hold, express, or publish any opinion or belief, including opinions or beliefs that align with those
of a foreign government, are expressed by a foreign government-backed campaign, or dissent from the majority.