Social bot

Source: Wikipedia, the free encyclopedia.

A social bot, also described as a social AI or social algorithm, is a software agent that communicates autonomously on social media. The messages (e.g. tweets) it distributes can be simple and operate in groups and various configurations with partial human control (hybrid) via algorithm. Social bots can also use artificial intelligence and machine learning to express messages in more natural human dialogue.[citation needed]

Uses

  • To persuade people, e.g. to advertise a product, support a political campaign, or boost social media engagement.[1]
  • To offer affordable customer service agents.
  • To provide automatic responses to frequently asked questions on social media platforms like Discord.

Lutz Finger identifies five immediate uses for social bots:[2][clarification needed]

  • foster fame: Simulating real success by having fake followers is unethical.
  • spamming: Having advertising bots in online chats is similar to email spam, but more direct.
  • mischief: For example, one unethical tactic is signing up an opponent with multiple fake identities and spamming the account to discredit them.
  • public opinion bias: Countless messages with similar content but different phrasings have the power to influence fads or trends.[3]
  • limit
    free speech
    : Allow automated bot messages to bury important messages.
  • to phish passwords or other personal data.

Some of another examples, such as:

  • Algorithmic curation: The curation (organizing and maintaining a collection) of online media using computer algorithms. This form of curation has changed how creators & businesses can escape social media algorithms to reach consumers.
  • Algorithmic radicalization: Users are led toward increasingly extreme content, generating polarizing media and self-confirmation of radicalized political views.
  • Influence-for-hire: The term "influencer economy" refers to the buying and selling of influence on social media platforms.
  • Ghost followers: Users on social media who don't interact through likes, comments, messaging, or posts are considered inactive.
  • Social influence bias: This refers to a phenomenon where users tend to overcompensate for negative ratings while amplifying positive ones.

History

Bots have coexisted with computer technology since its creation. Social bots have therefore risen in popularity simultaneously with the rise of social media. Social bots, besides being able to (re-)produce or reuse messages autonomously, also share many traits with spambots concerning their tendency to infiltrate large user groups.[4]

Artificial Social Networking Intelligence (ASNI) refers to the application of artificial intelligence within social networking services and social media platforms. It encompasses various technologies and techniques used to automate, personalize, enhance, improve, and synchronize user's interactions and experiences within social networks. ASNI is expected to evolve rapidly, influencing how we interact online and shaping their digital experiences. Transparency, ethical considerations, media influence bias, and user control over data will be crucial to ensure responsible development and positive impact.

internet personas
that are well capable of influencing real people.

Using social bots is against the terms of service of many platforms, such as Twitter and Instagram, although it is allowed to some degree by others, such as Reddit and Discord. Even for social media platforms that restrict social bots, a certain degree of automation is of course intended by making social media APIs available. Social media platforms have also developed their own automated tools to filter out messages that come from bots, although they are not advanced enough to detect all bot messages.[5]

The topic of legal regulation of social bots is becoming more urgent to policy makers in many countries, however, due to the difficulty of recognizing social bots and separating them from "eligible" automation via social media

spear-phishing attempts which are why there has been an emergence of urgency to create more research, policies, and detection of bots on the many platforms that they affect.[6]

Detection

The first generation of bots could sometimes be distinguished from real users by their often

pattern detection techniques, some of which are:[8]

  • cartoon figures as user pictures
  • sometimes also random real user pictures are captured (identity fraud)
  • reposting rate
  • temporal patterns[9]
  • sentiment expression
  • followers-to-friends ratio[10]
  • length of
    user names
  • variability in (re)posted messages
  • engagement rate (
    followers
    rate)
  • analysis of the time series of social media posts[11]

Social bots are always becoming increasingly difficult to detect and understand, some of the greatest challenges for the detection of bots include:

datasets, detect the bots' human-like behavior in the wild, ever-changing behavior of the bots, lack of appropriate visualization tools and the sheer volume of bots covering every platform.[12]

Botometer[13] (formerly BotOrNot) is a public Web service that checks the activity of a Twitter account and gives it a score based on how likely the account is to be a bot. The system leverages over a thousand features.[14][15] An active method that worked well in detecting early spam bots was to set up honeypot accounts where obvious nonsensical content was posted and then dumbly reposted (retweeted) by bots.[16] However, recent studies[17] show that bots evolve quickly and detection methods have to be updated constantly, because otherwise they may get useless after a few years.

One method still in development, but showing promise is the use of Benford's Law for predicting the frequency distribution of significant leading digits to detect malicious bots online. This study was first introduced at the University of Pretoria in 2020 and had successful trials in the field.[18]

Another method that has also proven to be quite successful in research and in the field is artificial-intelligence-driven detection which simply put, evens the playing field when putting artificial intelligence against itself. Some of the most popular sub-categories of this type of detection would be active learning loop flow, feature engineering, unsupervised learning and outliers identification, supervised learning, correlation discovery, and system adaptability.[12]

An important mode of operation of bots is by working together in a synchronized way. For example, ISIS used Twitter to amplify its Islamic content by numerous orchestrated accounts which further pushed an item to the Hot List news,[19] thus further amplifying the selected news to a larger audience.[20] This mode of synchronized bots accounts is an efficient method to further spread a desired news and is also used as a modern tool of propaganda as well as stock markets manipulations.[21]

Research and development to detect malicious bots continue to be an important topic throughout the tech world. Social media sites like Twitter, which are among the most affected with CNBC reporting up to 48 million of the 319 million users (roughly 15%) were bots in 2017, continue to fight against the spread of misinformation, scams and other harmful activities on their platforms.[22]

Platforms

Instagram

Instagram reached a billion active monthly users in June 2018,[23] but of those 1 billion active users it was estimated that up to 10% were being run by automated social bots. Instagram's unique platform for sharing pictures and videos makes it one of the biggest targets for malicious social bot attacks, especially porn bot accounts,[24] because imagery resonates with the platform's users more than simple words on platforms like Twitter.[25] While malicious propaganda posting bots are still popular, many individual users use engagement bots to propel themselves to a false virality, making them seem more popular on the app. These engagement bots can do everything from like, watch, follow, and comment on the users' posts.[26] Around the same time that the platform achieved the 1 billion monthly user plateau, Facebook (Instagram and WhatsApp's parent company) planned to hire 10 000 to provide additional security to their platforms, this would include combatting the rising number of bots and malicious posts on the platforms.[25] Due to increased security on the platform and enhanced detecting methods by Instagram, some botting companies are reporting issues with their services because Instagram imposes interaction limit thresholds based on past and current app usage and many payment and email platforms deny the companies access to their services, preventing potential clients from being able to purchase them.[27]

Twitter

Twitter's bot problem is being caused by the ease of use in creating and maintaining them. To create an account you must have a phone number, email address, and CAPTCHA recognition. The ease of creating the account as and the many APIs that allow for complete automation of the accounts are leading to excessive amounts of organizations and individuals using these tools to push their own needs.[22][28] CNBC claiming that about 15% of the 319 million Twitter users in 2017 were bots, the exact number is 48 million.[22] As of July 7, 2022, Twitter is claiming that they remove 1 million spam bots on their platform each and every day.[29] Twitter bots are not all malicious, some bots are used to automate scheduled tweets, download videos, set reminders and even send warnings of natural disasters.[30] Those are examples of bot accounts, but Twitter's API allows for real accounts (individuals or organizations) to use certain levels of bot automation on their accounts, and even encourages the use of them to improve user experiences and interactions.[31]

See also

References

  1. ^ "The influence of social bots". www.akademische-gesellschaft.com. Retrieved March 1, 2022.
  2. ^ Lutz Finger (February 17, 2015). "Do Evil - The Business Of Social Media Bots". forbes.com.
  3. ^ Frederick, Kara (2019). "The New War of Ideas: Counterterrorism Lessons for the Digital Disinformation Fight". Center for a New American Security. {{cite journal}}: Cite journal requires |journal= (help)
  4. S2CID 1914124
    .
  5. ^ Efthimion, Phillip; Payne, Scott; Proferes, Nicholas (July 20, 2018). "Supervised Machine Learning Bot Detection Techniques to Identify Social Twitter Bots". SMU Data Science Review. 1 (2).
  6. S2CID 51877148
    .
  7. .
  8. .
  9. .
  10. ^ "How to Find and Remove Fake Followers from Twitter and Instagram : Social Media Examiner".
  11. S2CID 253359560
    .
  12. ^ .
  13. ^ "Botometer".
  14. .
  15. ^ Varol, Onur; Emilio Ferrara; Clayton A. Davis; Filippo Menczer; Alessandro Flammini (2017). "Online Human-Bot Interactions: Detection, Estimation, and Characterization". Proc. International AAAI Conf. on Web and Social Media (ICWSM).
  16. ^ "How to Spot a Social Bot on Twitter". technologyreview.com. July 28, 2014. Social bots are sending a significant amount of information through the Twittersphere. Now there's a tool to help identify them
  17. S2CID 10464463
    .
  18. .
  19. .
  20. .
  21. .
  22. ^ a b c Newberg, Michael (March 10, 2017). "As many as 48 million Twitter accounts aren't people, says study". CNBC. Retrieved November 22, 2022.
  23. ^ Constine, Josh (June 20, 2018). "Instagram hits 1 billion monthly users, up from 800M in September". TechCrunch. Retrieved November 24, 2022.
  24. S2CID 204107862
    .
  25. ^ a b "Instagram's Growing Bot Problem". The Information. July 18, 2018. Retrieved November 24, 2022.
  26. ^ "Instagram Promotion Service (Real Marketing) – UseViral". August 15, 2021. Retrieved November 24, 2022.
  27. ^ Morales, Eduardo (March 8, 2022). "Instagram Bots in 2021 — Everything You Need To Know". Medium. Retrieved November 24, 2022.
  28. S2CID 33003478
    .
  29. ^ Dang, Sheila; Paul, Katie (July 7, 2022). "Twitter says it removes over 1 million spam accounts each day". Reuters. Retrieved November 23, 2022.
  30. ^ Reply, Huzaifa Azhar 2 months ago. "10 Best Twitter Bots You Should Follow in 2022 - TechPP". techpp.com. Retrieved November 24, 2022.{{cite web}}: CS1 maint: numeric names: authors list (link)
  31. ^ "Twitter's automation development rules | Twitter Help". help.twitter.com. Retrieved November 24, 2022.

External links