Twitter’s design encourages hostility and controversy. Here’s what needs to change

It’s about time the platform took another look at the way it surfaces content, and its engagement metrics

Who faces hostility on Twitter?

Hostility on Twitter is disproportionately directed towards women, people of color and marginalized groups. For example, in 2016 US comedianLeslie Joneswasinundated with racist tweetsfollowing the release of the filmGhostbusters.

Black and Indigenous sportspeople, such asAdam Goodes,Glen KamaraandLewis Hamilton, have also been subjected to racist abuse on Twitter and implored the platform to do more to respond to the situation.

More recently,racist tweetsagainst Black English footballers proliferated after the national team’s loss to Italy in the UEFA European Football Championship.

In 2018,Amnesty Internationalpublished a report detailing the extent of abuse directed at female users, describing Twitter as “a toxic place for women”. The report criticised Twitter for failing to respect women’s rights and respond to reports of violence in a transparent manner.

Twitter’s response: how platform design encourages hostile interaction

Part of the reason for the degree of hostility on Twitter is because of the way the platform is designed. Sociologist Ian Hutchby called this the “affordances” of a technology: the material possibilities a technology affords its users, the type of actions it enables and constrains.

Twitter’s affordances shape how users interact on the site. This includes platform features (such as likes, retweets, and mentions), accounts being public by default and the capacity for users to be anonymous. The character limit of tweets also facilitates brief, impulsive, inimical exchange.

In 2017, the company introduced changes to reduce hostility on Twitter. Notable changes included doubling thelength of tweetsfrom 140 to 280 characters. Twitter also introduced “threads” to connect a series of tweets into a longer commentary and provided the option tohide replies. These changes were an attempt to “help minimise unwanted replies and improve meaningful conversations” on the platform, but hostility on Twitter persists.

One reason for the degree of hostility on Twitter is that the site’s metrics can begamed to elevate controversial and abusive content. Research also shows false and misleading news is retweeted more than authentic stories, especially amonglike-minded groups.

In 2018, Twitter launched a “healthy conversation strategy”. This aimed to assess the “health” of interactions on Twitter with a view toimprovethem.

User strategies to avoid hostility on Twitter

In 2019 we conducted an online questionnaire to explore how internet users respond to hostility on Twitter.Our studyfound Twitter users deploy several common strategies to manage hostile interactions on the site.

These include the use of pseudonyms and multiple accounts to achieve a degree of anonymity and privacy, as well as blocking users and self-censoring to pre-emptively limit exposure to harassment and abuse.

Users know they are vulnerable on the platform and artfully manage their social interactions by anticipating hostility, managing the immediate information environment through protecting their tweets, adopting different personas via multiple accounts and limiting how they communicate online.

Saving face on Twitter and minimizing exposure

These observations suggest users are finding ways to “save face” online. The sociologistErving Goffmancalled this kind of activity “face-work”.

In Goffman’s model, weemploy different “faces”to adapt to specific interactions and environments:

The aim of face-work is to create a positive impression of ourselves to others. When we “have face”, we succeed in presenting a consistent self that others validate. In contrast, we “lose face” when information arises that undercuts ourpresentation of self.

Ourresearchextends the idea of face-work to examine the strategies Twitter users employ to interact with others.

We suggest that users adopt a type of “Twitter-face”: a face-work tactic of responding to hostile interactions in a way that will protect the user’s metaphorical face.

Hostile interactions on Twitter often take specific forms, such asdoxing,pile-onsandratioing. In each of these, a user’s face is confronted by coordinated attacks that disrupt the positive impression they are trying to present.

Two kinds of face-work

Face-work generally occurs intwo ways. The first is avoidance, in which people try to avoid face-threatening information or prevent others from seeing it. The second is correction, where people make efforts to apologize for their own face-threatening actions.

We can see an example of corrective face-work on Twitter below, where a person’s face is threatened, they attempt to correct the threatening information, and the conflict is resolved with an apology and acceptance.

Avoidance, on the other, often takes the form of blocking other Twitter users.

Our findings show Twitter users overwhelmingly use avoidance practices as a defensive strategy to prevent hostility on the site. Specific techniques include

Under normal circumstances both avoidance and correction are vital aspects of face-work, but on Twitter there appears to be an overemphasis on avoidance at the expense of correction.

Improving Twitter’s environment

This places Twitter in a difficult situation. Users desire greater control over how they interact, but new features allowing greater control seem to privilege avoidance and may reduce attempts to engage in restorative interactions.

Beyond introducing isolated features, which place responsibility on the individual user, Twitter needs to reconsider the algorithms and metrics (such as likes and retweets) that enable the company to profit from coordinated harassment campaigns, controversy and abuse. This could include hiding likes or removing re-tweets and algorithms.

This article byMichael James Walsh, Associate Professor,University of CanberraandStephanie Alice Baker, Senior Lecturer in Sociology,City, University of London, is republished fromThe Conversationunder a Creative Commons license. Read theoriginal article.

Story byThe Conversation

An independent news and commentary website produced by academics and journalists.An independent news and commentary website produced by academics and journalists.

Get the TNW newsletter

Get the most important tech news in your inbox each week.

Also tagged with

More TNW

About TNW

Twitter’s withdrawal from disinformation code draws ire of EU politicians

Musk claims moderation stifles free speech on Twitter. He’s wrong

Discover TNW All Access

Twitter/X is biggest source of social media disinformation, EU warns

Musk’s beef with Twitter and 3 other tweets that got him in trouble