Written evidence submitted by Reset (OSB0138)
Content in scope
Definition of harm
Enforcement
---
“One of the biggest issues social networks face is that, when left unchecked, people will engage disproportionately with more sensationalist and provocative content. [...] At scale it can undermine the quality of public discourse and lead to polarization.”[6]
The graph accompanying his blog post (Fig. 1) visualises this reality.
Fig. 1
Zuckerberg goes on to assert in this blog post (without evidence) that Facebook corrects for this human tendency to engage with extreme, disturbing, and provocative content by prioritising healthier conversations. Recent revelations from a whistleblower demonstrate that his public statements do not match his internal decisions. When presented with options for taming the artificial amplification of extremism in the NewsFeed, Zuckerberg himself rejected the changes so as not to reduce engagement, growth and revenue.[7]
Because we were looking at each entity individually, rather than as a cohesive movement, we were only able to take down individual Groups and Pages once they exceeded a violation threshold. We were not able to act on simple objects like posts and comments because they individually tended not to violate, even if they were surrounded by hate, violence, and misinformation.[23]
----
Below are some examples of how technology companies have introduced design features to reduce the amplification of harmful content. These changes were made after intense public campaigns of shaming the platforms with open letters, petitions, media pressure and forcing Twitter and Facebook to meet victims of harm. This further underscores why we need regulation of legal but harmful content, for the government to require these design changes and be able to assess their impact. Otherwise they will continue to happen in a piecemeal way, with campaigners and researchers spending years and millions of pounds for tweaks that the platforms can make happen overnight if they were required to by law. As the below examples demonstrate, tech firms do indeed have the agility and insights to stem the spread of harmful material at pace and at scale.
In 2020, Facebook announced that it would introduce a notification screen warning users if they try to share content that’s more than 90 days old. They’ll be given the choice to “go back” or to click through if they’d still like to share the story knowing that it isn’t fresh.
Facebook acknowledged that old stories shared out of their original context play a role in spreading misinformation. The social media company said “news publishers in particular” have expressed concern about old stories being recirculated as though they’re breaking news.
In a recent blog post, Twitter announced the trial of a new product feature that temporarily autoblocks accounts using harmful language, such that they’re stopped from being able to interact with a user’s account. In the post, Twitter states:
We are also continuing to roll out our replies prompts, which encourage people to revise their replies to Tweets when it looks like the language they use could be harmful. We found that, if prompted, 34% of people revised their initial reply or decided not to send their reply at all and, after being prompted once, people composed on average 11% fewer potentially harmful replies in the future.
In the days following the 2020 US Presidential election, misinformation about the election results flooded social media. In response, Facebook made a temporary change to its News Feed algorithm to give prominence to information from mainstream media outlets. To achieve this, Facebook dialled up the weighting of its “news ecosystem quality” (NEQ) score, a ranking Facebook assigns to news outlets based on signals about the quality of their journalism. According to internal sources at Facebook, the NEQ score usually plays a minor role in determining News Feed content, but concerns over the nature and scale of election disinformation drove senior executives including Mark Zuckerberg to temporarily increase NEQ’s weighting. This resulted in a spike in visibility for mainstream news outlets.
This intervention is another example of how design choices can be made to reduce the reach of harmful material, as well as counter false information with that which is more verifiable. While it would undoubtedly be preferable for an independent regulator to determine which content is harmful, rather than a tech platform, this approach demonstrates that companies can respond at pace when focused on harm reduction, and that such design choices are already available to them.
We were also able to add friction to the evolution of harmful movements and coordination through Break the Glass measures (BTGs). We soft actioned Groups that users joined en masse after a group was disabled for PP or StS, this allowed us to inject friction at a critical moment to prevent growth of another alternative after PP was designated, when speed was critical. We were also able to add temporary feature limits to the actors engaging in coordinating behaviors, such as the super posters and super-inviters in the Groups that were removed, to prevent them from spreading the movement on other surfaces. These sets of temporary feature limits allowed us to put the breaks on growth during a critical moment, in order to slow the evolution of adversarial movements, and the development of new ones. Our ongoing work through the disaggregating networks taskforce will help us to make more nuanced calls about soft actions in the future in order to apply friction to harmful movements.
In April 2020, TikTok removed its direct messaging features for users under the age of 16 in an attempt to reduce the amount of grooming and bullying taking place in close conversations. It was the first time a major social-media platform has blocked private messaging by teenagers, on a global scale.
During the first wave of the Covid-19 pandemic, WhatsApp sought to address the “infodemic” by imposing a limit on message forwarding to slow the spread of mis and disinformation. Any frequently forwarded message: ie: forwarded more than five times, would get slowed down with users able to only forward that on to one user at a time.
YouTube also took steps to address the massive spread of Covid-19 conspiracy videos on its site by prominently placing authoritative sources on the top of the search page. For instance, a search for “The truth about Covid-19” has a link at the top to an official NHS source.
Most of the main platforms use warning messages to inform users about sensitive content. These messages are overlaid on specific Tweets or Posts, warning users about the nature of the content and requiring them to click through before they can view it. The messages stay on the site - content is not removed.
These are generally applied to content that has been marked (either by the person Tweeting it or following reports by other users) as “sensitive”, such as media that includes adult content, excessive gore or violence. This reduces the risk of users inadvertently witnessing content they might find harmful or distressing, but allows users who do want to find such content to access it. Users can choose whether to turn this feature on/off, so they don’t have to click through to view sensitive content.
In June 2020, Twitter applied for the first time its “public exemption policy”. The policy states that when a Tweet contains harmful content but is deemed to be in the public interest, the Tweet will be placed behind a notice. Such content would include harassment or hateful conduct, content which is in breach of Twitter’s T&Cs and for the majority of users would have to be taken down. Instead, in such instances, the notice would be applied which still “allows people to click through to see the Tweet” but “limits the ability to engage with the Tweet through likes, Retweets, or sharing on Twitter, and makes sure the Tweet isn't algorithmically recommended by Twitter”. This is an example of what it means to protect free speech while challenging unlimited reach. The exception only applies to elected or government officials with over 100,000 followers, and aims to “limit the Tweet’s reach while maintaining the public’s ability to view and discuss it”.
Reset has worked with two leading experts at MIT and University of Regina in Canada to pursue an empirically grounded strategy to fight mis and disinformation. Their academic research distilled in this paper, Lazy not biased, found that most people actually share misinformation out of laziness and lack of friction in platform design, rather than inherent bias. They tested whether a simple reminder about the importance of accuracy in the social media feed (content neutral and non-partisan nudges or “accuracy prompts”) could yield an increase in cognitive discernment of true/false content to deter a significant percentage of harmful sharing of disinformation.
Using a random sample of 3000 likely US voters (studies were also carried out in France, Italy and Canada with similar results), they found that the “accuracy prompt” Facebook ad shows the potential to change the proportion of true and false content shared on social media by 7-9%. (This testing was done via YouGov). That may seem a modest change but for highly vulnerable groups (eg. those that share disinformation with great frequency) we are moving from a baseline of users that share disinformation as often or even more often as they do accurate information to a 7% advantage for accurate information. This represents as much as a 14% decrease in disinformation sharing.
----
The Online Safety Bill is a pioneering piece of legislation which aims to improve online safety for individual users, in particular for children. The companies in scope are predominantly the Big Tech platforms where the vast majority of harmful content is posted and witnessed. These platforms offer users unprecedented connectivity to friends, family, public figures, strangers, authorities and much more. They exist to bind us together and forge connections.
These are some of the wealthiest firms in the history of the world, whose primary source of revenue is advertising. The more time people spend online, the more ads they can sell. It is in their interest to keep users logged-on so that they stay clicking. To do this, platforms serve up engaging content and that which most grabs our attention tends to be sensationalist, subversive and controversial. Algorithms, which are trained to drive engagement at the expense of nothing else, push content and network recommendations which we wouldn’t necessarily seek out or find ourselves, but which we can’t resist. For example, an internal report by Facebook in 2016 found that 64% of all extremist group joins on the platform were due to recommendation algorithms.[30]
The attraction to sensationalist content is nothing new. It’s human nature. But in our digital world, and on large technology platforms, the pace and scale at which such content is witnessed is unprecedented. We are encouraged to push this content - be it on Pages, comments, stories or posts - with and beyond our network, while algorithms work furiously to secure engagement. We are rarely encouraged to assess the content we share for accuracy or potential harm, if we are encouraged to read it at all. This creates false levels of trust in content, which is given more credence by virtue of being shared by trusted contacts. Groups allow people to share perspectives with huge audiences with a single click, allowing new levels of reach which are impossible for individuals to replicate offline. Third parties can interact with, or insert themselves into, these networks with ease. The frictionless system ensures content spreads at speed and scale, feeding the recommendation algorithms with as much data as possible in order to improve targeted content. This perpetuates a vicious cycle in which the individual becomes indistinguishable from their network.
Collective harm
The result is an environment where harmful content becomes normalised - where what was once at the fringes becomes mainstream, dominating Feeds and Posts. Covid disinformation brought that into sharp relief. Research by the Institute for Strategic Dialogue (ISD) showed that Covid disinformation had twelve times the volume of interactions on disinformation sites as credible sites like the WHO and CDC.[31] When factual information relating to a highly contagious virus is drowned out by disinformation, this becomes a collective harm. It damages society at large, through the erosion of trust in official bodies and the growth of a culture where conspiracy theories are mainstream, as well as individuals. Separate research by ISD showed that false claims of voter fraud in the US Presidential election were shared at twice the rate of the best performing official pieces of information.[32] The resulting confusion and aggression this stirred up played out in a very public way. Facebook’s internal analysis of its role in the US election noted that categorising electoral disinformation campaigns “as a network allowed [them] to understand coordination in the movement and how harm persisted at the network level. This harm was more than the sum of its parts.”[33]
Similarly, with the recent sexual abuse scandal in UK schools, the Everyone’s Invited movement exposed how attitudes towards girls are being shaped by the normalistion of pornography and the volume at which pornographic material is viewed online.[34] Authorities have had to remind teenage boys that the pornographic material they watch online is fictional. When such boundaries become blurred, the result is both a very real harm for individual victims as well as a broader collective harm to society.
These are not isolated examples. Racist abuse, misogyny, climate change denial, homophobia, bullying, pro-suicide and self-harm and many other types of harmful content are as readily promoted to the widest possible audience, propagating harm at a network level. Often, this promotion of harmful material is coordinated by malicious groups with malign intent, as it is now far easier to target, access and influence groups online than it is offline. But just as problematic is how the business model of Big Tech prioritises content which will spread over that which is safe, nudging users to engage with harmful content at any cost. Harms to society often stem from harms to individuals, and vice versa. One cannot be considered distinct from the other.
Online Safety Bill
The definition of harm in the draft Bill is limited to content which has “a significant adverse physical or psychological impact” on an individual of “ordinary sensibilities”.[35] This differs from the language in the Online Harms White Paper which proposed “prioritising regulatory action to tackle harms that have the greatest impact on individuals or wider society.”[36] The narrower focus on individuals rather than particular demographics, groups or society as a whole fails to reflect the nature of digital technologies which forge connections, groups, networks and communities. Ignoring this leaves vast numbers of users, including children and vulnerable people, exposed to manipulation, abuse and bullying at a worrying scale - both online and offline. It means platforms will continue to distort public discourse, as well as long-held social and cultural norms. The Online Safety Bill should account for the fundamental networking principles of online platforms, which promote connections and groups, and ensure that any legislation prevents harm at a much broader level.
----
Below are a number of real world examples of disinformation which would not cross the illegality threshold and so would only be in scope of the OSB if they met the definition of “content that is harmful to adults”. This means, to be in scope, “the nature of the content is such that there is a material risk of the content having, or indirectly having, a significant adverse physical or psychological impact on an individual of ordinary sensibilities”.
It is unclear which of the below would reach that threshold. It could certainly be argued that all of these result in direct or indirect harm, either physical or psychological, to individuals. Many of them have a direct societal harm, however of course that is not included in the definition at present. It is a useful exercise to demonstrate where the definition creates uncertainty and where serious instances of disinformation may be free to circulate online without recourse.
| Headline | Example | Comments |
1 | State backed disinformation campaigns inauthentically amplifying partisan views on Scottish referendum
| Twitter and Facebook have identified swathes of fake accounts linked to the governments of Russia and Iran which amplify messages from pro-independence campaigners. Facebook has also identified, and deleted, a page called Free Scotland 2014 which was traced back to Iran and linked to fake news sites. | Unclear whether such campaigns would be captured under the current definition. Could be argued that encountering such information directly impacts views and outlooks; as well as indirectly pro-Unionists by damaging their cause and, by extension, the Union. Unlikely to qualify as “significant”. |
2 | Coordinated disinformation campaign against journalists. | This example of a disnfo campaign against an AP journalist, which resulted in her losing her job, is just one of the rising number of cases of journos being subject to disinfo campaigns. In this instance, the journalist was targeted by a political group who mischaracterized her old tweets to portray her as antisemitic. She lost her job as a result. Recently, Christina Lamb, Chief Foreign Correspondent for The Times, found herself in a similar situation but hept her job. And the reporter Jon Snow was physically assaulted by antivax campaigners who had encountered conspiracy theories online. | Disinfo campaigns against journalists are on the rise. Responses are generally a matter of internal policy for publishers but it is increasingly difficult for news outlets to distinguish between legitimate public reaction and coordinated attacks. It is unclear whether journalists are protected against such targeted campaigns by the OSB, and how the protections for journalists marry with the definition of harm. The attacks on Jon Snow and Nicholas Watt are examples of how online campaigns translate into offline harm. |
3 | Climate change | There are plenty of examples of disinformation campaigns sowing confusion and outright lies online about scientifically proven climate change realities. This includes false assertions about the evidence proving climate change, conspiracy theories about politicians calling for action and fake statistics about the environmental impact. | False information about climate change directly and indirectly impacts behaviour, health and wellbeing. In the medium to long term, the impact is highly significant and adverse. |
4 | The role of coordinated disinformation campaigns in loss of life and livelihood | The British co-founder of the White Helmets, a Syrian civil defence group, committed suicide following disinformation campaigns run by Russian and Syrian propagandists claiming the White Helmets faked evidence of Syrian atrocities. | An example of how a deluge of disinformation can have a direct and indirect detrimental psychological impact. |
5 | Islamophobic conspiracy theories (Eurabia and “No Go Zones”) inciting violence and racial tensions. Anders Brevik, who carried out atrocities in Norway, was a follower of such beliefs. | Widely debunked far-right conspiracy theories about Islam run rife on social media and news media sites/blogs. Examples include a deal between Western governments and the Middle East to exchange Western sovereignty for oil; a plot by Islamic nations to take over Europe to create “Eurabia”; claims of “No Go Zones” in Western nations which are run by Sharia law and bar non-Muslims and police. | Another national security issue linked to disinformation and fake news. Issues around how far-right conspiracy sites are categorised vis a vis “news publishers” and whether their content would be deemed out of scope. |
6 | Families, children and conspiracy theories | Sites and movements such as QAnon risk having an indirect adverse effect on families whose members start to believe anti-vax disinformation or wild conspiracy theories about e.g. the political elite. There are increasing examples of families torn apart by such movements, with the impact on children whose parents start to distrust authorities and remove their children from educational services. | While certainly indirect harm, the threshold of significance makes it unclear whether this would be in scope. Such disinformation has a long-tail, with the impact of children being removed from educational and healthcare services having a lasting - but perhaps not immediate - effect. |
7 | Manipulated video image of political candidates | Deep Fake Video of Speaker Nancy Pelosi, appears to show her slurring and impaired. The video was viewed more than 2.5 million times on Facebook and Facebook refused to take it down, despite clearly being a deep fake.
A deepfake video of Boris Johnson endorsing Jeremy Corbyn, created for academic and campaigning purposes, shows how else this technology can be applied. | An example of how deep fakes can be used to manipulate videos and messages. More advanced technology could have more severe implications. |
8 | Disinformation about the sexuality of footballers | In addition to the racial abuse received by footballers and their families, many are also on the receiving end of disinformation about their sexuality. Most recently, Premier League player James Rodriguez was the subject of false rumours claiming that he had not played because he had undergone gender reassignment surgery. The false information involved people mocking up images saying he would never play again. |
|
9 | Coordinated campaigns to stir-up racial tensions ahead of the 2016 presidential election | According to the US Senate Intelligence Committee, disinformation campaigns by Russia in the US 2016 Presidential election targeted African-American communities more than any other group.
“By far, race and related issues were the preferred target of the information warfare campaign designed to divide the country in 2016”.
This included false claims about police brutality, political party funding (such as Hillary Clinton received $ from the Klu Klux Klan) and about the BLM movement.
| Disinfo campaigns by state actors are far more nuanced than simply targeting individuals or groups of individuals with abusive content. In this case, the aim was to sow confusion and division in order to subtly change opinions and reduce voter engagement. This has major democratic implications. It could be argued that such a campaign fits the definition of harm, but it is the nature and scale of the content rather than the content itself which causes the harm. May be out of scope. |
10 | Coordinated response to discredit Sergei Skripal poisoning | Campaign by Russian officials and others calling into question the authenticity of images published by HMG, and promoting #skripalhoax
| Disinformation had no bearing on the poisoning itself but rather on post hoc public perception of the event and whether/how it took place. Weakens trust in security services and sows confusion, but does not clearly fulfil the definition. |
11 | US electoral fraud | Perhaps the most high profile conspiracy in modern democracies is the evidence-free claim that the 2020 elections were stolen and the results therefore illegitimate. This conspiracy is widely disseminated on social media and echoed by mainstream media and prominent political leaders in whole or in part. As a result, a substantial percentage of the electorate believes the election was fraudulent and has lost confidence in electoral institutions. | This is a particularly pernicious example because the disinformation is centrally about politics and elections in the most consequential way possible. Would it be subject to protection due to “democratic importance”? The platforms have all changed terms of service to prohibit election delegitimization; but they are failing to achieve that objective. Should the regulator insist they do? Or insist they shouldn’t?
Consider the possibility that this scenario occurs in a future Scottish referendum. |
13 | UN Global Compact on Migration | In December 2018, The UN Global Compact on Migration was subjected to a wave of misinformation from over 50,000 Twitter users in multiple languages, as well as manipulation of the YouTube recommendation algorithm by far-right outlets. The misleading content pushed the far right Great Replacement Theory and spread lies about the outcomes of the Compact. Countries including Brazil, the USA, Austria, Bulgaria, the Czech Republic, Slovakia, Poland, Australia and Israel pulled out. Later, the barrel of the Christchurch killer’s gun was found with the chilling inscription: ‘Here’s your global compact on migration.’
| Given this has evidently resulted in indirect physical impact (the Christchurch massacre) it feels this should qualify. However it’s unlikely such campaigns would be picked up before or unless a mass murder took place. |
Reset (www.reset.tech) was launched in March 2020 by Luminate in partnership with the Sandler Foundation. Reset seeks to improve the way in which digital information markets are governed, regulated and ultimately how they serve the public. We will do this through new public policy across a variety of areas – including data privacy, competition, elections, content moderation, security, taxation and education.
To achieve our mission, we make contracts and grants to accelerate activity in countries where specific opportunities for change arise. We hope to develop and support a network of partners that will inform the public and advocate for policy change. We are already working with a wide variety of organizations in government, philanthropy, civil society, industry and academia.
27 September 2021
[1] Online Nation – 2020 report, p37, Ofcom, 2020
[2] Facebook Executives Shut Down Efforts to Make the Site Less Divisive, Wall Street Journal, 26 March 2020
[3] Facebook Tried to Make its Platform a Healthier Place, Wall Street Journal, 15 September 2021.
[4] YouTube Regrets, Mozilla Foundation, July 2021
[5] Inside TikTok’s Algorithm, Wall Street Journal, 21 July 2021.
[6] A blueprint for governance and enforcement, blog post, Mark Zuckerberg, 2018 (updated 2021)
[7] Facebook Tried to Make its Platform a Healthier Place, Wall Street Journal, 15 September 2021.
[8] Facebook Struggles to Balance Civility and Growth, New York Times, 24 Nov 2020
[9] Facebook Knows Instagram is Toxic for Teenage Girls, Wall Street Journal, 14 September 2021.
[10] Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That’s Exempt. Wall Street Journal, 13 September 2021.
[11] Algorithms in social media: realistic routes to regulatory inspection
[12] We Research Misinformation on Facebook. It Just Disabled Our Accounts.
[13] See, for example, this investigation by the Süddeutsche Zeitung examining Facebook’s haphazard compliance with the German hate speech law.
[14] Anti-vaxxers harass Jon Snow as they storm ITN headquarters, The Telegraph, 23 August 2021
[15] Facebook shuts fake Scottish independence accounts | Scotland, The Times, 6 March 2021
[16] Facebook's Climate of Deception: How Viral Misinformation Fuels the Climate Emergency, Avaaz
[17] Sergei Skripal and the Russian disinformation game, BBC News, September 2018
[18] Here's where those 5G and coronavirus conspiracy theories came from, FullFact
[19] Director General Ken McCallum gives annual threat update 2021
[20] Carbis Bay G7 Summit Communique (PDF, 430KB, 25 pages)
[21] The New Atlantic Charter 2021
[22] Online Harms White Paper - GOV.UK
[23] Facebook Stopped Employees From Reading An Internal Report About Its Role In The Insurrection. You Can Read It Here.
[24] Online Safety Bill: Five thoughts on its impact on journalism, LSE Media Blog, June 20201
[25] Online Safety Bill: Explanatory Notes,
[26] Black and Asian women MPs abused more online, Amnesty International
[27] Facebook: Smoking and alcohol ads 'target Australian children'.
[28] The Big Tech business model poses a threat to democracy
[29] Revealed: Trump campaign strategy to deter millions of Black Americans from voting in 2016
[30] Facebook Executives Shut Down Efforts to Make the Site Less Divisive
[31] Public Health Disinformation - ISD
[32] Narratives of Violence around the 2020 Presidential Election - ISD
[33] Facebook Stopped Employees From Reading An Internal Report About Its Role In The Insurrection. You Can Read It Here., Buzzfeed, April 2021