Written evidence submitted by Digital Net Identity UK Ltd. (OSB0143)
Objectives
Digital Identity Net (DIN) believe that the Bill’s approach to introduce accountability for content on user-to-user platforms via legislation, combined with technical solutions for improving the online environment, is a good approach and will achieve the policy aim to make the UK the safest place to be online. Digital identity solutions enable users to easily prove who they are online, and can be used to set up new accounts or verify existing accounts. These solutions can support ‘real name displayed’ and pseudonymous accounts (where the platform has traceability to the verified real user in cases of illegal content).
The same technologies that can help with online safety will also help to support the UK’s economic growth, and better enable an inclusive society via mechanisms such as increased access to Government data sets for those who are under-served today. The UK can build on its global reputation for building open, standards-based frameworks such as Open Banking, to deliver a world-leading approach to online safety based on the platforms verifying that their users are real people who have not been banned. This will enable the platform providers to provide a better product for their users; an environment without abusive comments or trolls, that also protects children from exposure to harmful content.
Yes, children can be protected if platform providers verify the ages of their users. This could be done to prove the users are over 18, for instance, and unsuitable content not shown unless the account has been verified.
DIN agrees that the focus should be on systems and processes, and also include the ability for platforms to confirm answers to the questions ‘who is the user who is posting the content?’, ‘who is the user consuming the content?’ and ‘are they over 18?’. AI technologies to filter content are constantly improving, but will never be able to 100% remove the risk that inappropriate content will be shown. Having a verifiable age would provide an absolute control over serving content to a user. Knowing the real identity of the posting account will also help with removing offenders, preventing them from creating new accounts, and enable fast reporting to law enforcement agencies for prosecution in the most serious cases.
Freedom of expression is still enabled under the framework of the Bill. Having a verified account still allows for the poster to express themselves. The account can also be pseudonymous to allow for a ‘persona’ to present someone’s views without publicly declaring who they are. The ‘persona’ should be traceable by the platform back to a real person to dissuade posting of harmful or illegal content. Freedom of expression/ freedom of speech is not the same as ‘free to say anything’.
Content in Scope
DIN would like to see specific reference to known fraud types that originate on user-to-user services, e.g. impersonation fraud where an account pretends to be a vendor, and obtains data or payment from an unsuspecting user victim, or romance scams. UK Finance collates industry information on fraud types and volumes for the UK FS industry; it would be good to see collaboration between public and private sector to define initial fraud types that could be included in the Bill, targeted and measured to see if the Bill approach is effective, and then apply this approach to other fraud types over time.
Services in Scope
Algorithms and user agency
AI technologies can be used to identify illegal or harmful content, but the identity behind the account is also needed to hold the user who owns the account responsible (otherwise illegal content can be posted by anonymous accounts with no accountability). AI content identification and digital identity technologies combined would give the best approach; to find the illegal content being posted, and to know who is doing it. There is little point in finding content and closing an account if the offender can just create a new account and re-post the material without any consequences. Certified digital identity technologies should be included in the ‘Use of technology’ approach to enable accountability. The certification can be provided by DCMS’ Digital Identity Trust Framework, to join up different Government initiatives.
The role of Ofcom
If the platform providers adopt digital identity, this will enable a much-improved process for users to report misuse to the platforms, and for the platforms to report the real identity to the police and Ofcom in a timely manner. This will maximise the time that the police have to establish a case before the statutory window expires.
27 September 2021