Which risks from AI are the most urgent to mitigate?

Background

In the National AI Strategy, the government made commitments to enrich our understanding of AI as it impacts the economy and society more broadly. Additionally, we recently launched a steering board chaired by the heads of both the government analysis and scientific functions, to ensure cohesive cross government approaches to understanding AI impacts. An overview of the high-level questions we are asking in this regard are outlined in the section below. (https://www.gov.uk/government/publications/national-aistrategy)

Some priority work we are currently developing to meet these commitments include:

An analysis of the AI White Paper consultation to feed into the formal consultation response. This will allow us to take on feedback from the public and various key players in sectors across the economy, and better tailor policy interventions to support strategic AI aims.

Establishing the AI Safety Institute to advance the world’s knowledge of AI safety by carefully examining, evaluating, and testing new frontier AI systems. The Institute will conduct fundamental research on how to keep people safe in the face of fast and unpredictable progress in AI, improving our understanding of the capabilities and risks of AI systems.

A monitoring and evaluation framework for AI regulatory interventions in tandem with the AI regulatory white paper. This will develop our understanding of key metrics to monitor with regards to Ai governance and ecosystem impacts.

Research into the AI sector and supply. Updating the AI Sector Study to establish a consistent and comparable set of economic indicators for the AI sector in terms of producers and suppliers. This study helps us to best understand where the AI sector needs support, to grow sovereign capability of the UK in AI, in alignment with strategic priorities.

The development of a cross-economy national AI risk register. Developed in tandem with a responsibility register that garnered cross Whitehall agreement on which departments hold which risks with regards to AI. The risk register will provide a single source of truth on AI risks which regulators, government departments, and external groups can use to prioritise further action.

Further research into Compute and the best ways to leverage compute to support the AI sector. This will be key to informing our response to the future of compute review, and maximising the £1 billion+ investments in state-of-the-art compute.

Next steps

If you are keen to register your interest in working and connecting with DSIT Digital Technology and Telecoms Group and/or submitting evidence, then please complete the DSIT-ARI Evidence survey - https://dsit.qualtrics.com/jfe/form/SV_cDfmK2OukVAnirs.
Please view full details: https://www.gov.uk/government/publications/department-for-science-innovation-and-technology-areas-of-research-interest/dsit-areas-of-research-interest-2024

Related UKRI funded projects


  • ENFIELD: European Lighthouse to Manifest Trustworthy and Green AI

    ENFIELD will create a unique European Centre of Excellence that excels the fundamental research in the scientific pillars of Adaptive, Green, Human-Centric, and Trustworthy AI that are new, strategic and of paramount imp...

    Funded by: Horizon Europe Guarantee

    Lead research organisation: UNIVERSITY OF NOTTINGHAM

    Why might this be relevant?

    Partially relevant as it focuses on creating a European Centre of Excellence in AI research, but does not specifically address urgent risks in AI.

  • AI UK: Creating an International Ecosystem for Responsible AI Research and Innovation

    Artificial Intelligence (AI) can have dramatic effects on industrial sectors and societies (e.g., Generative AI, facial recognition, autonomous vehicles). AI UK will pioneer a reflective, inclusive approach to responsibl...

    Funded by: EPSRC

    Lead research organisation: University of Southampton

    Why might this be relevant?

    Partially relevant as it focuses on creating an international ecosystem for responsible AI research, but does not specifically address urgent risks in AI.

  • Seclea Platform - Responsible AI Tools for Everyone

    Artificial Intelligence has the potential to improve our lives with rapid, personalised and assistive services. It presents risks of negative effects on both society and individual citizens. Recent debacles have showed t...

    Funded by: Innovate UK

    Lead research organisation: SECLEA LTD

    Why might this be relevant?

    Fully relevant as it specifically addresses the risks associated with AI and provides tools for responsible AI development.

  • ELIAS: European Lighthouse of AI for Sustainability

    We live in a crucial historical moment, with tremendous challenges ahead, from climate change to the energy crisis. ELIAS emerges from the belief that AI will be a key discipline to help us tackle these challenges. At th...

    Funded by: Horizon Europe Guarantee

    Lead research organisation: THE UNIVERSITY OF MANCHESTER

  • Responsible AI for Long-term Trustworthy Autonomous Systems (RAILS): Integrating Responsible AI and Socio-legal Governance

    Society is seeing enormous growth in the development and implementation of autonomous systems, which can offer significant benefits to citizens, communities, and businesses. The potential for improvements in societal wel...

    Funded by: SPF

    Lead research organisation: University of Oxford

  • LEAP - Legal Ecosystem for AI Proliferation

    LEAP is the Legal Ecosystem for AI Proliferation. Innovation in AI and the potential impact it is having on current and emerging industries cannot be underestimated. The speed of this innovation means the impacts are bar...

    Funded by: Innovate UK

    Lead research organisation: CHARLTON STONEHILL LTD

  • Seclea – Building Trust in AI

    Artificial Intelligence has the potential to improve our lives with rapid, personalised and assistive services. At the same time, it presents risks of negative impacts on both society and individual citizens. Recent deba...

    Funded by: Innovate UK

    Lead research organisation: SECLEA

  • Multi-scale Operation-assurance evaluation Tool for AI (MOT4AI) Systems

    According to a UK government report (2019), approximately 50% of SMEs in the UK were using AI technology. The government has called for greater transparency in AI systems to ensure they are used ethically and fairly. We ...

    Funded by: Innovate UK

    Lead research organisation: DIGITAL READINESS & INTELLIGENCE LTD

  • GRAIL (Guide to Regulation & AI Laws) A natural language process approach to discovery of AI regulations & laws.

    AI has the potential to transform our understanding of the world and help resolve many persistent human and environmental concerns. This transformative potential has stimulated the creation of a plethora of guidelines, d...

    Funded by: Innovate UK

    Lead research organisation: ORBIT RRI LIMITED

    Why might this be relevant?

    Partially relevant as it focuses on AI regulations and laws, not specifically on risks.

  • UKRI Trustworthy Autonomous Systems Hub

    Public opinion on complex scientific topics can have dramatic effects on industrial sectors (e.g. GM crops, fracking, global warming). In order to realise the industrial and societal benefits of Autonomous Systems, they ...

    Funded by: SPF

    Lead research organisation: University of Southampton

Similar ARIs from other organisations