Contact

info@th3fus3.com



© 2025 TheFuse. All rights reserved.

UK AI Institute Expands

New Office in San Francisco

May 20, 2024 02:38 PM

Reading time: 2 minutes, 24 seconds

TL;DR The UK's AI Safety Institute is expanding internationally, opening its first overseas office in San Francisco this summer. This move aims to tap into the Bay Area's tech talent and strengthen global AI safety efforts. The expansion follows the UK's landmark AI Safety Summit.

The United Kingdom's Artificial Intelligence (AI) Safety Institute is set to expand internationally with a new location in the United States.

On May 20, Michelle Donelan, the U.K. Technology Secretary, announced that the institute will open its first overseas office in San Francisco in the summer.

Strategic Choice

The announcement said that the strategic choice of a San Francisco office will allow the U.K. to 'tap into the wealth of tech talent available in the Bay Area,' along with engaging with one of the world's largest AI labs located between London and San Francisco. Additionally, it said this move will help it 'cement' relationships with key players in the U.S. to push for global AI safety 'for the public interest.'

Growing Expertise

Already the London branch of the AI Safety Institute has a team of 30 that is on trajectory to scale and acquire more expertise particularly in the field of risk assessment for frontier AI models. Donelan said the expansion represents the U.K.'s leader and vision for AI safety in action. 'It is a pivotal moment in the UK's ability to study both the risks and potential of AI from a global lens, strengthening our partnership with the US and paving the way for other countries to tap into our expertise as we continue to lead the world on AI safety,' she added.

A Follow-Up to the AI Safety Summit

This follows the U.K.'s landmark AI Safety Summit that took place in London in November 2023. The summit was the first of its kind in focusing on AI safety on a global scale. The event boasted of leaders from around the world, including the U.S. and China, and leading voices in the AI space including Microsoft president Brad Smith, OpenAI CEO Sam Altman, Google and DeepMind CEO Demis Hassabiss, and Elon Musk.

Recent Results Released

In this latest announcement, the U.K. said it also is releasing a selection of the institute's recent results from safety testing it conducted of five publicly available advanced AI models. It anonymized the models and said the results provide a 'snapshot' of the capabilities of the models instead of designating them as 'safe' or 'unsafe.' Part of the findings included that several models could complete cyber security challenges, though others struggled with more advanced ones. Several models were found to have PhD-level knowledge of chemistry and biology.

Vulnerabilities and Capabilities

It concluded that all tested models were 'highly vulnerable' to basic jailbreaks and that the tested models were not able to complete more 'complex, time-consuming tasks' without human supervision. Ian Hogearth, the chair of the institute, said these assessments will help contribute to an empirical assessment of model capabilities. 'AI safety is still a very young and emerging field. These results represent only a small portion of the evaluation approach AISI is developing.'

Share this

Similar news
cryptocurrency

Peirce's Vision for Crypto

SEC's Crypto Mom on Regulatory Durability

March 27, 2025 01:15 AM
cryptocurrency

SEC Sets Target on Yet Another Crypto Market Maker

Crypto Market Maker Faces Legal Battle

October 11, 2024 01:00 PM
cryptocurrency

Striple Scores One-Day, 70-Country Stablecoin Payment Success

Stripe had previously discontinued Bitcoin payments due to high fees and slow confirmation times

October 11, 2024 11:59 AM
All results loaded