UK AI Safety Institute to open US office

The UK government's AI Safety Institute is expanding internationally by opening its first overseas office in San Francisco this summer.

The new office will allow the Institute to tap into the technology talent in the Bay Area, engage with major AI labs headquartered in London and San Francisco, and cement relationships with the United States to advance AI safety for the public interest.

The San Francisco office, expected to open in the coming months, will recruit an initial team of technical staff led by a Research Director. It will complement the Institute's London headquarters, which continues to grow and now has over 30 employees focused on assessing the risks of cutting-edge AI systems.

By establishing a presence in the US, the Institute aims to foster close collaboration, further the strategic partnership between the two countries on AI safety, share research, and conduct joint evaluations of AI models to inform global AI safety policies.

Announcing the expansion, Science and Technology secretary Michelle Donelan said, "This expansion represents British leadership in AI in action. It is a pivotal moment in the UK's ability to study both the risks and potential of AI from a global lens, strengthening our partnership with the US and paving the way for other countries to tap into our expertise as we continue to lead the world on AI safety."

The Institute has also released selected results from recent safety testing of five publicly available advanced AI models, making it the first government-backed organisation to unveil such evaluations. The tests assessed models across key risk areas, including the effectiveness of installed safeguards.

Findings included models completing some cybersecurity challenges but struggling with advanced ones, demonstrating PhD-level knowledge in chemistry and biology, vulnerability to basic "jailbreaks" enabling harmful outputs, and inability to complete complex, lengthy tasks without human oversight.

Institute Chair Ian Hogarth said the results contribute to assessing model capabilities and lack of robustness in existing safeguards, adding, "Our ambition is to continue pushing the frontier of this field by developing state-of-the-art evaluations, with an emphasis on national security related risks."



Share Story:

Recent Stories


Bringing Teams to the table – Adding value by integrating Microsoft Teams with business applications
A decade ago, the idea of digital collaboration started and ended with sending documents over email. Some organisations would have portals for sharing content or simplistic IM apps, but the ways that we communicated online were still largely primitive.

Automating CX: How are businesses using AI to meet customer expectations?
Virtual agents are set to supplant the traditional chatbot and their use cases are evolving at pace, with many organisations deploying new AI technologies to meet rising customer demand for self-service and real-time interactions.