UK’s AI Safety Institute Unveils Platform to Accelerate Safe AI Development

Security

The UK’s AI Safety Institute has made its AI testing and evaluation platform available to the global AI community as of 10 May, 2024.

The platform, called Inspect, is set to pave the way for the safe innovation of AI models, according to the AI Safety Institute and Department for Science, Innovation and Technology (DIST).

By making Inspect available to the global community, the Institute said it is helping accelerate the work on AI safety evaluations carried out internationally. The aim is that this leads to better safety testing and the development of more secure models.

It also allows for a consistent approach to AI safety evaluations around the world. according to the government.

Inspect is a software library which enables testers – from start-ups, academia and AI developers to international governments – to assess specific capabilities of individual models and then produce a score based on their results.

Inspect can be used to evaluate models in a range of areas, including their core knowledge, ability to reason, and autonomous capabilities. Released through an open-source license, it means Inspect it is now freely available for the AI community to use.

AI Safety Institute Chair, Ian Hogarth, commented: “We have been inspired by some of the leading open source AI developers – most notably projects like GPT-NeoX, OLMo or Pythia which all have publicly available training data and OSI-licensed training and evaluation code, model weights, and partially trained checkpoints. This is our effort to contribute back.”

“We hope to see the global AI community using Inspect to not only carry out their own model safety tests, but to help adapt and build upon the open source platform so we can produce high-quality evaluations across the board.”

The UK AI Safety Institute was announced by British Prime Minister Rishi Sunak at the AI Safety Summit, held in Bletchley Park, England, in November, 2023.

At the time, Sunak said the UK government’s ambition for this new entity is to make it a global hub tasked with testing the safety of emerging types of AI.

Products You May Like

Articles You May Like

Italy’s Data Protection Watchdog Issues €15m Fine to OpenAI Over ChatGPT Probe
LockBit Developer Rostislav Panev Charged for Billions in Global Ransomware Damages
Thousands Download Malicious npm Libraries Impersonating Legitimate Tools
Attackers Exploit Microsoft Teams and AnyDesk to Deploy DarkGate Malware
Sophisticated TA397 Malware Targets Turkish Defense Sector

Leave a Reply

Your email address will not be published. Required fields are marked *