Artificial Intelligence

Google Announces Bug Bounty Program and Other Initiatives to Secure AI

Google announces a bug bounty program and other initiatives for increasing the safety and security of AI.

Google announces a bug bounty program and other initiatives for increasing the safety and security of AI.

Google today announced several initiatives meant to improve the safety and security of AI, including a bug bounty program and a $10 million fund.

The new vulnerability reporting program (VRP), Google says, will reward researchers for finding vulnerabilities in generative AI, to address concerns such as the potential for unfair bias, hallucinations, and model manipulation.

The internet giant is looking for reports detailing attack scenarios leading to prompt injections, data leaks, tampering with the behavior of a model, misclassifications in security controls, and the extraction of exact architecture or weights of a confidential/proprietary model.

According to Google, security researchers may also be rewarded for other vulnerabilities or behaviors in AI-powered tools that could lead to a valid security or abuse issue.

“Our scope aims to facilitate testing for traditional security vulnerabilities as well as risks specific to AI systems. It is important to note that reward amounts are dependent on severity of the attack scenario and the type of target affected,” the internet giant explains.

To improve the security of AI, Google has introduced a Secure AI Framework (SAIF), which aims to secure the “critical supply chain components that enable machine learning (ML)”.

Today, the company is announcing the first prototypes for model signing and attestation verification, which leverage Sigstore and SLSA to verify the identity of software and improve supply chain resilience.

The initiative focuses on increasing the transparency of the ML supply chain throughout the development lifecycle, to help mitigate the recent increase in supply chain attacks.

Advertisement. Scroll to continue reading.

“Based on the similarities in development lifecycle and threat vectors, we propose applying the same supply chain solutions from SLSA and Sigstore to ML models to similarly protect them against supply chain attacks,” Google notes.

Additionally, together with Anthropic, Microsoft, and OpenAI, Google today announced the creation of a $10 million AI Safety Fund, focused on promoting research in the field of AI safety.

Related: Google Expands Bug Bounty Program With Chrome, Cloud CTF Events

Related: Google Workspace Introduces New AI-Powered Security Controls

Related: Hacker Conversations: Natalie Silvanovich From Google’s Project Zero

Related Content

Artificial Intelligence

When not scamming other criminals, criminals are concentrating on the use of mainstream AI products rather than developing their own AI systems.

Artificial Intelligence

Israeli AI security firm Apex has received $7 million in seed funding for its detection, investigation, and response platform.

Artificial Intelligence

Japan's Prime Minister unveiled an international framework for regulation and use of generative AI, adding to global efforts on governance for the rapidly advancing...

Malware & Threats

Researchers can earn as much as $450,000 for a single vulnerability report as Google boosts its mobile vulnerability rewards program.

Application Security

Adobe is providing incentives for bug bounty hackers to report security flaws in its implementation of Content Credentials and Adobe Firefly.

Artificial Intelligence

AI-Native Trust, Risk, and Security Management (TRiSM) startup DeepKeep raises $10 million in seed funding.

Artificial Intelligence

Microsoft provides an easy and logical first step into GenAI for many organizations, but beware of the pitfalls.

Artificial Intelligence

CEOs of major tech companies are joining a new artificial intelligence safety board to advise the federal government on how to protect the nation’s...

Copyright © 2024 SecurityWeek ®, a Wired Business Media Publication. All Rights Reserved.

Exit mobile version