Artificial Intelligence

US, UK Cybersecurity Agencies Publish AI Development Guidance

New guidance from US and UK cybersecurity agencies provides recommendations for secure AI system development.

The US and UK cybersecurity agencies CISA and NCSC have published security-focused guidance for the developers of systems that leverage AI.

The document, titled Guidelines for Secure AI System Development (PDF), promotes the implementation of secure-by-design principles, as well as transparency and accountability, and prioritizes ownership of security outcomes for customers.

The guidelines, the two agencies note, apply to all types of AI/ML systems, regardless of whether built from scratch or on top of third-party resources, to address issues related to AI, cybersecurity, and critical infrastructure.

Developed in collaboration with over 20 domestic and international cybersecurity organizations, the document has been broken down into four sections, covering different stages of the AI system development lifecycle, namely design, development, deployment, and operation and maintenance.

Meant to be applied in conjunction with cybersecurity, incident response, and risk management best practices, the recommendations demand investments in features, mechanisms, and tools that protect customer data at all layers, throughout the entire system lifecycle, CISA and NCSC say.

“Providers should implement security controls and mitigations where possible within their models, pipelines and/or systems, and where settings are used, implement the most secure option as default,” the two agencies note.

CISA and NCSC also say that providers are responsible for informing users of risks that cannot be mitigated and of advising them how to use systems securely, and that they should treat all cybersecurity risks as critical.

Providers are advised to assess the threats to their systems, focus on security, functionality, and performance during the design stage, to secure their supply chain, protect their assets, secure their infrastructure and protect their model continuously, implement incident response, monitor the system’s behavior and inputs, and implement a secure-by-designed approach to updates.

Advertisement. Scroll to continue reading.

The guidelines, the two agencies say, are primarily aimed at providers of AI systems, either hosted by an organization or accessed via external APIs. However, all stakeholders, “including data scientists, developers, managers, decision-makers, and risk owners”, are encouraged to read the document “to make informed decisions about the design, development, deployment and operation of their AI systems,” the two agencies note.

Related: Pentagon’s AI Initiatives Accelerate Hard Decisions on Lethal Autonomous Weapons

Related: The $64k Question: How Does AI Phishing Stack Up Against Human Social Engineers?

Related: White House Unveils New Efforts to Guide Federal Research of AI

Related Content

Artificial Intelligence

China’s official Xinhua news agency said the two sides would take up issues including the technological risks of AI and global governance.

Artificial Intelligence

When not scamming other criminals, criminals are concentrating on the use of mainstream AI products rather than developing their own AI systems.

Vulnerabilities

CISA and the FBI warn of threat actors abusing path traversal software vulnerabilities in attacks targeting critical infrastructure.

Government

The White House has published a national security memorandum focusing on critical infrastructure security and resilience.

Artificial Intelligence

Israeli AI security firm Apex has received $7 million in seed funding for its detection, investigation, and response platform.

Artificial Intelligence

Japan's Prime Minister unveiled an international framework for regulation and use of generative AI, adding to global efforts on governance for the rapidly advancing...

Artificial Intelligence

AI-Native Trust, Risk, and Security Management (TRiSM) startup DeepKeep raises $10 million in seed funding.

Artificial Intelligence

Microsoft provides an easy and logical first step into GenAI for many organizations, but beware of the pitfalls.

Copyright © 2024 SecurityWeek ®, a Wired Business Media Publication. All Rights Reserved.

Exit mobile version