Cloud Security Alliance Launches AI Safety Effort

The Cloud Security Alliance (CSA) has formed an AI Safety Initiative coalition of working groups, per a Tuesday announcement.

The AI Safety Initiative includes Amazon, Anthropic, Google, Microsoft and OpenAI as partners. It also includes participation by academic experts, plus government agencies, such as U.S. Cybersecurity and Infrastructure Security Agency (CISA).

The initiative has already attracted the greatest number of CSA program participants since its founding, with more than "1,500 expert participants" working in the following four core research groups:

  • AI Technology and Risk Working Group
  • AI Governance & Compliance Working Group
  • AI Controls Working Group
  • AI Organizational Responsibilities Working Group

The aim of the AI Safety Initiative is to create safety and security guidelines for artificial intelligence (AI), with an initial focus on generative AI.

Generative AI Security
CSA has already published its first publication on the topic, "Security Implications of ChatGPT," which offers information about the kind of concerns that the CSA AI Safety Initiative aims to address.

The publication (released in August 2023), outlined "how AI-driven systems can be exploited in different aspects of cyberattacks, including enumeration, foothold assistance, reconnaissance, phishing and the generation of polymorphic code." Those items are all approaches that attackers might try against computing environments.

CSA's publication also mentioned the issue of users attacking generative AI by "prompt injection," delivering so-called "malicious prompts." Generative AI systems could be corrupted to "spread disinformation, interfere with its functioning, or deceive the AI into producing unsuitable responses by feeding it false or misleading information." Attackers may also use role-playing prompts to bypass generative AI's "filters and security protocols."

CSA's "Security Implications of ChatGPT" publication did not have specific guidelines for organizations on how to use generative AI securely. However, it promised a follow-up paper on the topic. Organizations meanwhile should develop guidelines and policies on the use of AI tools, protect sensitive information, protect communication channels against adversary-in-the-middle attacks, audit AI use and stay up to date on security developments.

Other Efforts
CSA's initiative on AI and security comes shortly after other parallel efforts have launched.

In late October, the Biden administration announced voluntary guidelines with an executive order on AI security. The executive order included concerns that AI use might have "chemical, biological, radiological, nuclear and cybersecurity risks," along with fraud and privacy implications.

In late November, the UK's National Cyber Security Centre and CISA announced joint guidelines for the secure development of AI systems, which were said to complement the White House's guidelines. This "Guidelines for Secure AI System Development" publication offers a framework for the secure design, development, deployment and operations of AI systems.

The NCSC-CISA guidelines suggested that AI security is somewhat different than other systems. Attackers can leverage "adversarial machine learning" to skew the model's performance or classifications, or extract sensitive model information, for instance. AI providers should take responsibility for security at present since users "do not typically have sufficient visibility and/or expertise to fully understand, evaluate or address risks associated with systems they are using," the guidelines argued.

Automated AI attacks are still a few years off, according to a report published last year, "The Security Threat of AI-Enabled Cyberattacks," by security solutions firm WithSecure, which was commissioned by the Finnish Transport and Communications Agency Traficom. However, it's currently possible to use AI to impersonate people's voices and use it for "spear phishing" messaging (targeting specific individuals in an organization).

About the Author

Kurt Mackie is senior news producer for 1105 Media's Converge360 group.


  • Image of a futuristic maze

    The 2024 Microsoft Product Roadmap

    Everything Microsoft partners and IT pros need to know about major Microsoft product milestones this year.

  • SharePoint Embedded Becomes Generally Available

    After a six-month preview, SharePoint Embedded, an API-based version of SharePoint that developers and ISVs can use to embed Microsoft 365 capabilities into their apps, is now generally available.

  • Copilot in Microsoft 365 Getting Agents, Extensions and Team (Not Teams) Support

    Microsoft is adding more functionality to its Copilot AI assistant aimed at improving business collaboration, processes and workflows for Microsoft 365 users.

  • Microsoft Giving Startups Templates To Build AI Apps

    A new perk for businesses enrolled in the Microsoft for Startups Founders Hub program aims to fast-track their ability to build AI-powered applications.