AI 安全包括用于保障 AI 应用开发安全、管控员工 AI 使用以及保护 AI 驱动应用和模型的所有资源。
阅读本文后,您将能够:
复 制文章链接
Artificial intelligence (AI) security is the set of controls that prevent cyber attacks against, and ensure the safe behavior of, AI deployments. Just as cybersecurity in general protects IT systems and digital data, AI security protects the AI lifecycle — from building models, training data, and developing interfaces to deploying downstream applications. AI security technologies, processes, and practices do the following:
As AI adoption surged, AI security became a necessity. AI was adopted quickly: according to McKinsey, GenAI usage in organizations leaped from 33% in 2023 to 71% in 2024. By 2025, as many as 78% of organizations reported using AI in at least one business function.
For many organizations, the rapid increase in AI adoption vastly outpaced the capabilities of traditional security. AI made the organizational attack surface much more complex. AI systems comprise multiple interlocking layers — data pipelines, model training, model hosting, protocols, APIs, user interfaces, plugins, agents — that all must be secured.
For instance, a customer support bot — if manipulated by prompt injection or another AI-specific attack — could leak sensitive employee data or trade secrets. An attacker could abuse a model by overloading it with requests, causing AI resource overconsumption or denial of service. Understanding the key AI security risks and best practices, as well as security approaches tailored to generative and agentic AI, can help organizations prevent these kinds of attacks.
Shadow AI is the incorporation of AI models and tools without IT or security oversight. There are two types of shadow AI:
One survey found that 85% of IT decision makers report that employees are adopting AI tools faster than their IT teams can assess them. That same survey found that 93% of employees input information into AI tools without approval. Without a comprehensive view of the tools being used by the workforce, sensitive company data, such as proprietary code or personally identifiable information (PII), may be uploaded to AI services that fail to meet required security thresholds.
Large language models (LLMs) offer attractive targets for cybercriminals because they are so widely used, and in some cases are embedded into organizational infrastructure. OWASP's Top 10 Risks for LLMs list includes attacks like:
See the full list of top LLM risks.
Adopting AI at scale also introduces compliance and legal challenges. Organizations in highly regulated industries (finance and healthcare, for instance) face stiff penalties for failing to comply with data privacy regulations, including the Health Insurance Portability and Accountability Act (HIPAA) in the United States and the General Data Protection Regulation (GDPR) in the EU. AI can pose risks to private information in a couple of primary ways:
Security posture is a system’s readiness to mitigate attacks. Effectively managing it means taking a proactive approach to identifying, assessing, and acting on threats and vulnerabilities.
安全态势管理本质上很复杂,AI 进一步加剧了这种复杂性。由于 AI 系统涉及数据、模型、接口、API 以及往往采用异步通讯方式的智能体,AI 安全态势管理 (AI-SPM) 成为一项多维挑战。企业必须确保一致性、监测偏移风险、检测异常,以及将 AI 风险整合到企业风险框架。他们需要既能帮助促进 AI 采用,同时仍能维护企业网络和数据安全性以及隐私性的工具。
IT leaders can reduce the complexity of securing AI by looking for solutions that support some basic practices:
Securing GenAI usage, including LLMs and chat tools, requires a layered strategy. Organizations need to identify the GenAI tools in use, how users interact with those tools, and what happens to the outputs from those interactions.
一些最佳实践包括:
| GenAI risk | Security best practice |
|---|---|
| Shadow AI | Shadow AI discovery |
| Prompt injection | Model guardrails |
| Training data poisoning | Access control, encryption |
| PII leakage | Data loss prevention (DLP) |
AI agents are AI-powered programs that can autonomously make decisions, call external tools, and chain tasks. AI agents introduce their own risks. Agents can be manipulated over sessions and hijacked to execute unintended actions.
智能体式 AI 的主要风险包括:
遵循这些基本原则有助于保护 AI 智能体:
Core to agentic AI security is Model Context Protocol (MCP) security. AI agents rely on MCP servers in order to access external databases and tools, just as classic applications rely on external APIs. Learn more about MCP and MCP security.
AI security involves implementing controls that block cyber attacks and maintain the integrity of AI systems. Their goal is to safeguard the entire AI lifecycle — covering everything from initial model development and data training to the final deployment of applications and interfaces.
Shadow AI occurs when employees use AI tools or developers integrate models without official oversight from IT or security teams. This creates a visibility gap where sensitive company data or proprietary code might be uploaded to unapproved services that do not meet security standards.
Attackers use several methods to compromise LLMs, such as prompt injection to override built-in instructions, or data poisoning to corrupt training sets and skew model behavior. They may also attempt to steal proprietary models through API queries or launch denial-of-service attacks to exhaust computing resources.
Organizations should implement a layered strategy that includes identifying all AI traffic to uncover unauthorized tools. Key steps include using data loss prevention to stop sensitive information uploads, applying the principle of least privilege for access, and using guardrails to block harmful or inappropriate prompts.
Cloudflare AI Security for Apps helps defend public-facing applications against major threats like model theft and prompt injection. Their services also monitor user interactions to prevent the accidental exposure of private data in prompts or model outputs.