Stay organized with collections
Save and categorize content based on your preferences.
AI safety includes a set of design and operational techniques to follow to
avoid and contain actions that can cause harm, intentionally or unintentionally.
For example, do AI systems behave as intended, even in the face of a security
breach or targeted attack? Is the AI system robust enough to operate safely
even when perturbed? How do you plan ahead to prevent or avoid risks? Is the AI
system reliable and stable under pressure?
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-02-25 UTC."],[[["AI safety encompasses design and operational techniques to prevent harm, ensuring AI systems behave as intended, even under pressure or attack."],["Adversarial testing is a key safety technique where AI systems are intentionally challenged with malicious or harmful input to assess their robustness."],["Google's Responsible AI Practices provide recommendations for protecting AI systems, including methods for adversarial testing and safeguarding against attacks."]]],[]]