Stay organized with collections
Save and categorize content based on your preferences.
Accountability means owning responsibility for the effects of an AI system.
Accountability typically involves transparency, or sharing information about
system behavior and organizational process, which may include documenting and
sharing how models and datasets were created, trained, and evaluated. The
following sites explain two valuable modes of accountability documentation:
Another dimension of accountability is interpretability, which involves the
understanding of ML model decisions, where humans are able to identify features
that lead to a prediction. Moreover, explainability is the ability for a
model's automated decisions to be explained in a way for humans to understand.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-02-25 UTC."],[[["\u003cp\u003eAccountability in AI involves taking ownership for the effects of a system, often achieved through transparency about the system's development and behavior.\u003c/p\u003e\n"],["\u003cp\u003eTransparency can be enhanced using documentation practices like Model Cards and Data Cards, which provide information about models and datasets.\u003c/p\u003e\n"],["\u003cp\u003eInterpretability and explainability are crucial aspects of accountability, enabling understanding of model decisions and providing human-understandable explanations for automated actions.\u003c/p\u003e\n"],["\u003cp\u003eFostering user trust in AI systems requires focusing on explainability and transparency, with further resources available in Google's Responsible AI Practices and Explainability Resources.\u003c/p\u003e\n"]]],[],null,["\u003cbr /\u003e\n\n**Accountability** means owning responsibility for the effects of an AI system.\nAccountability typically involves **transparency**, or sharing information about\nsystem behavior and organizational process, which may include documenting and\nsharing how models and datasets were created, trained, and evaluated. The\nfollowing sites explain two valuable modes of accountability documentation:\n\n- [Model Cards](https://modelcards.withgoogle.com/about)\n- [Data Cards](https://sites.research.google/datacardsplaybook/)\n\nAnother dimension of accountability is **interpretability** , which involves the\nunderstanding of ML model decisions, where humans are able to identify features\nthat lead to a prediction. Moreover, **explainability** is the ability for a\nmodel's automated decisions to be explained in a way for humans to understand.\n\nRead more about building user trust in AI systems in the [Explainability +\nTrust](https://pair.withgoogle.com/chapter/explainability-trust/) section of the\n[People + AI Guidebook](https://pair.withgoogle.com/guidebook).\nYou can also check out [Google's Explainability Resources](https://explainability.withgoogle.com/)\nfor real life examples and best practices."]]