🧠 All Things AI
Intermediate

Model Cards

A model card is a short document that accompanies a trained machine learning model and provides essential transparency information about it. Introduced by Mitchell et al. at Google (2019), model cards are now a standard practice for any AI system released publicly or deployed in a high-stakes context. They serve multiple audiences simultaneously: developers who want to know if a model is fit for their use case, affected individuals who want to understand what decisions the model informs, and regulators who need evidence of responsible deployment.

Standard Model Card Structure

SectionContent
Model detailsName, version, type, training date, organisation, licence, contact, links to related resources
Intended usePrimary intended uses; primary intended users; out-of-scope uses; explicitly prohibited uses
FactorsRelevant factors that affect model performance — demographic groups, environmental conditions, instrumentation, data formats
MetricsPerformance metrics used; thresholds for acceptable performance; decision thresholds if applicable
Evaluation dataDataset(s) used for evaluation; why they were chosen; preprocessing applied; any known limitations of the eval set
Training dataTraining data description (can be high-level for proprietary systems); notable characteristics or known issues
Quantitative analysesDisaggregated performance metrics by relevant subgroups (race, gender, age, geography)
Ethical considerationsKnown risks; potential for harm; sensitive use cases; mitigation measures taken
Caveats and recommendationsWhat the model should not be used for; recommendations for monitoring; open questions

Performance Disaggregation by Subgroup

The most valuable section of a model card — and the one most frequently omitted — is disaggregated performance. Reporting a single aggregate accuracy hides differential performance across demographic or contextual subgroups.

What good disaggregation looks like

  • Report the same primary metric (accuracy, F1, recall) per subgroup
  • Identify which subgroups are relevant to the deployment context
  • Include confidence intervals — small subgroups have high variance
  • Flag gaps that exceed a defined threshold (e.g., >5 percentage points)
  • Document what was done if gaps were found (mitigation, accepted residual risk)

Common omissions that weaken model cards

  • Aggregated metrics only — no per-subgroup breakdown
  • Vague ethical considerations section with no specific risks named
  • Intended use stated so broadly that it does not exclude any use case
  • Training data described as "large-scale web data" with no further characterisation
  • No update date — cards become stale after model updates

Model Cards for LLMs

Large language models require adaptations to the model card format. Key differences:

  • Intended use is harder to bound: LLMs are general-purpose. Model cards must describe what use cases the model was optimised for and which high-risk use cases are explicitly out of scope (medical diagnosis, legal advice, crisis counselling).
  • Performance metrics are multi-dimensional: Reporting a single accuracy metric is not meaningful. LLM cards report: benchmark results (MMLU, HumanEval, etc.), safety evaluations (refusal rates, toxicity scores), and alignment evaluation results.
  • Known failure modes: Hallucination rates, calibration on factual questions, performance degradation for non-English languages, and susceptibility to known prompt injection patterns should be documented.
  • Alignment and safety section: RLHF or RLAIF alignment methods used, red team testing performed, safety classifiers applied as guardrails.
  • Usage restrictions: Terms of service restrictions that apply — which sectors or use cases require explicit permission or are prohibited (e.g., weapons development, surveillance, political disinformation).

Regulatory Requirements for Model Documentation

FrameworkDocumentation requirementWho must comply
EU AI Act (high-risk)Technical documentation including model performance, training data, testing methodology, intended purpose, known limitationsProviders of high-risk AI systems (Annex III uses)
EU AI Act (GPAI)Model cards or equivalent transparency document required for all general-purpose AI models released publiclyGPAI model providers above capability threshold
NIST AI RMFAI Transparency documentation as part of GOVERN function; model documentation as part of MAP outputsVoluntary — referenced in US federal procurement
ISO 42001 (Clause 8.7)Information for AI systems: intended use, limitations, performance characteristics, user guidanceOrganisations seeking ISO 42001 certification
FDA (AI/ML SaMD)Software documentation requirements including algorithm description, performance on intended population, known limitationsAI-based medical device software in the US

Keeping Model Cards Current

A model card created at initial deployment and never updated is a compliance liability. Model cards must be treated as living documents:

  • Trigger updates: Model retrained or fine-tuned; new evaluation data available; new failure mode discovered in production; deployment context expands; regulatory requirements change
  • Version the card: Each model version should have a corresponding model card version with changelog. Do not silently overwrite previous versions.
  • Link to model registry: Model card should be accessible from the model registry entry — when a model is looked up, the card should be one click away
  • Assign an owner: Designate a person responsible for keeping the card current — typically the model owner or AI product manager

Checklist: Do You Understand This?

  • Name the nine standard sections of a model card and what each covers.
  • Why is disaggregated performance the most important section of a model card?
  • How does a model card for an LLM differ from one for a traditional classification model?
  • Under the EU AI Act, which types of AI systems require model documentation? What must the documentation include?
  • What triggers a model card update and who should own that responsibility?