Model Card Template

Use this template to document your ML models following the standard model card framework established by Mitchell et al. (2019). Each section includes guidance on what to include.

1. Model Details

The official name of the model, including version identifier.

[Your content here]

Person or organization that developed the model. Include contact information.

[Your content here]

Date the model was developed or last updated (YYYY-MM-DD).

[Your content here]

Version number or identifier. Use semantic versioning if applicable.

[Your content here]

Architecture type (e.g., transformer, CNN, random forest) and training algorithm.

[Your content here]

License under which the model is released (e.g., Apache 2.0, MIT, CC-BY-4.0).

[Your content here]

Paper or resource to cite when using this model.

[Your content here]

2. Intended Use

The primary intended use cases for the model. Be specific about tasks, domains, and deployment contexts.

[Your content here]

Who is the model designed for? (e.g., researchers, developers, end-users, specific industries).

[Your content here]

Use cases that the model is explicitly not designed for or should not be used for. Include known failure modes.

[Your content here]

3. Factors

Demographic or phenotypic groups, instrumentation, and environment factors relevant to model performance.

[Your content here]

Factors that were explicitly tested during evaluation. List the groups and conditions evaluated.

[Your content here]

4. Metrics

Metrics used to evaluate the model (e.g., accuracy, F1, AUC-ROC, BLEU). Justify each choice.

[Your content here]

Thresholds used for classification decisions and the rationale for choosing them.

[Your content here]

How performance variation is measured (e.g., confidence intervals, standard deviation across folds).

[Your content here]

5. Evaluation Data

Names and descriptions of evaluation datasets used. Include size, source, and any filtering applied.

[Your content here]

Why these datasets were chosen. How do they represent the intended use cases?

[Your content here]

Preprocessing steps applied to evaluation data (e.g., tokenization, normalization, augmentation).

[Your content here]

6. Training Data

Names and descriptions of training datasets. Include size, source, collection methodology, and time period.

[Your content here]

Preprocessing and data augmentation steps applied during training.

[Your content here]

Origin and chain of custody for training data. Include any licensing or consent information.

[Your content here]

7. Quantitative Analyses

Overall performance metrics on the evaluation dataset(s). Report all metrics listed in Section 4.

[Your content here]

Disaggregated performance across relevant factors and their intersections. Report metrics for each subgroup.

[Your content here]

8. Ethical Considerations

Known or potential harms from model use, including allocation harms, quality-of-service harms, and representational harms.

[Your content here]

Particularly sensitive applications where extra caution is needed (e.g., healthcare, criminal justice, hiring).

[Your content here]

Steps taken to reduce potential harms (e.g., debiasing techniques, output filtering, human-in-the-loop).

[Your content here]

9. Caveats and Recommendations

Known limitations of the model. Include scenarios where performance degrades or the model is unreliable.

[Your content here]

Recommendations for model users, including suggested monitoring, testing before deployment, and update cadence.

[Your content here]

Areas for improvement, planned updates, and research directions.

[Your content here]

Learn more: Read about evaluation methodology to understand how to select metrics and design evaluation protocols for Sections 4-7.