Category Archives: Artificial Intelligence

Model Cards for LLMs — what they are, why they matter, and how the EU AI Act gives them a boost

TL;DR: Model cards are structured “fact sheets” for AI models. They document purpose, training data, performance, risks, limitations, and responsibilities — making them the fastest route to reliable transparency. Under the EU AI Act, exactly these information duties for general-purpose/LLM models are becoming binding in stages: since 2 August 2025, transparency and copyright obligations apply in the EU, including a public summary of training data; particularly capable models face additional safety and risk obligations. A well-crafted model card helps fulfill these duties efficiently.

What are model cards (for LLMs)?

Model cards were proposed in 2018/2019 as a standard for model-accompanying documentation: concise, structured documents that disclose a model’s intended use, performance metrics (including subgroups), known limits, and ethical considerations. For LLMs, this additionally means pretraining/finetuning details, the RLHF process, benchmarks (e.g., MMLU/MT-Bench), resilience to hallucinations and jailbreaks, privacy tests, and the energy footprint.

In practice, model cards have become a de facto standard — for example on the Hugging Face Hub as a README.md plus metadata.

What does this have to do with the EU AI Act?

1) Timeline and scope

The EU AI Act entered into force on 1 August 2024. For general-purpose AI (GPAI) — which includes LLMs — transparency and copyright obligations have applied in stages since 2 August 2025; models placed on the market earlier benefit from transition periods until 2 August 2027.

2) Concrete obligations for LLM/GPAI providers

  • Transparency and technical documentation: Provide information on capabilities, limitations, and safe use (Art. 53).
  • Copyright: Maintain policies/processes for copyright compliance.
  • Public training-data summary: Publish a “sufficiently detailed” summary of the content used for training — following the EU template made available by the Commission in 2025 (Art. 53(1)(d)).
  • Systemic risks (very large models only): For models above a compute threshold (e.g., 10^25 FLOPs), additional obligations apply, including risk assessment, red teaming, incident reporting, cybersecurity, and more.

The Act also requires transparency toward users — for example, labeling synthetic content and informing people when they interact with AI (Art. 50).

In short: The information the AI Act requires forms the core of a good model card.

How to build an “AI-Act-ready” model card for LLMs

Below is a practical structure — with an indication of which AI-Act idea it supports:

  1. Model profile (name, version, date, contact) — technical documentation/accountability.
  2. Intended uses and out-of-scope usessafe use, minimization of misuse.
  3. Training-data summary (source classes, curation logic, geographies/languages, licensing mix, exclusion criteria) — public summary per the EU template (link/appendix).
  4. Copyright policy (handling opt-outs/opt-ins, TDM exceptions, rights preservation) — copyright compliance.
  5. Model development (pretraining objective, finetuning, RLHF, safety filters) — transparency/technical documentation.
  6. Performance and evaluation (benchmarks, methodological notes, domains/languages, subgroup analyses) — traceability and fairness assessment.
  7. Risk profile (hallucinations, bias, privacy leakage, jailbreaks, known failure modes) — risk management/systemic-risk topics for large models.
  8. Red teaming and safety (test design, adversarial testing, content moderation, incident process) — safety obligations for systemic risks.
  9. Energy and operations (training compute, inference cost/footprint, efficiency measures) — best-practice transparency.
  10. User guidance (prompt examples, limitations, labeling duties for synthetic media, monitoring tips) — Art. 50 transparency toward end users.
  11. Governance and maintenance (versioning/changelog, deprecation policy, incident contact channels) — continuity and compliance evidence.

Common pitfalls (and how the model card helps)

  • “We can’t disclose our training data.” — You don’t have to: what’s required is a summary, not a raw data dump, and there is an official template.
  • “Our model is open source, so we’re exempt.” — Only partially: open GPAI models still need to provide a training-data summary and a copyright policy, among other things; exemptions do not apply where systemic risk is concerned.
  • “Transparency ≠ marketing brochure.” — A model card is technical documentation that should be audit-ready — not just nicely phrased.

Conclusion

Model cards were “good practice” for years. With the EU AI Act, they become an operational lever for implementing transparency, copyright, and safety requirements concretely and verifiably. Put more personally: a good model card ultimately saves time — because it answers the right questions up front.

https://arxiv.org/abs/1810.03993 “Model Cards for Model Reporting”
https://huggingface.co/docs/hub/en/model-cards “Model Cards”
https://commission.europa.eu/news-and-media/news/ai-act-enters-force-2024-08-01_en “AI Act enters into force – European Commission”
https://artificialintelligenceact.eu/article/53/ “Article 53: Obligations for Providers of General-Purpose AI …”
https://digital-strategy.ec.europa.eu/en/policies/contents-code-gpai “The General-Purpose AI Code of Practice”
https://digital-strategy.ec.europa.eu/en/faqs/general-purpose-ai-models-ai-act-questions-answers “General-Purpose AI Models in the AI Act – Questions & Answers”
https://digital-strategy.ec.europa.eu/en/news/eu-rules-general-purpose-ai-models-start-apply-bringing-more-transparency-safety-and-accountability “EU rules on general-purpose AI models start to apply, bringing …”
https://artificialintelligenceact.eu/article/50/ “Article 50: Transparency Obligations for Providers and …”
https://artificialintelligenceact.eu/high-level-summary/ “High-level summary of the AI Act”

Apertus: Switzerland Opens the Door to a Transparent AI Future

From ETH Zurich to Swisscom: Apertus marks the birth of Switzerland’s first large open-source language model — a European signal for transparency, research, and digital sovereignty.

An Open Countermodel to the AI Powerhouses

While the global spotlight shines on the major U.S. tech companies — OpenAI, Anthropic, Google, and Meta — a remarkable alternative is taking shape in Switzerland: Apertus. Unveiled in 2025, this large language model is not only the first of its kind developed in Switzerland but also one of the most transparent AI initiatives worldwide. At its core lies a bold vision: artificial intelligence as a public good, not a closed commercial product.

A Collaboration Between ETH, EPFL, and Swisscom

Apertus is the result of a collaboration among Switzerland’s leading research and technology institutions:

  • ETH Zurich and EPFL Lausanne lead the scientific development.
  • The Swiss National Supercomputing Centre (CSCS) in Lugano provides the computational power via its Alps supercomputer.
  • Swisscom integrates the model into its Swiss AI Platform, bringing it to enterprise clients.

This combination of academic excellence and industrial application makes Apertus a showcase of applied AI innovation in Europe.

Technical Foundation: GPT-Like, but with Swiss Precision

Apertus is built on the transformer architecture, similar to the GPT family of models. Two versions have been released:

  • a model with 8 billion parameters, and
  • a larger one with 70 billion parameters,

each available in instruction-tuned variants optimized for conversational use.

The training corpus consists of about 15 trillion tokens across more than a thousand languages. Special attention was given to underrepresented languages such as Swiss German and Romansh.

A custom embedding system was designed to prevent English from dominating the dataset — a frequent bias in global AI models.

Transparency as a Principle

What truly distinguishes Apertus is its radical openness.
The entire training process, source code, and datasets have been made public — an unprecedented level of transparency among large-scale language models.

This makes Apertus fully compliant with the EU AI Act’s transparency requirements and sets a new standard for responsible AI development.
Researchers, developers, and companies can trace exactly how and with what data the model was trained.

It’s not just a technical decision — it’s a political statement about accountability and democratic oversight in artificial intelligence.

Data Protection and Ethical Boundaries

The project’s creators emphasize that development followed Swiss data protection and copyright laws.

At the same time, they openly acknowledge that not every textual source used in training was formally licensed — a transparency that reflects intellectual honesty rather than negligence.

Apertus thus navigates the realistic gray zone that many AI research projects inhabit, where the ideals of openness intersect with the complex realities of digital content rights.

Not Yet a Chatbot — but a Powerful Foundation

As of autumn 2025, Apertus exists purely as a text-based model.
It does not yet support multimodal capabilities such as image, video, or audio processing. However, the development roadmap envisions extensions that could eventually make Apertus a fully fledged open-source assistant.

Even today, the model serves as a robust foundation for specialized AI applications in research, public administration, and industry.

Why Apertus Matters

Apertus is more than a research project — it is a message to Europe.
At a time when AI infrastructure is increasingly concentrated in private hands, Switzerland demonstrates that technological sovereignty and openness can coexist.

The benefits are clear:

  • Independence from U.S. platforms
  • Compliance with European data protection standards
  • Promotion of multilingual and local innovation

In this sense, Apertus stands as a symbol of an independent European AI strategy — small, precise, and transparent.

Conclusion: Apertus as a Blueprint for Open AI

With Apertus, ETH, EPFL, and Swisscom have set a new benchmark:
a powerful, auditable, and freely accessible language model designed to serve science, business, and society alike.

In a world where many AI systems operate as opaque “black boxes,” Switzerland’s Apertus — true to its Latin name meaning “open” — offers a window into a transparent, verifiable AI future.

https://de.wikipedia.org/wiki/Apertus

https://ethz.ch/de/news-und-veranstaltungen/eth-news/news/2025/09/medienmitteilung-apertus-ein-vollstaendig-offenes-transparentes-und-mehrsprachiges-sprachmodell.html

https://huggingface.co/collections/swiss-ai/apertus-llm

https://publicai.co/