Skip to content
MDU
Publication 2026

Open Benchmarks for Civic LLMs

  • Evaluation
  • Responsible AI
  • Public Services

This publication proposes a benchmark suite for evaluating language models used in public services.

We emphasize:

  • failure modes that matter for citizens,
  • documentation that procurement teams can read,
  • and audit trails that make evaluation reproducible.
Partners
  • Municipal Innovation Office
Contact
Note
Citations, downloadable artifacts, and a versioned record of outputs are available on request—contact our press office at press@mdu.edu.kg.