Publication 2026
Open Benchmarks for Civic LLMs
- Evaluation
- Responsible AI
- Public Services
This publication proposes a benchmark suite for evaluating language models used in public services.
We emphasize:
- failure modes that matter for citizens,
- documentation that procurement teams can read,
- and audit trails that make evaluation reproducible.
Partners
- Municipal Innovation Office
Contact
Note
Citations, downloadable artifacts, and a versioned record of outputs are available on request—contact our press office at press@mdu.edu.kg.