Skip to main content

CIRCLE: A framework for evaluating AI from a real-world lens.

Westling, C., Schwartz, R., Briggs, M., Carlyle, M., Holmes, M., Fadaee, M., Waters, G., Taik, A. and Lacerda, T., 2026. CIRCLE: A framework for evaluating AI from a real-world lens. In: Intellisys 2026, 3-4 September 2026, Amsterdam, The Netherlands. (In Press)

Full text available as:

[thumbnail of CIRCLE_final.pdf] PDF
CIRCLE_final.pdf - Accepted Version
Restricted to Repository staff only until 30 September 2026.
Available under License Creative Commons Attribution Non-commercial.

1MB

Official URL: https://saiconference.com/IntelliSys

Abstract

Most AI evaluations rely on static benchmarks that measure model outputs in isolation, offering little evidence about how systems behave once embedded in real-world workflows. As a result, decision- makers lack systematic evidence about downstream effects, operational risks, and long-term impacts that matter for deployment, governance, and procurement. We introduce CIRCLE, a six-stage lifecycle-based framework that links stakeholder concerns to context-sensitive evaluation methods, longitudinal measurement, and ongoing monitoring of deployed AI systems. The framework integrates evaluation methods such as A/B testing, field testing, red teaming, and longitudinal studies into a coordinated evaluation pipeline rather than treating them as isolated activities. Together, these methods support more contextualized, iterative, and decision-relevant assessments of AI systems. By aligning constructs, methods, and metrics with real deployment contexts, CIRCLE supports more actionable, iterative, and governance-relevant evaluation of AI systems and their secondary and tertiary effects.

Item Type:Conference or Workshop Item (Paper)
Group:Faculty of Media, Science and Technology
ID Code:41721
Deposited By: Symplectic RT2
Deposited On:12 Feb 2026 15:54
Last Modified:12 Feb 2026 15:54

Downloads

Downloads per month over past year

More statistics for this item...
Repository Staff Only -