|COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring.|
Evaluating User-Adaptive Systems: Lessons from Experiences with a Personalized Meeting Scheduling Assistant
If you have a question about this talk, please contact Microsoft Research Cambridge Talks Admins.
This event may be recorded and made available internally or externally via http://research.microsoft.com. Microsoft will own the copyright of any recordings made. If you do not wish to have your image/voice recorded please consider this before attending
We present experiences from evaluating the learning performance of a user-adaptive personal assistant agent. This work was part of the CALO project, which led to spin-out Siri acquired by Apple in 2010. We discuss the challenge of designing adequate evaluation and the tension of collecting adequate data without a fully functional, deployed system. Reflections on negative and positive experiences point to the challenges of evaluating user-adaptive agent systems. Lessons learned concern early consideration of evaluation and deployment, characteristics of AI technology and domains that make controlled evaluations appropriate or not, holistic experimental design, implications of “in the wild” evaluation, and the effect of AI-enabled functionality and its impact upon existing tools and work practices.
This talk is part of the Microsoft Research Cambridge, public talks series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
Other listscambridge architecture society Computer Laboratory Computer Architecture Group Meeting euroscicon
Other talksVarious TBC Sustainable Farming - making it happen. The Science, delivery and pitfalls of Integrated Farm Management The Affective Economy of Social Relations in Early Modern England Coffee&Cakes The 2015 Ageing Summit