|COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring.|
Evaluating User-Adaptive Systems: Lessons from Experiences with a Personalized Meeting Scheduling Assistant
If you have a question about this talk, please contact Microsoft Research Cambridge Talks Admins.
This event may be recorded and made available internally or externally via http://research.microsoft.com. Microsoft will own the copyright of any recordings made. If you do not wish to have your image/voice recorded please consider this before attending
We present experiences from evaluating the learning performance of a user-adaptive personal assistant agent. This work was part of the CALO project, which led to spin-out Siri acquired by Apple in 2010. We discuss the challenge of designing adequate evaluation and the tension of collecting adequate data without a fully functional, deployed system. Reflections on negative and positive experiences point to the challenges of evaluating user-adaptive agent systems. Lessons learned concern early consideration of evaluation and deployment, characteristics of AI technology and domains that make controlled evaluations appropriate or not, holistic experimental design, implications of “in the wild” evaluation, and the effect of AI-enabled functionality and its impact upon existing tools and work practices.
This talk is part of the Microsoft Research Cambridge, public talks series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
Other listsRollo Davidson Lectures CUCRS Looking at Language Acquisition (LALA) XIII - A meeting of Essex and Cambridge PhD students
Other talksInferno XV, Purgatorio XV, Paradiso XV Controlling Cancer Summit Meditation Session at the Botanical Gardens William Courten (1642–1702) and natural history Numerical Methods for (Quasi)Variational Inequalities - Part II Pre-Processing for Approximate Bayesian Computation in Image Analysis