|COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring.|
Evaluating User-Adaptive Systems: Lessons from Experiences with a Personalized Meeting Scheduling Assistant
If you have a question about this talk, please contact Microsoft Research Cambridge Talks Admins.
This event may be recorded and made available internally or externally via http://research.microsoft.com. Microsoft will own the copyright of any recordings made. If you do not wish to have your image/voice recorded please consider this before attending
We present experiences from evaluating the learning performance of a user-adaptive personal assistant agent. This work was part of the CALO project, which led to spin-out Siri acquired by Apple in 2010. We discuss the challenge of designing adequate evaluation and the tension of collecting adequate data without a fully functional, deployed system. Reflections on negative and positive experiences point to the challenges of evaluating user-adaptive agent systems. Lessons learned concern early consideration of evaluation and deployment, characteristics of AI technology and domains that make controlled evaluations appropriate or not, holistic experimental design, implications of “in the wild” evaluation, and the effect of AI-enabled functionality and its impact upon existing tools and work practices.
This talk is part of the Microsoft Research Cambridge, public talks series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
Other listsConservation seminars women@CL all Wolfson Research Event
Other talksDevelopment and the cell cycle Histopathology 2014: Advances in research and techniques Production Processes Group Seminar - "Drop Impact on to Moving Liquid Pools " Title 'TBA' Free webinar - Total information risk management - making the business case for data quality programmes Molecular excited states and QMC