University of Cambridge > Talks.cam > pb400's list > Perception Model for people with Visual Impairments

Perception Model for people with Visual Impairments

Add to your list(s) Download to your calendar using vCal

If you have a question about this talk, please contact Dr Pradipta Biswas.

Scientists from many different disciplines (including physiology, psychology, and engineering) have worked on modelling visual perception. However this field has been less extensively studied in the context of computer science, as most existing perception models work only for very specific domains such as menu searching or icon searching tasks. We are developing a perception model that works for any application. It takes a list of mouse events, a sequence of bitmap images of an interface and locations of different objects in the interface as input, and produces a sequence of eye-movements as output. We have identified a set of features to differentiate among different screen objects and using those features, our model has reproduced the results of previous experiments on visual perception in the context of HCI . It can also simulate the effects of different visual impairments on interaction. Our model of visual impairments are more detailed and accurate than the previous models and show probable interaction pattern of visually impaired users.

This talk is part of the pb400's list series.

Tell a friend about this talk:

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2020 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity