BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//talks.cam.ac.uk//v3//EN
BEGIN:VTIMEZONE
TZID:Europe/London
BEGIN:DAYLIGHT
TZOFFSETFROM:+0000
TZOFFSETTO:+0100
TZNAME:BST
DTSTART:19700329T010000
RRULE:FREQ=YEARLY;BYMONTH=3;BYDAY=-1SU
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:+0100
TZOFFSETTO:+0000
TZNAME:GMT
DTSTART:19701025T020000
RRULE:FREQ=YEARLY;BYMONTH=10;BYDAY=-1SU
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
CATEGORIES:Darwin College Humanities and Social Sciences Semi
 nars
SUMMARY:The Social Impact of Automatic Hate Speech Detecti
 on - Dr. Stefanie Ullmann\, Postdoctoral Research 
 Associate\, Centre for Humanities and Social Chang
 e
DTSTART;TZID=Europe/London:20191022T131000
DTEND;TZID=Europe/London:20191022T140000
UID:TALK133069AThttp://talks.cam.ac.uk
URL:http://talks.cam.ac.uk/talk/index/133069
DESCRIPTION:In this talk\, I explore quarantining as a more et
 hical method for delimiting the spread of Hate Spe
 ech via online social media platforms. Currently\,
  companies like Facebook\, Twitter\, and Google ge
 nerally respond reactively to such material: offen
 sive messages that have already been posted are re
 viewed by human moderators if complaints from user
 s are received. The offensive posts are only subse
 quently removed if the complaints are upheld\; the
 refore\, they still cause the recipients psycholog
 ical harm. In addition\, this approach has frequen
 tly been criticised for delimiting freedom of expr
 ession\, since it requires the service providers t
 o elaborate and implement censorship regimes. In t
 he last few years\, an emerging generation of auto
 matic Hate Speech detection systems has started to
  offer new strategies for dealing with this partic
 ular kind of offensive online material. Anticipati
 ng the future efficacy of such systems\, the prese
 nt article advocates an approach to online Hate Sp
 eech detection that is analogous to the quarantini
 ng of malicious computer software. If a given post
  is automatically classified as being harmful in a
  reliable manner\, then it can be temporarily quar
 antined\, and the direct recipients can receive an
  alert\, which protects them from the harmful cont
 ent in the first instance. The quarantining framew
 ork is an example of more ethical online safety te
 chnology that can be extended to the handling of H
 ate Speech. Crucially\, it provides flexible optio
 ns for obtaining a more justifiable balance betwee
 n freedom of expression and appropriate censorship
 .\n
LOCATION:The Richard King Room\, Darwin College
CONTACT:Dr Nanna K L Kaalund
END:VEVENT
END:VCALENDAR
