BEGIN:VCALENDAR
VERSION:2.0
PRODID:researchseminars.org
CALSCALE:GREGORIAN
X-WR-CALNAME:researchseminars.org
BEGIN:VEVENT
SUMMARY:Olle Häggström (Chalmers University of Technology\, Sweden)
DTSTART:20230822T140000Z
DTEND:20230822T150000Z
DTSTAMP:20260423T005715Z
UID:Wisla23/3
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/Wisla23/3/">
 AI risk and AI alignment</a>\nby Olle Häggström (Chalmers University of 
 Technology\, Sweden) as part of Mapping the Interdisciplinary Horizons of 
 AI: Safety\, Functional Programming\, Information Geometry\n\n\nAbstract\n
 The planetary dominance over other species that humanity has attained has 
 very little to do with muscular strength and physical endurance: it is all
  about intelligence. This makes the present moment in history\, when we ar
 e automating intelligence and handing over this crucial skill to machines\
 , the most important ever. The research area that has become known as AI a
 lignment deals with how to make sure that the first superintelligent machi
 nes have goals and values that are sufficiently aligned with ours and that
  sufficiently prioritize human flourishing. This needs to succeed\, becaus
 e otherwise we face existential catastrophe. In these lectures I will outl
 ine key challenges in AI alignment\, what is being done to solve them\, an
 d how all this relates to the breakneck speed at which AI is presently adv
 ancing.\n
LOCATION:https://researchseminars.org/talk/Wisla23/3/
END:VEVENT
END:VCALENDAR
