AI alignment researcher
Jan Leike (born 1986 or 1987)[1] is an AI alignment researcher who has worked at DeepMind and OpenAI. He joined Anthropic in May 2024.
Education
Jan Leike obtained his undergraduate degree from the University of Freiburg in Germany. After earning a master's degree in computer science, he pursued a PhD in machine learning at the Australian National University under the supervision of Marcus Hutter.[2]
Career
Leike made a six-month postdoctoral fellowship at the Future of Humanity Institute before joining DeepMind to focus on empirical AI safety research,[2] where he collaborated with Shane Legg.[1]
OpenAI
In 2021, Leike joined OpenAI.[1] In June 2023, he and Ilya Sutskever became the co-leaders of the newly introduced "superalignment" project, which aimed to determine how to align future artificial superintelligences within four years to ensure their safety. This project involved automating AI alignment research using relatively advanced AI systems. At the time, Sutskever was OpenAI's Chief Scientist, and Leike was the Head of Alignment.[3][1] Leike was featured in Time's list of the 100 most influential personalities in AI, both in 2023[1] and in 2024.[4] In May 2024, Leike announced his resignation from OpenAI, following the departure of Ilya Sutskever, Daniel Kokotajlo and several other AI safety employees from the company. Leike wrote that "Over the past years, safety culture and processes have taken a backseat to shiny products", and that he "gradually lost trust" in OpenAI's leadership.[5][6][7]
In May 2024, Leike joined Anthropic, an AI company founded by former OpenAI employees.[8]
References
External links