Anthropic hires former OpenAI researcher to set up 'Super Alignment Team' to strengthen AI safety and security



Anthropic, an AI company developing chat AI such as 'Claude,' has hired

Jan Reich , a former AI researcher at OpenAI, and established a new 'Super Alignment Team' focused on AI safety and security.

Anthropic hires former OpenAI safety lead to head up new team | TechCrunch
https://techcrunch.com/2024/05/28/anthropic-hires-former-openai-safety-lead-to-head-up-new-team/



Anthropic's goal is to 'emphasize safety more than OpenAI,' and the newly established Super Alignment Team will focus on various aspects of AI safety and security, particularly 'scalable supervision,' ' weak-to-strong generalization, ' and 'automated alignment research.'




According to the people, Reich will report to Anthropic's chief scientific officer, Jared Kaplan, and Anthropic researchers working on scalable surveillance at the time of writing will join Reich's team as the team is formed.




Reich previously led the SuperAlignment team at OpenAI with Ilya Satskivar, but will leave the company in May 2024, citing 'disagreements with OpenAI's leadership and core priorities.'




He also said, 'Over the past few years, safety culture and process has lagged behind flashy products,' and suggested that 'OpenAI must become a safety-first AGI company.'




Having lost key AI researchers such as Reich and Satskivar, OpenAI disbanded the SuperAlignment team in May 2024.

OpenAI's 'Super Alignment' team, which was researching the control and safety of superintelligence, has been disbanded, with a former executive saying 'flashy products are being prioritized over safety' - GIGAZINE

in Software, Posted by log1r_ut