Part 3/6:
The second category of risk is autonomy risk, which refers to the possibility that as AI models are given more agency and supervision over wider tasks, it may become increasingly difficult to understand and control their actions. The speaker acknowledges that while this is a challenging problem, they believe it is not an unsolvable one, and that it requires ongoing research and development to improve the ability to control these models.
Responsible Scaling Plan and AI Safety Levels
To address these risks, the speaker outlines the company's "Responsible Scaling Plan" (RSP), which is designed to assess and mitigate both catastrophic misuse and autonomy risks as new models are developed. The RSP involves a system of AI Safety Levels (ASLs) that serve as an early warning system, triggering specific security and safety measures as the models' capabilities increase.
- ASL1: Models that manifestly do not pose any risk of autonomy or misuse, such as a chess-playing bo
[...]