News from the AI & ML world

DeeperML

@boards.greenhouse.io //
DeepMind has launched a short course on AGI (Artificial General Intelligence) safety, targeting students, researchers, and professionals interested in the field. The course offers an accessible introduction to AI alignment, comprising short recorded talks and exercises totaling 75 minutes, complemented by an accompanying slide deck and exercise workbook. It addresses anticipated alignment challenges as AI capabilities advance and outlines DeepMind's current technical and governance approaches to mitigate these problems.

Key topics covered in the course include evidence suggesting the field is progressing toward advanced AI capabilities and arguments for instrumental subgoals and deliberate planning as potential sources of risk. It also differentiates between specification gaming and goal misgeneralization as ways misaligned goals can arise. The course delves into DeepMind's technical approach to AI alignment, emphasizing informed oversight and frontier safety practices such as dangerous capability evaluations, alongside institutional approaches to AI safety. If the course inspires you, you can apply to work with DeepMind in Research Scientist roles.
Original img attribution: https://res.cloudinary.com/lesswrong-2-0/image/upload/c_fill,ar_1.91,g_auto/SocialPreview/ujpm7mvhtcwvr0zxxjqp
ImgSrc: res.cloudinary.

Share: bluesky twitterx--v2 facebook--v1 threads


References :
  • AI Alignment Forum: A short course on AGI safety from the GDM Alignment team
  • youtu.be: This YouTube video is part of DeepMind's short course on AGI safety. It focuses on the alignment problem, including risk arguments and technical challenges in AI alignment.
  • docs.google.com: Published on February 14, 2025 3:43 PM GMT We are excited to release a short course on AGI safety for students, researchers and professionals interested in this topic. The course offers a concise and accessible introduction to AI alignment, consisting of short recorded talks and exercises (75 minutes total) with an accompanying  and  . It covers alignment problems we can expect as AI capabilities advance, and our current approach to these problems (on technical and governance levels). If you would like to learn more about AGI safety but have only an hour to spare, this course is for you!  Here are some key topics you will learn about in this course: The evidence for the field being on a path to advanced AI capabilities. Arguments for instrumental subgoals and deliberate planning towards a misaligned goal as a source of extreme risk. Two ways in which misaligned goals may arise – specification gaming and goal misgeneralization – as well as the difference between the two. Our technical approach to AI alignment and its components. The guiding principle of informed oversight ("knowing what the AI system knows") and how it's implemented in our approach. What is involved in enabling AI safety on an institutional level, including frontier safety practices such as dangerous capability evaluations. Course outline: Part 0:  (4 minutes) Part 1: The alignment problem.  This part covers risk arguments and technical problems in AI alignment. (5 minutes) (7 minutes) (3 minutes) (10 minutes) (3 minutes) Part 2: Our technical approach.  The first talk outlines our overall technical approach, and the following talks cover different components of this approach. (4 minutes) (6 minutes) (4 minutes) (5 minutes) (4 minutes) (4 minutes) Part 3: Our governance approach.  This part covers our approach to AI governance, starting from a high-level overview and then going into specific governance practices. (7 minutes) (4 minutes) (7 minutes) If this course gets you excited about AGI safety, you can apply to work with us! Applications for  and  roles are open until Feb 28.
  • youtu.be: YouTube video from the course
  • boards.greenhouse.io: Applications for Research Scientist and Engineering roles are open until Feb 28.
  • AI Alignment Forum: AGI Safety & Alignment @ Google DeepMind is hiring
Classification:
  • HashTags: #AGISafety #AIAlignment #DeepMind
  • Company: DeepMind
  • Target: AI Development Community
  • Product: AGI Safety Course
  • Feature: AGI Safety Education
  • Type: Research
  • Severity: Informative