Newsletter 12

Newsletter 12

Last week, we hosted the first event of the AI Safety Talks series in collaboration with inzva!

We teamed up with inzva for insightful discussions on AI safety featuring our co-founders, Bengüsu Özcan and Berke Çelik, followed by a thought-provoking session by Erdem Bıyık on Reinforcement Learning from Human Feedback.

🎥 You can watch the session here!

ANNOUNCEMENTS 🔊

SERI MATS Program

SERI MATS, arguably the most prestigious research upskilling program in AI Safety has opened applications Independent research and educational seminar program connecting talented scholars with top mentors in the fields of AI alignment, interpretability, and governance. For 10 weeks, MATS scholars will conduct research while also attending talks, workshops, and networking events with other members of the Berkeley alignment research community.

🗓️ Register by: April 18th

Economics of Transformative AI Hackathon

Join Apart Research’s new hackathon that brings technical insights with policy ideas: what the economy should look like in the age of transformative AI?

🗓️ Register by: April 25th

Bluedot Impact Courses

Join 9-12 weeks of online training and discussion sessions, tailored for cutting-edge, transformative AI topics. Choose the topic for your background: technical AI alignment, AI governance, or transformative AI economics.

🗓️ Rolling applications

Networking for impact in AI Safety

Networking event from BlueDot Impact aimed at sparking meaningful conversations, uncovering powerful collaborations, and accelerating your personal and collective impact in the field. The session will blend purposeful 1-on-1 conversations with speed networking structured based on attendee input.

🗓️ Register by: April 17th

AI Alignment Evals Hackathon

We want AIs to have certain values in robust and scalable ways. There are methods proposed to get these values into models, but it’s highly ambiguous how well they actually work. This week-long hackathon from AI-Plans aims to help remove that ambiguity.

🗓️ Register by: April 25th

LASR Labs

Research program working towards reducing the risk of loss of control from advanced AI, focusing on action-relevant questions tackling concrete threat models. Participants are matched into teams of 3-4 and work with a supervisor to write an academic-style paper, with support and management from LASR. The program will be in person, in London.

🗓️ Register by: April 26th

MAISU - Minimal AI Safety Unconference

A flexible, participant-driven unconference for those working to prevent AI catastrophe, organized by the AI Safety Camp team.

🗓️ Register by: April 18th

AI Futures World-building Course

Foresight Institute just launched a free, self-paced course designed to help more people think clearly and creatively about how AI might shape the future. This course is for people who want to explore how AI might shape our shared future and what we can do about it. It is designed to be accessible and thought-provoking, with no technical background required.

🗓️ Register by: April 17th

TOP PICKS 📑 🎧

GATE: Modeling the Trajectory of AI and Automation

If you’re seeking credible analysis on AI progress, Epoch AI continues to be our recommended source. In their new paper, Ege Erdil et. al builds a general framework for modelling the relationship between AI progress and its economic impacts, and provides forecasts regarding how AI will transform the economy. Their research shows that full automation is quite plausible within a decade, and even with conservative assumptions, full automation seems quite likely within two decades.

AI 2027: A Plausible Scenario For Super Intelligence In This Decade

A group of researchers, including ex-OpenAI employee Daniel Kokotajlo known for his accurate forecasts for frontier AI development so far, has published an extensive scenario of a technical and geopolitical pathway that leads up to artificial super intelligence in this decade. This project provides a great opportunity for you to inform your own opinion: what are the dependencies, where you agree and disagree.