Join our community and participate in AI safety research and learning
Subscribe to Events CalendarThe Mechanistic Interpretability Course is an intensive 1-month program focused on techniques for understanding the internal mechanisms of neural networks. This course combines theoretical learning with hands-on projects.
Mechanistic interpretability is a key area of AI safety research, aiming to make AI systems more transparent and understandable.
The AGI Safety Fundamentals cohort is an 8-week guided course covering the essential concepts in AI alignment and safety. Participants read selected materials and meet weekly to discuss the readings with a facilitator.
This program is based on the AGI Safety Fundamentals curriculum by BlueDot and provides a structured introduction to the field of AI safety.
Our Weekly Discussion Group provides a casual forum for discussing recent papers, concepts, and developments in AI safety. These sessions are open to anyone interested in the field, regardless of prior knowledge.
Each week features a different topic, announced in advance through our mailing list and Telegram group.
No registration is required. Simply show up! If you're attending for the first time, we recommend arriving 10 minutes early to meet the organizers.
The Paper Reading Club conducts deep dives into foundational and recent papers in AI safety research. Unlike the more casual discussion group, this activity involves a thorough examination of specific research papers.
Participants are expected to read the selected paper in advance and come prepared to discuss its methods, results, and implications.