AI Safety Camp connects you with a research lead to collaborate on a project – to see where your work could help ensure future AI is safe.
Apply to collaborate online from January to April 2024.
We value diverse backgrounds. Many roles but definitely not all require some knowledge in one of: AI safety, mathematics or machine learning.
Click to see skills needed to join various projects.
- Art, design, photography
- Humanistic academics
- Legal expertise
- Project management
- Interpretability methods
- Using LLMs
- Reading scientific papers
- Know scientific methodologies
- Think and work independently
- Familiarity of AI risk research landscape
Applications have closed for all projects for the 2024 edition.
To be notified if 2025 edition applications open, please leave your email address here.
To not build uncontrollable AI
Projects to restrict corporations from recklessly scaling the training and uses of ML models. Given controllability limits.
- 1. Towards realistic ODDs for foundation model based AI offerings Closed
2. Luddite Pro: information for the refined luddite Closed
3. Lawyers (and coders) for restricting AI data laundering Closed
4. Assessing the potential of congressional messaging campaigns for AIS Closed
Diverse other projects, including technical control of AGI in line with human values.
5. Modelling trajectories of language models Closed
6. Towards ambitious mechanistic interpretability Closed
7. Exploring toy models of agents Closed
8. High-level mechanistic interpretability and activation engineering library Closed
9. Out-of-context learning interpretability Closed
10. Understanding search and goal representations in transformers Closed
- Evaluating and Steering Models
11. Benchmarks for stable reflectivity Closed
12. SADDER: situational awareness datasets for detecting extreme risks Closed
13. TinyEvals: how language models speak coherent English? Closed
14. Evaluating alignment evaluations Closed
15. Pipelines for evaluating and steering LLMs towards faithful reasoning Closed
16. Steering of LLMs through addition of activation vectors with latent ethical valence Closed
- Agent Foundations
17. High actuation spaces Closed
18. Does sufficient optimization imply agent structure? Closed
19. Discovering agents in raw bytestreams Closed
20. The science algorithm Closed
- Miscellaneous Alignment Methods
21. SatisfIA – AI that satisfies without overdoing it Closed
22. How promising is automating alignment research? (literature review) Closed
23. Personalized fine-tuning token for AI value alignment Closed
24. Self-other overlap @AE Studio Closed
25. Asymmetric control in LLMs: model editing and steering that resists control for unalignment Closed
26. Tackling key challenges in Debate Closed
27. AI-driven economic safety nets: restricting the macroeconomic disruptions of AGI deployment Closed
28. Policy-based access to powerful models Closed
29. Organise the next Virtual AI Safety Unconference Closed
Please write your application with the research lead of your favorite project in mind. Research leads will directly review applications this round. We organizers will only assist when a project receives an overwhelming number of applications.
Apply if you…
- want to consider and try out roles for helping ensure future AI function safely;
- are able to explain why and how you would contribute to one or more projects;
- previously studied a topic or trained in skills that can bolster your new team’s progress;
- can join weekly team calls and block out 5 hours of work each week from January to April 2024.
Dec 1-22: Interviews. You may receive an email for an interview, from one or more of the research leads whose project you applied for.
By 28 Dec: Final decisions. You will definitely know if you are admitted. Hopefully we can tell you sooner, but we pinky-swear we will by 28 Dec.
Jan 13-14: Opening weekend. First meeting with your teammates and one-on-one chats.
Jan 15 – Apr 28: Research is happening. Teams meet weekly, and plan in their own work hours.
April 25-28: Final presentations spread over four days.
For as long as you want: Some teams keep working together after the official end of AISC.
When you start the project, we recommend that you don’t make any commitment beyond the official length of the program. However if you find that you work well together as a team, we encourage you to keep going even after AISC is officially over.
Every team will have:
- one Research Lead (RL)
- one Team Coordinator (TC)
- other team members
All team members are expected to work at least 5 hours per week on the project (this number can be higher for specific projects), which includes joining weekly team meetings, and communicating regularly with other team members about their work.
Research Lead (RL)
The RL is the person behind the research proposal. They will guide the research project, and keep track of relevant milestones. When things inevitably don’t go as planned (this is research after all) the RL is in charge of setting the new course.
The RL is part of the research team and will be contributing to research the same as everyone else on the team.
Team Coordinator (TC)
The TC is the ops person of the team. If you are the TC then you are in charge of making sure meetings are scheduled, checks in with individuals on their task progress, etc.
The role of the TC is important but not expected to take too much time (except for project management-heavy teams). Most of the time, the TC will act like a regular team member contributing to the research, same as everyone else on the team.
Each project proposal states whether the looking for someone like you to take on this role.
Other team members
Other team members will work on the project under the guidance of the RL and the TC. Team members will be selected based on relevant skills, understandings and commitments to contribute to the research project.
Check out our frequently asked questions, in case you can find the answer there.
- For questions on a project, please contact the research lead. Find their contact info at the bottom of their project doc.
- For questions about the camp in general, or if you can’t reach the specific research lead, please email firstname.lastname@example.org.
May take 5 days for organizers to reply.
We are fundraising!
Organizers are volunteering this round, since we had to freeze our salaries. This is not sustainable.
To make next editions happen, consider making a donation. For larger amounts, feel free to email Remmelt.
See our funding case here.