Jeffrey Brunner
Mathematical enforcement of AI behavioral boundaries at deployment time. 0% attack success rate on 1,300 adversarial prompts vs 30.8% baseline.
Mateusz Bagiński
One Month to Study, Explain, and Try to Solve Superintelligence Alignment
Boyd Kane
by buying gift cards for the game and handing them out at the OpenAI offices
Igor Labutin
A tech-infused immersive musical. Experience the future of storytelling where artificial intelligence meets the depths of human emotion.
Jacob Steinhardt
Krishna Patel
Expanding proven isolation techniques to high-risk capability domains in Mixture of Expert models
Lawrence Wagner
Finn Metz
Funding 5–10 AI security startups through Seldon’s second SF cohort.
Sara Holt
Short Documentary and Music Video
Avinash A
Formalizing the "Safety Ceiling": An Agda-Verified Impossibility Theorem for AI Alignment
Sean Peters
Measuring attack selection as an emergent capability, and extending offensive cyber time horizons to newer models and benchmarks
Preeti Ravindra
AI Safety Camp 2026 project: Bidirectional Failure modes between security and safety
Parker Whitfill
Mirco Giacobbe
Developing the software infrastructure to make AI systems safe, with formal guarantees
Gergő Gáspár
Help us solve the talent and funding bottleneck for EA and AIS.
Alex Leader
Measuring whether AI can autonomously execute multi-stage cyberattacks to inform deployment decisions at frontier labs
Ella Wei
A prototype safety engine designed to relieve the growing AI governance bottleneck created by the EU AI Act and global compliance demands.
Chris Canal
Enabling rapid deployment of specialized engineering teams for critical AI safety evaluation projects worldwide
Miles Tidmarsh
Training AI to generalize compassion for all sentient beings using pretraining-style interventions as a more robust alternative to instruction tuning