Jash Vira
Does harmful fine-tuning data cause broad misalignment only when the model already recognises the target behaviour as a norm violation?
yesh
Mohamed Mohamed Ramadan
AI infrastructure for turning natural language hardware ideas into manufacturable devices
Christian Lysenstøen
Building an AI research agent that can propose, test, and write up small ML findings
Dev D. Goyal
Small API budget to test whether semantic reporting failures in an OpenAI-validated VCG market simulation replicate across Claude and Gemini.
Avyukt Jindal
Joseph A Wecker
First-principles theory of the adaptive feedback loop — one structural account governing Kalman filters, RL agents, software teams, and frontier LLMs alike.
Lucas Kempe
Make "which tool did the agent pick, and why?" an inspectable artifact instead of a vector lookup.
Nathan Thornhill
An ORCID-gated submission pipeline where a multi-model AI panel plus quality-control layer delivers rigorous peer review without institutional gatekeeping.
Andwar Cheng
Open-source replay tool and benchmark prototype for identifying step-level semantic risk transitions in multi-agent AI traces — CI-backed and independently repr
Aashka Patel
Inspiring India’s Middle‑Schoolers to pursue AI Safety, Governance, and X‑Risk Work
Joshua Michael Sparks
Stage 0 bridge for a focused-ultrasound research program targeting the neural architecture that maintains chronic suffering.
Vangelis Gkagkelis
A 6-month pilot testing probabilistic forecasting for AI, misinformation, institutional trust, and social risk in Greece.
Jessica Pu Wang
Germany’s talents are critical to the global effort of reducing catastrophic risks brought by artificial intelligence.
Matthew A Cator
Funding the open-source launch of a working claim-state system and the local firewall bridge that carries verification before voice into governed agent action.
José Wheeler
Identifying and auditing reasoning circuits in LLMs within Algoverse 2026 using Sparse Autoencoders (SAEs).
Ida-Emilia Kaukonen
A 15,000+ page corpus on long-term interaction, symbolic language, unusual model behavior, and safety edge cases.
Sardor Razikov
Measuring k*: the phase transition where LLMs surrender independent reasoning under authority. 7 frontier models tested, all data open.
Francisco Antonio Da Costa Barroso
Independent researcher in Brazil scaling a validated sparse architecture to 1B, plus open interpretability tooling for expert routing. 6-month runway grant.
Zaelani
18+ preprints across multiple fields, all written on a 2GB RAM phone. $600 removes the only thing standing between me and the next body of work.