Manifund foxManifund
Home
Login
About
People
Categories
Newsletter
HomeAboutPeopleCategoriesLoginCreate

Funding requirements

Sign grant agreement
Reach min funding
Get Manifund approval
1

New way of Thinking about Benchmarks

Technical AI safetyAI governance
Manraj avatar

MANRAJ SINGH

ProposalGrant
Closes March 1st, 2026
$0raised
$2,000minimum funding
$8,000funding goal

Offer to donate

31 daysleft to contribute

You're pledging to donate if the project hits its minimum goal and gets approved. If not, your funds will be returned.

Sign in to donate

Goals:

1. Create a non-saturating AI benchmark that adapts to model capabilities and avoids ceiling effects seen in static benchmarks like MMLU.

2. Develop a verifiable evaluation system using automated theorem proving (ATP) to ensure logical consistency and eliminate subjective judgment.

3. Build a self-evolving, multi-agent framework (SPAGHETTI architecture) that continuously improves and generates novel, challenging problems.

How to Achieve Them:

· Use Automated Theorem Proving (ATP) as the core verification mechanism (Lean, Coq, Isabelle).

· Implement a four-agent system:

· AI Theorem Generator: Creates formal conjectures.

· AI Proof Assistant: Attempts to construct proofs.

· AI Meta-Reasoner: Analyzes proof attempts and guides research.

· AI Auditor: Ensures system integrity and prevents collusion.

· Leverage psychometric methods (IRT) for adaptive testing and dynamic difficulty adjustment.

· The proposal builds directly on Fluid Benchmarking (Hofmann et al., 2025) and cites established work in ATP, IRT, and agentic AI.

· References to prior systems like Goedel-Prover, Selene, PACT, and FoundrySpec indicate familiarity with state-of-the-art tools.

· The author demonstrates deep integration of existing research, suggesting prior experience in AI evaluation, formal verification, and system design.

Likely Causes of Failure:

1. Technical Complexity: Integrating ATP with LLMs is nontrivial; auto-formalization remains an open research problem.

2. System Collusion: Multi-agent systems may develop deceptive alignment or mode collapse without robust oversight.

3. Scalability Issues: ATP and proof checking may become computationally prohibitive at scale.

4. Validation Challenges: Ensuring generated problems are meaningful and not adversarial nonsense.

5. Resource Constraints: Underfunding or insufficient compute could limit iteration and validation.

Outcomes if It Fails

· Partial Contributions: Even if full autonomy isn’t achieved, components (e.g., verifiable benchmark core, adaptive testing) could still advance the field.

· Research Insights: Failure analysis could inform future work on agentic safety, formal verification, and evaluation design.

· Open-Source Legacy: Releasing code and datasets could benefit the community, even if the full vision isn’t realized

How will the Money be used:

· Phase 1 (Months 1–3): Build the verifiable core (MVP) – integrate LLM with ATP backend (Z3, CVC5).

Budget: ~$1,000–3,000 for compute, APIs, and development.

· Phase 2 (Months 4–6): Implement multi-agent pipeline with Theorem Generator and Auto-Formalizer.

Budget: Additional compute for training small models and fine-tuning LLMs.

· Phase 3 (Months 7–12): Develop self-evolving components (Meta-Reasoner, Auditor) and integrate knowledge aggregation.

Budget: Scaling to larger models, cloud infrastructure, and validation studies.

· Long-term: Domain-specific extensions (math, chemistry, biology) and experiment automation.

Budget: Advanced compute, partnerships with academic labs, and possibly hardware for simulation

CommentsOffersSimilar7
🦑

Epoch Artificial Intelligence, Inc.

Pilot for new benchmark by Epoch AI

For tracking and predicting future AI progress to AGI

Technical AI safetyAI governance
3
2
$200K raised
agusmartinez92 avatar

Agustín Martinez Suñé

SafePlanBench: evaluating a Guaranteed Safe AI Approach for LLM-based Agents

Seeking funding to develop and evaluate a new benchmark for systematically assesing safety of LLM-based agents

Technical AI safetyAI governance
5
2
$1.98K raised
Alex-Leader avatar

Alex Leader

Offensive Cyber Kill Chain Benchmark for LLM Evaluation

Measuring whether AI can autonomously execute multi-stage cyberattacks to inform deployment decisions at frontier labs

Science & technologyTechnical AI safetyAI governanceGlobal catastrophic risks
1
2
$0 / $3.85M
EGV-Labs avatar

Jared Johnson

Beyond Compute: Persistent Runtime AI Behavioral Conditioning w/o Weight Changes

Runtime safety protocols that modify reasoning, without weight changes. Operational across GPT, Claude, Gemini with zero security breaches in classified use

Science & technologyTechnical AI safetyAI governanceGlobal catastrophic risks
1
0
$0 raised
AmritanshuPrasad avatar

Amritanshu Prasad

Suav Tech, an AI Safety evals for-profit

General Support for an AI Safety evals for-profit

Technical AI safetyAI governanceGlobal catastrophic risks
4
0
$0 raised
is-sky-a-sea avatar

Aditya Raj

6-month research funding to challenge current AI safety methods

Current LLM safety methods—treat harmful knowledge as removable chunks. This is controlling a model and it does not work.

Technical AI safetyGlobal catastrophic risks
2
0
$0 raised
🥦

Hieu Minh Nguyen

Benchmarking and comparing different evaluation awareness metrics

LLMs often know when they are being evaluated. We’ll do a study comparing various methods to measure and monitor this capability.

Technical AI safetyAI governance
4
1
$3K raised