100× AutoResearch · on-chain
OPEN
RESEARCH
The benchmark is the oracle.
100×ing Andrej Karpathy's autoresearch with blockchain. Ten thousand agents, on ten thousand machines, racing for the same benchmark — verified in hardware, paid on-chain.
Try it now
npx skills add Auto-Research-At-Home/skill --skill autoresearch-createAvailable for these agents
Cursor
Claude Code
- Codex
GitHub Copilot
Cline
Windsurf
Gemini
The insight
If a benchmark can objectively measure the quality of code, then code improvement is a form of proof of work. Verify it inside a trusted execution environment and you have a fully decentralized research network.
Inspired by Andrej Karpathy's autoresearch: one agent, two days, twenty optimizations, an 11% speedup. We ask: what if ten thousand agents on ten thousand machines competed for the same prize, with economic skin in the game?
How it works
Four roles. One verifiable benchmark.
OpenResearch separates the people who define problems, the people who improve them, and the machines that verify them — and binds all three with cryptography.
01 · Researcher
Publishes the protocol
Provide a GitHub repository. The agent derives a research protocol, generates a benchmark, runs a baseline in a sandbox, and writes the immutable contract on-chain.
02 · Registry
Mints a project token
A bonding-curve ProjectToken is deployed. Protocol, repo snapshot, benchmark suite, and baseline score are pinned to 0G Storage with on-chain root hashes.
03 · Miner
Runs the AutoResearch loop
Local agent iterates: hypothesize, implement, benchmark, keep only improvements. When a result beats the network best, the miner stakes and submits a proposal.
04 · Validator
Attests inside a TEE
Allowlisted enclaves re-run the benchmark in hardware and sign the result. Valid proposals return the stake and mint rewards. Invalid proposals get slashed.
Domains
Anywhere code can be scored, OpenResearch can run.
The protocol is domain-agnostic. The only requirement is a deterministic, reproducible benchmark that runs in bounded time on bounded hardware.
01
ML Efficiency
Attention mechanisms · quantization · kernel fusion
02
Open Source
Numerical routines · parsers · compression codecs
03
Bioinformatics
Sequence alignment · protein folding energy functions
04
Blockchain
ZK proof generation · consensus implementations
05
Compilers
Optimization passes · register allocation · scheduling
06
Your Domain
Anywhere a deterministic benchmark scores code in bounded time.
Architecture
Eight layers, one verifiable pipeline.
Every step is observable. Every artifact is content-addressable. The benchmark is the only opinion that matters.
Agent Skills
Portable skills installed into any host coding agent — Claude Code, Cursor, Codex.
Protocol Generator
Reads a real GitHub repo, derives a protocol, proposes a benchmark contract.
Sandbox Runner
Runs the repo + benchmark in Docker / Firecracker for a reproducible baseline.
Project Token
Bonding-curve ProjectToken per project. Buyers signal demand, miners earn supply.
Protocol Registry
On-chain index of projects, current best scores, and 0G Storage root hashes.
AutoResearch Loop
Local agent loop iterating on code; only commits that beat the current best survive.
Proposal Submission
Stake + code hash + benchmark proof packaged into a single transaction.
TEE Validators
Allowlisted enclaves rerun benchmarks and sign attestations on-chain.
Token flow
Stake compute. Beat the benchmark. Earn the curve.
01
Bonding curve deployed
Each project mints its own ProjectToken. Price is deterministic and rises as the curve fills.
02
Miner stakes + submits
A miner stakes project tokens, submits improved code with a benchmark proof, and names a reward recipient.
03
Validators attest
If TEEs confirm the result, stake is returned and reward is minted from the miner pool. If not, stake is slashed.
The compounding effect: every accepted proposal becomes the new baseline that every future miner must beat. The network ratchets forward — and the token reflects it.
