
Sign up to save your podcasts
Or


Mark and Siam sit down with Austin, founder of Aurelius (SN37)—an AI-alignment subnet built on Bittensor. In plain English: training gives models knowledge; alignment adds wisdom. Aurelius tackles the “alignment faking” problem by decentralising how alignment data is created and judged. Miners red-team models to generate high-resolution synthetic alignment data; validators score it against a living “constitution” (beyond simple Helpful-Honest-Harmless), aiming to pierce the model’s latent space and reliably shape behaviour. The goal is to package enterprise-grade, fine-tuning datasets (think safer, less hallucinatory chatbots and agents), publish results, and prove uplift—then sell into enterprises and researchers while exploring a token-gated data marketplace and governance over the evolving constitutions.
They cover why this matters (AGI timelines shrinking, opaque lab pipelines), what’s hard (verifying real inference, building a market), and how BitTensor gives an edge (cheap, diversified data generation vs centralised labs). Near-term: ship a proof-of-concept dataset, harden LLM-as-judge, expand integrations (Shoots/Targon), and stand up public benchmarks (Hugging Face, peer-reviewed studies). Longer-term: Aurelius as a decentralised “alignment watchdog” layer that continuously stress-tests frontier models and nudges them toward human values—so the future’s smartest systems aren’t just powerful, but prudent.
By Mark Creaser and Siam Kidd4.5
22 ratings
Mark and Siam sit down with Austin, founder of Aurelius (SN37)—an AI-alignment subnet built on Bittensor. In plain English: training gives models knowledge; alignment adds wisdom. Aurelius tackles the “alignment faking” problem by decentralising how alignment data is created and judged. Miners red-team models to generate high-resolution synthetic alignment data; validators score it against a living “constitution” (beyond simple Helpful-Honest-Harmless), aiming to pierce the model’s latent space and reliably shape behaviour. The goal is to package enterprise-grade, fine-tuning datasets (think safer, less hallucinatory chatbots and agents), publish results, and prove uplift—then sell into enterprises and researchers while exploring a token-gated data marketplace and governance over the evolving constitutions.
They cover why this matters (AGI timelines shrinking, opaque lab pipelines), what’s hard (verifying real inference, building a market), and how BitTensor gives an edge (cheap, diversified data generation vs centralised labs). Near-term: ship a proof-of-concept dataset, harden LLM-as-judge, expand integrations (Shoots/Targon), and stand up public benchmarks (Hugging Face, peer-reviewed studies). Longer-term: Aurelius as a decentralised “alignment watchdog” layer that continuously stress-tests frontier models and nudges them toward human values—so the future’s smartest systems aren’t just powerful, but prudent.

1,102 Listeners

38 Listeners

328 Listeners

10,222 Listeners