👋 We are an AI alignment research collective investigating how fundamental principles from biology and economics can inform safer, more aligned AI systems.
Our work centres on homeostasis, multi-objective balancing, sustainability, and universal human values — drawing from nature's time-tested strategies for maintaining equilibrium — to develop benchmarks that expose dangerous failure modes in current AI approaches.
We also research frameworks that mitigate these risks. We believe that shifting AI design from "maximise forever" toward "maintain a healthy equilibrium" is a crucial and underexplored part of the alignment solution space.
- Alignment with fundamental biological & economical principles
- Homeostatic bounded objectives
- Multi-objective balancing (bounded & unbounded objectives)
- Concave utility functions
- Universal human values
- Runaway conditions — benchmarking & mitigation
- Multi-objective multi-agent extended gridworlds
- Sustainability
- Proactive horizon scanning of side effects
- Accountability mechanisms and whitelisting