"4000+ lines of architectural foundation code. Built by a developer obsessed with one vision: Reinforcement Learning agents that learn to play, fight, and cooperate in procedurally generated tactical worlds."
This isn't just another game engine. This is a Reinforcement Learning training ground disguised as a tactical game. Every architectural decision serves one ultimate purpose: teaching AI agents to think tactically.
- Currently building: Modular tile-map components for procedural level assembly
- Vision: Like Lego blocks - snap together terrain, obstacles, objectives to create infinite training scenarios
- Purpose: RL agents need diverse environments to learn robust strategies
- "The foundation bricks are being crafted - procedural generation comes next"
- Message-driven AI framework - perfect for RL agent communication
- Behavior composition system - agents can dynamically learn new abilities
- Promise-based decision making - async responses ideal for neural network inference
- Action retry mechanisms - natural exploration/exploitation for RL algorithms
- "Every message an agent sends is a learning opportunity"
- Action queuing system - RL agents can plan multi-step strategies
- Recursive action processing - complex tactical decisions broken into learnable components
- Conflict resolution - teaches agents to handle competitive scenarios
- State tracking - every action recorded for training data collection
- "Turn-based tactical thinking, perfect for RL step-by-step learning"
- Journal System - captures every game state change for training datasets
- Timestamp-based logging - frame-perfect training data for temporal learning
- Event tracking - complete observation space for RL algorithms
- "Every game session becomes a training dataset"
- Actor collections - manage multiple RL agents simultaneously
- Grid-based world - discrete action spaces perfect for RL
- Message broker - inter-agent communication for cooperative/competitive scenarios
- "Ready for multi-agent reinforcement learning experiments"
Message Broker Promise System:
# Actors don't just move - they negotiate
message = Message(sender="Player", body=intention_to_move)
promise = messenger.send_message(message, enemy_unit)
# Enemy can respond with counter-actions, blocks, or agreementsBehavior Composition Magic:
# Build complex NPCs through behavior stacking
unit.add_behaviour(Behaviours.DISCRETE_MOVER)
unit.add_behaviour(Behaviours.BUFFERED_MOVER)
unit.add_behaviour(Behaviours.AGGRESSIVE)
# This unit now moves tactically AND fights intelligently- Protocol-based architecture - every component has clear interfaces
- Type-safe generic collections - actors, behaviors, animations all properly typed
- Event-driven design - loosely coupled, highly extensible
- Zero hard dependencies - just Pygame and petname
- Behavior composition framework ready for ML integration
- Message-passing architecture perfect for reinforcement learning agents
- Action retry mechanisms - natural fit for exploration/exploitation algorithms
- State tracking infrastructure already built for training data collection
- Separation of concerns - rendering, logic, input, AI all decoupled
- Generic type system - T-bound protocols ensure type safety
- Factory patterns - level creation, name generation, behavior instantiation
- Command pattern implementation - every action is an object
- Tile-map Lego system → Algorithmic level assembly
- Static mazes → Infinite diverse training environments
- Hand-crafted levels → Procedural challenge generation
- OpenAI Gym environment wrapper
- PyTorch/TensorFlow agent interfaces
- Custom reward systems for tactical learning
- Multi-agent competitive/cooperative scenarios
- Curriculum learning - agents graduate through difficulty levels
- Meta-learning - agents that learn to learn new behaviors
- Neural behavior composition - AI that creates new AI behaviors
- Self-play tournaments - agents teaching each other
- Complete RL environment - not just a toy problem
- Real tactical complexity - multi-agent, partial observability, long-term planning
- Built-in data collection - every experiment generates rich training data
- Extensible architecture - add your own RL algorithms easily
- Beyond scripted NPCs - truly learning game characters
- Emergent gameplay - behaviors that surprise even the developer
- Academic-quality codebase - learn from production-ready RL architecture
We're building toward AGI in miniature - artificial agents that can:
- Learn complex strategies through trial and error
- Cooperate and compete with other agents
- Adapt to new environments and challenges
- Exhibit emergent tactical intelligence
What's Built:
- 4000+ lines of RL-ready architectural foundation
- Complete multi-agent message-passing system
- Training data collection infrastructure
- Sophisticated behavior composition framework
What's Coming:
- Tile-map procedural generation (currently in development)
- RL agent integration (the main goal)
- Neural network behavior learning (the dream)
What Makes This Special:
- RL-first mindset from day one - not bolted on afterward
- Academic rigor meets practical game development
- Built by someone obsessed with the vision of truly intelligent game AI
"This isn't a game with some AI. This is an AI laboratory that happens to look like a game."
Seeking RL enthusiasts and tactical AI dreamers who want to:
- Build the infrastructure for truly learning game agents
- Push beyond scripted behaviors into emergent intelligence
- Create training environments for the next generation of game AI
- Be part of something that bridges gaming and AI research
Ready to teach machines to think tactically?
Let's build the future of intelligent agents together.