START HERE: New to this repository? Read this file first. Created: 2025-10-22 Purpose: Single entry point for AI agents and human contributors
Repository: D&D Session Processor (VideoChunking) Purpose: Automated transcription, speaker diarization, and campaign management for tabletop RPG sessions Stack: Python 3.10+, Gradio (UI), faster-whisper (transcription), PyAnnote (diarization), Ollama (LLM) Current State: Production-ready core pipeline, active feature development
Total onboarding time: ~45-60 minutes for full context
Read these files first to understand what this project does and how we work:
File: CLAUDE.md
- What to learn: Comprehensive guide covering codebase structure, workflows, conventions, and common tasks
- Why read this: All-in-one reference designed specifically for AI assistants
- Key sections: "Operator Workflow", "Development Standards", "Common Tasks Reference"
File: docs/PROJECT_SUMMARY.md
- What to learn: What this project does, key features, technology stack
- Why read this: You need to understand the problem space before touching code
- Key sections: "Core Features", "Architecture Overview"
File: AGENTS.md
- What to learn: Repository guidelines, coding style, testing standards, Operator Workflow
- Why read this: This defines HOW we work - required for all contributors
- Key sections:
- "Operator Workflow" (lines 53-62) - CRITICAL: The plan -> implement -> document -> test loop
- "AI Agent Workflows" - Critical Reviewer methodology
- "Character Encoding: ASCII-only" - Use only ASCII characters in all files to prevent encoding issues across different systems and editors
- "Prompt & Changelog Formatting" - NEW REQUIREMENT: Every update must include a UTC timestamp line plus a dated changelog summary.
- Key sections:
File: docs/QUICKREF.md
- What to learn: Common commands, directory structure, configuration
- Why read this: Practical reference for day-to-day work
File: docs/MCP_SERVERS.md
- What to learn: Available MCP (Model Context Protocol) tools for file operations, diagnostics, and knowledge management
- Why read this: MCP tools provide powerful capabilities for automation, testing, and knowledge graph operations
- Key sections: "Available MCP Servers", "Agent Usage Scenarios"
- Quick overview:
- Filesystem: Read/write files, search, directory operations
- VideoChunking-Dev: Project diagnostics (pipeline health, test coverage, session management)
- Context7: Fetch up-to-date library documentation
- Memory: Knowledge graph operations (create entities, relations, search)
- IDE: VS Code integration (diagnostics, Jupyter code execution)
[DONE] Stage 1 Complete: You now understand WHAT this project does and HOW we work
File: docs/CRITICAL_REVIEW_WORKFLOW.md
- What to learn: How to implement features with documented reasoning and get critical review
- Why read this: REQUIRED workflow for all implementations
- Key sections: "Phase 1: Implementation", "Real-World Examples"
File: .claude/agents/critical-reviewer.md
- What to learn: The skeptical review methodology, checklists, philosophy
- Why read this: Understand the "assume issues exist" mindset
- Key sections: "Review Process", "Critical Review Checklist"
[DONE] Stage 2 Complete: You now understand our quality standards and review process
File: ROADMAP.md
- What to learn: All planned features (P0-P4), priorities, effort estimates
- Why read this: See the big picture and current priorities
- Key sections: "P0: Critical / Immediate", "Quick Reference Guide"
- Look for
IMPLEMENTATION_PLANS.mdand related files in the root directory for active plans. - Files in
docs/archive/are historical and should not be used for current work. - These contain detailed subtasks, code examples, templates
- Read the plan for any feature before implementing it
[DONE] Stage 3 Complete: You know what needs to be built
CRITICAL: This is how ALL work is done in this repository. The pipeline now supports checkpointing and resume, meaning you can restart a process from the last completed stage if it's interrupted.
Remember: Every status reply must start with a UTC date/time + note line and include a dated changelog section summarizing the work completed that day.
1. START FROM THE PLAN
|
v Read ROADMAP.md or implementation plans before coding
2. WORK IN SMALL STEPS
|
v Implement one subtask at a time
v Update plan immediately (checkboxes, notes)
v **Leverage Checkpoints**: If your work involves running the main pipeline, it will automatically save progress. If interrupted, you can restart, and it will resume from the last completed stage.
3. DOCUMENT REASONING
|
v Add "Implementation Notes & Reasoning" as you go
v Explain WHY, not just WHAT
4. VALIDATE CONTINUOUSLY
|
v Run tests after each change (pytest -q)
v Note gaps or failures
5. REPORT WITH CONTEXT
|
v Reference plan sections you advanced
v List tests executed
6. REQUEST CRITICAL REVIEW
|
v "Is there truly no issues with [feature]?"
v Address findings and iterate
7. MERGE AFTER APPROVAL
|
v Update documentation
+-> Loop back to step 1 for next task
- Read
CLAUDE.md(comprehensive AI assistant guide - START HERE) - Read
docs/PROJECT_SUMMARY.md - Read
AGENTS.md(focus on "Operator Workflow") - Read
docs/QUICKREF.md - Read
docs/MCP_SERVERS.md(understand available MCP tools)
- Read
docs/CRITICAL_REVIEW_WORKFLOW.md - Read
.claude/agents/critical-reviewer.md
- Read
ROADMAP.md - Choose a feature to work on (start with P0)
- Read the specific implementation plan for your chosen feature
- Understand the subtasks and success criteria
- Index:
docs/README.md- Complete documentation index - All docs:
docs/directory - MCP Tools:
docs/MCP_SERVERS.md- Available MCP servers and tools
- Main pipeline:
src/pipeline.py - UI:
app.py- Gradio web interface - CLI:
cli.py- Command-line interface - Core modules:
src/- chunker, transcriber, diarizer, etc. - Tests:
tests/
- Roadmap:
ROADMAP.md- All features (P0-P4) - Implementation Plans:
IMPLEMENTATION_PLANS*.md(if they exist)
- Repository guidelines:
AGENTS.md- Coding style, testing, Operator Workflow - Critical Review:
docs/CRITICAL_REVIEW_WORKFLOW.md - Review Agent:
.claude/agents/critical-reviewer.md
- MCP Servers:
docs/MCP_SERVERS.md- 50+ tools for file ops, diagnostics, knowledge graphs - Health Check: Use
mcp__videochunking-dev__check_pipeline_health() - Run Tests: Use
mcp__videochunking-dev__run_specific_test(test_path="...")
Audio Input (M4A/MP3/WAV)
|
v Audio Conversion (FFmpeg -> 16kHz WAV)
|
v Chunking (VAD-based smart chunking)
|
v Transcription (faster-whisper)
|
v Overlap Merging
|
v Speaker Diarization (PyAnnote)
|
v IC/OOC Classification (Ollama LLM)
|
v Output Generation
- Skeptical by default: Assume issues exist until proven otherwise
- Socratic questioning: Challenge assumptions
- Documented reasoning: Every decision needs a "why"
- Learning feedback loop: Quality compounds over time
All features MUST include:
- Implementation Notes & Reasoning - Design decisions, alternatives, trade-offs
- Code Review Findings - Issues identified, recommendations, merge verdict
- Tests - Unit tests for new code
- Documentation - Update relevant docs
- P0: Critical/Immediate (bugs, crashes, refactoring blockers)
- P1: High Impact (features that unlock major value)
- P2: Important Enhancements
- P3: Future Enhancements
- P4: Infrastructure & Quality
- Coding without reading the plan -> You'll miss requirements
- Leaving documentation until the end -> Context is lost
- Not keeping the plan in sync -> Plan becomes stale
- Skipping tests -> Bugs slip through
- Not requesting critical review -> Issues ship to production
- Start from the plan -> Read before writing code
- Document as you go -> Update plan after each subtask
- Keep plan synchronized -> Plan is single source of truth
- Write tests continuously -> Test as you go
- Request skeptical review -> "Is there truly no issues?"
You're successfully onboarded when you can:
- Explain what this project does
- Navigate the codebase and find relevant modules
- Follow the Operator Workflow loop
- Read an implementation plan and understand subtasks
- Implement a feature with proper documentation
- Request and respond to critical review
Essential Context (5 min):
1. CLAUDE.md (AI Assistant comprehensive guide - RECOMMENDED START)
2. docs/PROJECT_SUMMARY.md
3. AGENTS.md (focus: Operator Workflow, ASCII-only)
4. docs/QUICKREF.md
5. docs/MCP_SERVERS.md (MCP tools available)
Development Standards (15 min):
6. docs/CRITICAL_REVIEW_WORKFLOW.md
7. .claude/agents/critical-reviewer.md
What to Build (20 min):
8. ROADMAP.md
9. IMPLEMENTATION_PLANS*.md (if they exist)
"Quality emerges from dialogue, not perfection on first try. Every implementation deserves skeptical analysis, and every decision deserves a documented 'why'."
Core principles:
- Plans are living documents - Keep them in sync
- Reasoning is required - Document the "why"
- Skepticism is professionalism - "Revisions requested" is normal
- Feedback loops create quality - Review -> Document -> Learn -> Improve
Welcome to the team! [ROCKET]
Next step: Choose your first task from ROADMAP.md and read its implementation plan.