Skip to content

feat: AI-native framework integration (v1.3.0)#1

Merged
savvides merged 6 commits intomainfrom
feat/ai-native-framework
Apr 1, 2026
Merged

feat: AI-native framework integration (v1.3.0)#1
savvides merged 6 commits intomainfrom
feat/ai-native-framework

Conversation

@savvides
Copy link
Copy Markdown
Owner

@savvides savvides commented Apr 1, 2026

Summary

Every skill is now AI-native aware. When a founder's product involves AI, skills detect whether the AI is load-bearing (AI-native) or decorative (bolted-on) and adapt guidance accordingly.

Foundation:

  • New data/ai-native-framework.md reference data: 4 AI-native criteria, 5 bolted-on indicators, removal test, Karpathy hierarchy, architecture patterns, pricing models

Heavy integration (new scoring dimensions):

  • /product-review — 6th scoring dimension: AI Architecture (criteria-based evaluation)
  • /idea-validation — AI Architecture Fit signal with higher-ed cross-referencing

Medium integration (adapted guidance):

  • /go-to-market — Usage-based pricing playbook, model improvement positioning
  • /pitch-review — "Improves with models" investor narrative, AI-native VC targeting
  • /sales-strategy — AI-native objection handling, demo flow, procurement narrative
  • /fundraising-guide — AI-native vs traditional VC targeting

Light integration (specific additions):

  • /edtech-landscape — AI-native vs bolted-on competitive mapping
  • /evidence-check — Behavior change evidence dimension
  • /accessibility-check — AI bias, transparency, explainability concerns
  • /pilot-design — AI-specific pilot metrics (accuracy, hallucination, trust)

The framework is diagnostic, not prescriptive. Bolted-on AI can be a valid strategy.

Pre-Landing Review

No issues found. All changes are markdown files (SKILL.md, reference data, documentation).

Adversarial Review

8 issues auto-fixed (wording consistency across skills, trigger logic, unsourced claim, missing next-skill recommendation). 3 pre-existing issues noted but not blocking.

Plan Completion

16/16 plan items DONE. 0 PARTIAL, 0 NOT DONE.

Scope Drift

Scope Check: CLEAN

TODOS

  • Created TODOS.md with P3 item: AI-native framework consistency check (CI step)

Test plan

  • Run /product-review with AI-native product description — verify AI Architecture dimension appears
  • Run /idea-validation with AI idea — verify AI Architecture Fit signal
  • Run /go-to-market with AI product — verify usage-based pricing guidance
  • Run any skill with non-AI product — verify AI sections are skipped

🤖 Generated with Claude Code

savvides and others added 6 commits April 1, 2026 15:55
New data/ai-native-framework.md with 4 AI-native criteria, 5 bolted-on
indicators, the removal test, Karpathy hierarchy, architecture patterns,
and AI-native pricing models. CLAUDE.md updated to document the new file.
product-review gets a 6th scoring dimension (AI Architecture) with the
removal test and 4-criteria evaluation. idea-validation gets AI Architecture
Fit as a validation signal with higher-ed cross-referencing.
go-to-market adds usage-based pricing playbook. pitch-review coaches
the "improves with models" narrative. sales-strategy adds AI-native
objection handling. fundraising-guide targets different VC pools.
edtech-landscape maps AI-native competitors. evidence-check adds behavior
change dimension. accessibility-check flags AI bias and transparency.
pilot-design adds AI-specific metrics.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Copy link
Copy Markdown

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request integrates a comprehensive AI-native framework into the EdTech Founder Stack, upgrading the system to version 1.3.0. The changes include a new reference data file defining AI-native vs. bolted-on criteria and the systematic update of all ten skills to include AI posture detection, specialized metrics, and tailored strategic guidance for AI products. Review feedback correctly identifies several minor inconsistencies where specific mapping instructions or scoring rubrics were omitted in certain skill files compared to others, ensuring the framework operates uniformly across the repository.

- AI is a minor or planned feature — optional, supplementary, or not yet built
- No AI component

If AI-native or borderline: read `data/ai-native-framework.md` and add "behavior change" as an evidence dimension in Phase 2. AI-native products need to demonstrate they create behavior change (users work fundamentally differently), not just engagement or satisfaction.
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The mapping instructions for the AI posture detection question are missing in this skill. All other skills that implement this detection block include explicit mapping for the LLM to classify the product as 'AI-native' or 'Borderline'. Adding this ensures consistent behavior across the framework.

Suggested change
If AI-native or borderline: read `data/ai-native-framework.md` and add "behavior change" as an evidence dimension in Phase 2. AI-native products need to demonstrate they create behavior change (users work fundamentally differently), not just engagement or satisfaction.
Map: "AI IS the product" = AI-native. "Significant feature" = Borderline. "Minor/planned" or "No AI" = Skip AI-specific evidence guidance.
If AI-native or borderline: read `data/ai-native-framework.md` and add "behavior change" as an evidence dimension in Phase 2. AI-native products need to demonstrate they create behavior change (users work fundamentally differently), not just engagement or satisfaction.

Comment on lines +154 to +159
Read `data/ai-native-framework.md`. Evaluate:

- **Removal test:** Remove the AI from this idea. Does it still solve the problem? If yes, the AI is supplementary, not essential.
- **Model improvement trajectory:** Will this idea get better as base models improve? Or is the AI a static feature that won't evolve?
- **Behavior change potential:** Would users of this product work fundamentally differently, or just slightly faster?
- **Higher ed cross-reference:** If the founder targets higher ed, check `data/higher-ed-jobs-atlas.md`. Some jobs are naturally AI-native (adaptive tutoring, intelligent scheduling, personalized content generation). Others don't need AI (financial aid navigation, transfer credit mapping). Flag if the founder is forcing AI into a job that doesn't need it.
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

This dimension lacks the scoring rubric defined in the /product-review skill. Since the scorecard on line 186 expects a score for 'AI Architecture Fit', providing a consistent rubric (using even numbers 2-10) ensures the LLM evaluates the idea accurately and maintains consistency across the Founder Stack.

Suggested change
Read `data/ai-native-framework.md`. Evaluate:
- **Removal test:** Remove the AI from this idea. Does it still solve the problem? If yes, the AI is supplementary, not essential.
- **Model improvement trajectory:** Will this idea get better as base models improve? Or is the AI a static feature that won't evolve?
- **Behavior change potential:** Would users of this product work fundamentally differently, or just slightly faster?
- **Higher ed cross-reference:** If the founder targets higher ed, check `data/higher-ed-jobs-atlas.md`. Some jobs are naturally AI-native (adaptive tutoring, intelligent scheduling, personalized content generation). Others don't need AI (financial aid navigation, transfer credit mapping). Flag if the founder is forcing AI into a job that doesn't need it.
Read `data/ai-native-framework.md`. Evaluate the idea against the 4 AI-native criteria (Token scaling, Model improvement, Removal test, Behavior change).
Score on the AI-native criteria (use even numbers to distinguish from the 1-10 scale of other dimensions):
- **10:** All 4 criteria met. Architecture is model-agnostic. Clear token economics. Memory/personalization across sessions.
- **8:** 3 of 4 criteria met. Strong AI integration but missing memory, behavior change, or model-agnostic design.
- **6:** 2 of 4 criteria met. AI is meaningful but not load-bearing. The product is in a borderline zone.
- **4:** 1 criterion met. AI is a significant feature but the product works without it. Bolted-on with potential.
- **2:** 0 criteria met but product claims to be "AI-powered." Sparkle icons and a chat pane.
**Higher ed cross-reference:** If the founder targets higher ed, check `data/higher-ed-jobs-atlas.md`. Some jobs are naturally AI-native (adaptive tutoring, intelligent scheduling, personalized content generation). Others don't need AI (financial aid navigation, transfer credit mapping). Flag if the founder is forcing AI into a job that doesn't need it.

- AI is a minor or planned feature — optional, supplementary, or not yet built
- No AI component

If AI-native or borderline: read `data/ai-native-framework.md` and add AI-specific pilot metrics to Phase 2.
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The mapping instructions for the AI posture detection question are missing here. Including the explicit mapping for 'AI-native' and 'Borderline' ensures the LLM correctly triggers the AI-specific pilot metrics in Phase 2, maintaining consistency with the other skills in the repository.

Suggested change
If AI-native or borderline: read `data/ai-native-framework.md` and add AI-specific pilot metrics to Phase 2.
Map: "AI IS the product" = AI-native. "Significant feature" = Borderline. "Minor/planned" or "No AI" = Skip AI-specific pilot metrics.
If AI-native or borderline: read `data/ai-native-framework.md` and add AI-specific pilot metrics to Phase 2.

**If Evidence Readiness scored < 7:**
"You need stronger evidence before institutional buyers will take you seriously. Run `/evidence-check` to assess your evidence tier and build a plan."

**If AI Architecture scored < 5 (and AI is involved):**
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

There is a minor inconsistency between the warning threshold defined in Phase 2 (line 162, score < 6) and the recommendation trigger here in Phase 5 (score < 5). While logically equivalent given the even-number scoring system (2, 4, 6, 8, 10), using < 6 here would be more consistent and clearer for maintenance.

Suggested change
**If AI Architecture scored < 5 (and AI is involved):**
**If AI Architecture scored < 6 (and AI is involved):**

@savvides savvides merged commit 8ca4724 into main Apr 1, 2026
1 check passed
@savvides savvides deleted the feat/ai-native-framework branch April 1, 2026 22:59
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant