You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/home/prompting-guide.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -654,14 +654,14 @@ powerful optimizers find solutions that technically satisfy the
654
654
objective but are practically useless.
655
655
656
656
**CLI commands as prompts** ("*Run `ctx status`*") interleave
657
-
*reasoning with acting* — the model thinks, acts on external tools,
657
+
*reasoning with acting*: The model thinks, acts on external tools,
658
658
observes results, then thinks again. Grounding reasoning in real
659
659
tool output reduces hallucination because the model can't ignore
660
660
evidence it just retrieved.
661
661
<br>Yao et al., [ReAct: Synergizing Reasoning and Acting in Language Models](https://arxiv.org/abs/2210.03629) (2022).
662
662
663
663
**Task decomposition** ("*Prompts by Task Type*") applies
664
-
*least-to-most prompting* — breaking a complex problem into
664
+
*least-to-most prompting*: Breaking a complex problem into
665
665
subproblems and solving them sequentially, each building on the last.
666
666
This is the research version of "plan, then implement one slice."
667
667
<br>Zhou et al., [Least-to-Most Prompting Enables Complex Reasoning in Large Language Models](https://arxiv.org/abs/2205.10625) (2022).
@@ -674,7 +674,7 @@ understanding the problem.
674
674
<br>Wang et al., [Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models](https://arxiv.org/abs/2305.04091) (2023).
675
675
676
676
**Session reflection** ("*What did we learn?*", `/ctx-reflect`) is
677
-
a form of *verbal reinforcement learning* — improving future
677
+
a form of *verbal reinforcement learning*: Improving future
678
678
performance by persisting linguistic feedback as memory rather than
679
679
updating weights. This is exactly what `LEARNINGS.md` and
680
680
`DECISIONS.md` provide: a durable feedback signal across sessions.
Copy file name to clipboardExpand all lines: site/home/prompting-guide/index.html
+3-3Lines changed: 3 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -4570,13 +4570,13 @@ <h2 id="why-do-these-approaches-work">Why Do These Approaches Work?<a class="hea
4570
4570
powerful optimizers find solutions that technically satisfy the
4571
4571
objective but are practically useless.</p>
4572
4572
<p><strong>CLI commands as prompts</strong> ("<em>Run <code>ctx status</code></em>") interleave
4573
-
<em>reasoning with acting</em> — the model thinks, acts on external tools,
4573
+
<em>reasoning with acting</em>: The model thinks, acts on external tools,
4574
4574
observes results, then thinks again. Grounding reasoning in real
4575
4575
tool output reduces hallucination because the model can't ignore
4576
4576
evidence it just retrieved.
4577
4577
<br>Yao et al., <ahref="https://arxiv.org/abs/2210.03629">ReAct: Synergizing Reasoning and Acting in Language Models</a> (2022).</p>
4578
4578
<p><strong>Task decomposition</strong> ("<em>Prompts by Task Type</em>") applies
4579
-
<em>least-to-most prompting</em> — breaking a complex problem into
4579
+
<em>least-to-most prompting</em>: Breaking a complex problem into
4580
4580
subproblems and solving them sequentially, each building on the last.
4581
4581
This is the research version of "plan, then implement one slice."
4582
4582
<br>Zhou et al., <ahref="https://arxiv.org/abs/2205.10625">Least-to-Most Prompting Enables Complex Reasoning in Large Language Models</a> (2022).</p>
@@ -4587,7 +4587,7 @@ <h2 id="why-do-these-approaches-work">Why Do These Approaches Work?<a class="hea
4587
4587
understanding the problem.
4588
4588
<br>Wang et al., <ahref="https://arxiv.org/abs/2305.04091">Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models</a> (2023).</p>
4589
4589
<p><strong>Session reflection</strong> ("<em>What did we learn?</em>", <code>/ctx-reflect</code>) is
4590
-
a form of <em>verbal reinforcement learning</em> — improving future
4590
+
a form of <em>verbal reinforcement learning</em>: Improving future
4591
4591
performance by persisting linguistic feedback as memory rather than
4592
4592
updating weights. This is exactly what <code>LEARNINGS.md</code> and
4593
4593
<code>DECISIONS.md</code> provide: a durable feedback signal across sessions.
0 commit comments