Skip to content

Commit 1bfc1e9

Browse files
committed
Docs.
Signed-off-by: Jose Alekhinne <alekhinejose@gmail.com>
1 parent d650c25 commit 1bfc1e9

3 files changed

Lines changed: 7 additions & 7 deletions

File tree

docs/home/prompting-guide.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -654,14 +654,14 @@ powerful optimizers find solutions that technically satisfy the
654654
objective but are practically useless.
655655

656656
**CLI commands as prompts** ("*Run `ctx status`*") interleave
657-
*reasoning with acting* — the model thinks, acts on external tools,
657+
*reasoning with acting*: The model thinks, acts on external tools,
658658
observes results, then thinks again. Grounding reasoning in real
659659
tool output reduces hallucination because the model can't ignore
660660
evidence it just retrieved.
661661
<br>Yao et al., [ReAct: Synergizing Reasoning and Acting in Language Models](https://arxiv.org/abs/2210.03629) (2022).
662662

663663
**Task decomposition** ("*Prompts by Task Type*") applies
664-
*least-to-most prompting* — breaking a complex problem into
664+
*least-to-most prompting*: Breaking a complex problem into
665665
subproblems and solving them sequentially, each building on the last.
666666
This is the research version of "plan, then implement one slice."
667667
<br>Zhou et al., [Least-to-Most Prompting Enables Complex Reasoning in Large Language Models](https://arxiv.org/abs/2205.10625) (2022).
@@ -674,7 +674,7 @@ understanding the problem.
674674
<br>Wang et al., [Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models](https://arxiv.org/abs/2305.04091) (2023).
675675

676676
**Session reflection** ("*What did we learn?*", `/ctx-reflect`) is
677-
a form of *verbal reinforcement learning* — improving future
677+
a form of *verbal reinforcement learning*: Improving future
678678
performance by persisting linguistic feedback as memory rather than
679679
updating weights. This is exactly what `LEARNINGS.md` and
680680
`DECISIONS.md` provide: a durable feedback signal across sessions.

site/home/prompting-guide/index.html

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4570,13 +4570,13 @@ <h2 id="why-do-these-approaches-work">Why Do These Approaches Work?<a class="hea
45704570
powerful optimizers find solutions that technically satisfy the
45714571
objective but are practically useless.</p>
45724572
<p><strong>CLI commands as prompts</strong> ("<em>Run <code>ctx status</code></em>") interleave
4573-
<em>reasoning with acting</em> — the model thinks, acts on external tools,
4573+
<em>reasoning with acting</em>: The model thinks, acts on external tools,
45744574
observes results, then thinks again. Grounding reasoning in real
45754575
tool output reduces hallucination because the model can't ignore
45764576
evidence it just retrieved.
45774577
<br>Yao et al., <a href="https://arxiv.org/abs/2210.03629">ReAct: Synergizing Reasoning and Acting in Language Models</a> (2022).</p>
45784578
<p><strong>Task decomposition</strong> ("<em>Prompts by Task Type</em>") applies
4579-
<em>least-to-most prompting</em> — breaking a complex problem into
4579+
<em>least-to-most prompting</em>: Breaking a complex problem into
45804580
subproblems and solving them sequentially, each building on the last.
45814581
This is the research version of "plan, then implement one slice."
45824582
<br>Zhou et al., <a href="https://arxiv.org/abs/2205.10625">Least-to-Most Prompting Enables Complex Reasoning in Large Language Models</a> (2022).</p>
@@ -4587,7 +4587,7 @@ <h2 id="why-do-these-approaches-work">Why Do These Approaches Work?<a class="hea
45874587
understanding the problem.
45884588
<br>Wang et al., <a href="https://arxiv.org/abs/2305.04091">Plan-and-Solve Prompting: Improving Zero-Shot Chain-of-Thought Reasoning by Large Language Models</a> (2023).</p>
45894589
<p><strong>Session reflection</strong> ("<em>What did we learn?</em>", <code>/ctx-reflect</code>) is
4590-
a form of <em>verbal reinforcement learning</em> — improving future
4590+
a form of <em>verbal reinforcement learning</em>: Improving future
45914591
performance by persisting linguistic feedback as memory rather than
45924592
updating weights. This is exactly what <code>LEARNINGS.md</code> and
45934593
<code>DECISIONS.md</code> provide: a durable feedback signal across sessions.

site/search.json

Lines changed: 1 addition & 1 deletion
Large diffs are not rendered by default.

0 commit comments

Comments
 (0)