You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
i18n: complete EN/KO parity for Beyond RAG section
Add the 17 rag.* keys (rag.label, rag.title, rag.intro, rag.viz.*,
rag.chunk.*, rag.doc.*, rag.complementary.*, rag.card1-3.*, rag.pipeline.*)
to both EN and KO dictionaries. The HTML markup referenced these keys but
they were missing from the i18n tables, so the Beyond RAG section fell
back to the literal key strings on language toggle.
EN dict: 162 keys, KO dict: 162 keys, HTML data-i18n keys: 162 — verified
matching by diff.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Copy file name to clipboardExpand all lines: site/index.html
+36-2Lines changed: 36 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -896,7 +896,24 @@ <h2 style="margin-bottom:1rem" data-i18n="cta.title">Try It Yourself</h2>
896
896
"glossary.gguf.term": "GGUF",
897
897
"glossary.gguf.def": "The standard file format for quantized LLM model weights, created by the llama.cpp project. quant.cpp loads GGUF models directly.",
898
898
"cta.title": "Try It Yourself",
899
-
"cta.desc": "Three lines of Python. No GPU, no API key, no setup."
899
+
"cta.desc": "Three lines of Python. No GPU, no API key, no setup.",
900
+
"rag.label": "Movement",
901
+
"rag.title": "Beyond RAG",
902
+
"rag.intro": "Traditional RAG splits documents into 512-token chunks, embeds them in a vector database, and retrieves fragments. This was a reasonable engineering compromise when LLMs had 2K context windows. <strong>Now they have 128K. The compromise should have started disappearing.</strong>",
903
+
"rag.viz.title": "Chunk-Level RAG vs Document-Level RAG",
904
+
"rag.chunk.title": "Chunk-Level RAG",
905
+
"rag.chunk.result": "✗ Cross-page info lost",
906
+
"rag.doc.title": "Document-Level RAG",
907
+
"rag.doc.result": "✓ Full document understanding",
908
+
"rag.complementary.title": "Complementary, Not Competitive",
909
+
"rag.complementary.desc": "RAG decides <strong>which documents</strong> to look at. Long-context decides <strong>how deeply</strong> to understand them. Each does what it's best at.",
"rag.card1.d": "Chunk boundaries lose cross-page relationships. Multi-hop reasoning fails. Long-context keeps the full document — no information loss.",
"cta.desc": "Python 3\uC904. GPU\uB3C4, API \uD0A4\uB3C4, \uC124\uCE58\uB3C4 \uD544\uC694 \uC5C6\uC2B5\uB2C8\uB2E4."
1063
+
"cta.desc": "Python 3\uC904. GPU\uB3C4, API \uD0A4\uB3C4, \uC124\uCE58\uB3C4 \uD544\uC694 \uC5C6\uC2B5\uB2C8\uB2E4.",
1064
+
"rag.label": "운동",
1065
+
"rag.title": "Beyond RAG",
1066
+
"rag.intro": "전통적인 RAG는 문서를 512토큰 청크로 나누고, 벡터 DB에 임베딩하고, 조각을 검색합니다. 이것은 LLM이 2K 컨텍스트만 가졌을 때 합리적인 엔지니어링 타협이었습니다. <strong>지금은 128K입니다. 그 타협은 사라지기 시작했어야 합니다.</strong>",
1067
+
"rag.viz.title": "Chunk-Level RAG vs Document-Level RAG",
1068
+
"rag.chunk.title": "Chunk-Level RAG",
1069
+
"rag.chunk.result": "✗ 페이지 간 정보 손실",
1070
+
"rag.doc.title": "Document-Level RAG",
1071
+
"rag.doc.result": "✓ 전체 문서 이해",
1072
+
"rag.complementary.title": "경쟁이 아닌 상호 보완",
1073
+
"rag.complementary.desc": "RAG는 <strong>어떤 문서를</strong> 볼지 결정합니다. Long-context는 <strong>얼마나 깊이</strong> 이해할지 결정합니다. 각자 잘하는 것을 합니다.",
1074
+
"rag.card1.t": "RAG의 약점 → Long-Context가 해결",
1075
+
"rag.card1.d": "청크 경계에서 페이지 간 관계가 사라집니다. Multi-hop 추론은 실패합니다. Long-context는 전체 문서를 유지합니다 — 정보 손실 없음.",
1076
+
"rag.card2.t": "Long-Context의 약점 → RAG가 해결",
1077
+
"rag.card2.d": "100K 문서를 한 번에 컨텍스트에 넣을 수 없습니다. Prefill이 느립니다. RAG는 검색을 2-3개 관련 문서로 좁혀줍니다.",
1078
+
"rag.card3.t": "한 번 읽고, 영원히 질문",
1079
+
"rag.card3.d": "문서를 .kv 파일로 사전 처리 (GPU, 1회). 어떤 노트북에서든 즉시 로드 (0.5초). 오프라인, 무제한, 프라이빗 질문.",
0 commit comments