Skip to content

Commit bda42b3

Browse files
committed
Errors--
1 parent f67cc19 commit bda42b3

File tree

1 file changed

+3
-8
lines changed

1 file changed

+3
-8
lines changed

website/docs/methodology/lesson-5-grounding.md

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,8 @@ sidebar_label: 'Lesson 5: Grounding'
44
title: 'Grounding: Anchoring Agents in Reality'
55
---
66

7-
{/_ TODO: Create these visual components
87
import UShapeAttentionCurve from '@site/src/components/VisualElements/UShapeAttentionCurve';
98
import GroundingComparison from '@site/src/components/VisualElements/GroundingComparison';
10-
_/}
119

1210
LLMs have a fundamental limitation: they only "know" what's in their training data (frozen at a point in time) and what's in their current context window (~200K/400K tokens for Claude Sonnet 4.5 / GPT-5 respectively). Everything else is educated guessing.
1311

@@ -19,11 +17,7 @@ This lesson covers the engineering techniques that turn agents from creative fic
1917

2018
**Scenario:** You're debugging an authentication bug in your API.
2119

22-
{/_ TODO: Add GroundingComparison visual component here showing side-by-side comparison _/}
23-
**Without grounding:** Generic advice like "Check your JWT validation logic"
24-
**With grounding:** Specific fix referencing your actual code: "In src/auth/jwt.ts:45, the validateJWT() function isn't checking token expiration"
25-
26-
The difference is clear: ungrounded responses are generic and potentially wrong. Grounded responses reference your actual code and current best practices.
20+
<GroundingComparison />
2721

2822
## RAG: Retrieval-Augmented Generation
2923

@@ -84,7 +78,8 @@ Claude Sonnet 4.5 has a 200K token context window. In practice, you'll get relia
8478

8579
### The Problem
8680

87-
{/_ TODO: Add UShapeAttentionCurve visual component here showing attention dropoff _/}
81+
<UShapeAttentionCurve />
82+
8883
**The U-shaped attention curve:** Information at the **beginning** and **end** of your context gets strong attention. Information in the **middle** gets skimmed or missed entirely. It's not a bug—it's how transformer attention mechanisms work under realistic constraints.
8984

9085
When you retrieve documentation and code chunks directly in your orchestrator context, they rapidly fill the window with search results, pushing critical constraints into that ignored middle. A few semantic searches return 10+ code chunks each (30K tokens), web docs add more (15K tokens)—your context fills with search mechanics before research completes, and the orchestrator forgets initial constraints.

0 commit comments

Comments
 (0)