He Open-Sourced His Claude Folder. 68K Stars on GitHub.
How a folder of markdown files beat BMAD, GSD, and most of the agent-framework ecosystem in 90 days.
May 10th 2026An ex-Apple engineer benchmarks ref.tools and Exa AI against Cursor on a live Tailwind v4 refactor — and Claude Code wins at 2,800 tokens vs 98,000.
Documentation used to eat 100,000 tokens before a single line of code was written. Ray Fernando — ex-Apple engineer, decade-plus shipping production software — found the fix in two MCP servers that pull only the context the agent needs, precisely when it needs it. The proof is live on screen: a full Tailwind v4 design-token audit across a real codebase clocks in at 2,800 tokens — 1.4% of what Cursor burned on the same task.
stated at 00:02“I'm gonna show you how to get two of my favorite MCPs set up.”delivered at 16:50

Ray introduces the concept of context rot — LLMs getting dumb as the window fills with irrelevant documentation. Establishes why targeted doc-fetching MCPs beat brute-force token dumps.

One-command install via `claude mcp add`, API key walkthrough, and security warning about committing keys to public repos.

Same pattern as ref.tools. API key generation and paste into terminal. Both MCPs now available globally across any project.

Haiku 4.5 runs a full codebase audit using both MCPs. Ray watches Claude Code research documentation, build a phased implementation plan, and hit /context — revealing only 2,800 tokens used.

Cursor's plan mode finished the same task using 98,000 tokens. Side-by-side makes the 35x gap visceral. Ray notes Cursor asks clarifying questions; Claude Code doesn't.

Codex uses config.toml instead of JSON. Ray edits the file in Cursor, pastes the API key, verifies with `codex mcp list`.

Copy the MCP JSON block from Cursor's tools panel directly into .factory/mcp.json. Same pattern, different file path.

Always call MCPs explicitly in your prompt. Use plan mode (shift-tab) before writing code. Closes with a pitch for his 1337 coaching intensive and a forward-looking take on agentic AI.
The degradation of LLM output quality as the context window fills with broad, untargeted documentation — the agent gets dumb before it writes a line of code.
Pull only the context needed for the specific sub-task, not everything that might be relevant. Combine ref.tools (documentation precision) with Exa (coding-task search quality) and always call them by name in the prompt.
“I used to use up almost 100k tokens just feeding in tons of documentation.”
“LLMs actually operate best when they have just the right information for just the right specific task.”
“Our MCP servers and tool calls only use a total of 2,800 tokens — which is only 1.4% of the context window.”
“It's almost like as if I hired several developers to read the documentation and implement the code for us.”
“Make sure you use plan mode so that it gathers all that specific context before you start writing code.”
“I do have a couple more spots that are open for my one three three seven intensive.”
Soft close after delivering all value. Thirty-minute sessions over five days. Credentialed with Apple engineering background. Low pressure but well-timed after a high-value tutorial.
The 35x token gap isn't about Claude Code being smarter than Cursor — it's about giving the agent a scalpel instead of a firehose.
Your AI assistant is probably burning through its memory reading docs it doesn't need — and getting dumber because of it.
00:00
00:15
00:31
00:43
00:57
01:17
01:31
01:40
02:00
02:14
02:28
02:42
02:56
03:08
03:28
03:40
03:50
04:07
04:21
04:35
04:49
05:03
05:18
05:38
05:46
05:54
06:13
06:28
06:46
06:57
07:17
07:25
07:39
07:53
08:05
08:28
08:30
08:47
09:04
09:18
09:32
09:46
10:00
10:14
10:29
10:43
10:57
11:11
11:25
11:39
11:53
12:08
12:22
12:36
12:51
13:04
13:18
13:32
13:47
14:01
14:15
14:29
14:43
14:58
15:07
15:26
15:40
15:54
16:08
16:22
16:36
16:49
17:04
17:14
17:37
17:47
17:57
18:13
18:25
18:37How a folder of markdown files beat BMAD, GSD, and most of the agent-framework ecosystem in 90 days.
May 10th 2026A 19-minute live demo of Auto Claude, the free open-source Kanban orchestrator that runs parallel Claude Code agents in git worktrees while you sleep.
December 17th 2025A 13-minute verdict: CLI inside VS Code beats Cursor, the desktop app, and the extension — and a live app build is the receipt.
December 30th 2025GosuCoder rapid-fires 14 Claude Code tips — a swipe-file of prompt templates and CLI shortcuts disguised as a YouTube listicle.
June 14th 2025318 commits in May. One outline canvas. One creator's honest pricing breakdown.
June 5th 2025A 12-minute tutorial covering slash commands, sub-agents, hooks, think modes, and CLAUDE.md rules — each delivered as a single copy-paste prompt.
July 26th 2025