The bait, then the rug-pull.
Keith opens with the line every over-tabbed developer already knows is true: one terminal, one task, one spinning cursor while three other things wait. In twelve seconds he names the real bottleneck — attention, not compute — and introduces OpenCode as the fix. What follows is the most thorough walkthrough of the tool currently on YouTube: install to mobile SSH in under half an hour.
What the video promised.
stated at 00:22“This is a step by step tutorial on how I actually use OpenCode, a real workflow under real constraints to ship faster without a team.”delivered at 28:00
Where the time goes.

01 · Hook + Keith intro
Names the bottleneck (attention), introduces OpenCode and Keith background.

02 · Old workflow vs OpenCode
Whiteboard diagram: before = chaos of 3 terminals; after = OpenCode task list with usage remaining.

03 · What is OpenCode + agenda
Task orchestration for AI coding. Agenda: install, connect providers, build/plan modes, MCPs, skills, live demo.

04 · Install: Desktop, Extension, CLI, Web
Desktop app, Cursor/Windsurf extension, curl one-liner for terminal, opencode --web for remote browser access.

05 · Settings and appearance
Dark mode, Dracula/Tokyo Night theme, sound effects on task complete, shortcuts.

06 · Build vs Plan mode + model strategy
Plan with Gemini 3 Pro (high thinking), execute with GPT-5.2 Codex; token budget panel; split features into separate sessions.

07 · Connecting AI providers
Desktop: /model > Connect Provider. Terminal: /select. GLM 4.7 and MiniMax rank above GPT-5.2 Flash on benchmarks.

08 · Live demo: parallel sessions on health app
Bug fix via screenshot + Gemini planning + UI/UX improvement — three tasks simultaneously. Gamification (streaks) built live.

09 · Advanced: agents.md + /review
/init reads codebase and writes agents.md for token efficiency. /review catches bugs inline.

10 · Advanced: MCP servers (Context7)
Context7 MCP via global opencode.json. Add rule to agents.md so model calls Context7 automatically.

11 · Terminal CLI: split panes + model picker
Control-T = token cap, Tab = plan/build, Control-P = model. Terminal exposes more models than desktop.

12 · Custom skills and commands
Skills = reusable instruction sets in skills.md. Commands = slash shortcuts. Added git commit+push live.

13 · Mobile: OpenCode on iPhone via Termius
SSH remote login on Mac, Termius app, local WiFi. Cloudflare tunnel for remote access is a future video.

14 · Outro + community CTA
Like/subscribe, free AI community at rumjahn.substack.com, vibe coding course.
Visual structure at a glance.
Named ideas worth stealing.
Sequential vs Parallel Coding
Old way = one task at a time until credits run out. New way = fleet of sessions in parallel; developer intervenes only when needed.
Model Budget Strategy ($20 Plans)
- Gemini 3 Pro Preview for planning (high thinking effort)
- GPT-5.2 Codex for complex execution
- GPT-5.2 for simple bugs
- Free models (GLM 4.7, MiniMax) as fallback
Match model capability to task complexity to stay inside $20/month budgets.
agents.md initialization
Run /init to have OpenCode read codebase and generate agents.md with code style, naming, error handling. Future sessions need fewer tokens.
Lines you could clip.
“My job now isn't to watch AI work. My job now is to intervene when it's only needed.”
“I'd trade money for time any day of the week.”
“Instead of making it one long continuous chat one after another, which consumes a lot of tokens, I split them up into new sessions each one so that it saves me money.”
How they spent the runtime.
Things they pointed at.
How they asked for the click.
“Join my free AI community. You can join my community for free inside the description.”
Soft close, no hard sell. Vibe coding course as secondary offer. Non-pushy given the dense tutorial content.
Word for word.
Steal the parallel-agent frame.
The bottleneck isn't AI speed — it's your attention. The moment you frame your tool as a supervision cockpit instead of a chat window, everything changes.
- Lead with the bottleneck not the feature: 'your attention is the limit' lands harder than 'run multiple sessions.'
- The agents.md / CLAUDE.md pattern is the same insight — front-load context once, save tokens on every future task.
- Token budget awareness as a workflow decision is a strong content angle for the $20 stack audience.
- The Sequential vs Parallel whiteboard diagram is a clean visual for a short — one frame, one thesis, steal the format.
- His Claude API warning is worth a JoeFlow response video — clarify what JoeFlow actually does differently.
- The Termius mobile SSH demo is an easy hook for a short: 'I coded from my iPhone on the bus.'
How to actually use AI to ship more with less time.
Stop watching AI think. Run three tasks at once and only look up when something needs you.
- Start every new project with /init — it reads your codebase and writes a context file so future sessions need fewer tokens.
- Use Gemini for planning (better at thinking through logic), then switch to GPT for the actual code.
- Split features into separate sessions instead of one long chat — saves context and keeps tasks clean.
- The /review command is a free code audit — run it after adding a batch of features to catch bugs before they ship.
- Add Context7 MCP once and your AI reads library docs automatically instead of hallucinating APIs.



































































