Loading course…
Loading course…
Created by Petter Smit
You will map and implement the core reasoning engine patterns used in modern AI agents, from ReAct control loops to planning, branching search, and evaluator-driven self-correction. By the end, you can choose the right reasoning pattern for a task, wire it into an agent loop with termination gates, and manage long-horizon memory with summarization and scaffolding tradeoffs.
7 modules • Each builds on the previous one
Define the reasoning engine as a control loop over state, tools, and memory, and map where prompting patterns fit (policy, planner, critic, summarizer) versus what belongs in the orchestrator.
Implement ReAct as a tight Thought→Action→Observation loop with explicit state updates, error handling, and termination criteria; focus on correctly feeding observations back into the next step.
Use CoT as a reasoning scaffold while managing exposure: structured intermediate steps, hidden reasoning patterns, and testable decomposition without leaking sensitive traces.
Separate planning from execution: generate an explicit roadmap (subgoals, tool plan, checks) before solving, then execute with verification and replanning triggers.
Move from linear CoT to explicit search: generate multiple candidate thoughts, score them with evaluators/lookahead, then prune/expand under a token/latency budget.
Add evaluator passes to detect mistakes and constraint violations: reflection for strategy correction, criticism with explicit rubrics (functional, security, style, policy) and multi-criteria scoring.
Use hierarchical/rolling summaries to keep long-horizon agents coherent: compress history into stable abstractions, preserve constraints/rubrics, and prevent goal drift under context limits.
Begin your learning journey
In-video quizzes and scaffolded content to maximize retention.