Claude Certified Architect – Foundations: Visual Exam Cheat Sheet

Ciprian Ciprian · · 13 min read

Claude Certified Architect – Foundations
Visual Exam Cheat Sheet

Built from the exam guide. This is optimized for pattern recognition, not prose reading. The exam is mostly testing whether you choose the most structured, explicit, scoped, deterministic answer in realistic scenarios.

Force > Ask If the workflow must be correct every time, enforce it with prerequisites, hooks, tool choice, or schemas.
Split > Big Prompt Large reviews and open-ended workflows should be decomposed into passes or subagents.
Tools > Prompts When tool selection is bad, first fix tool descriptions, names, boundaries, and access scope.
Structure > Free Text Prefer tool_use, JSON schema, explicit fields, structured errors, claim-source mappings, and manifests.
Limit > Expand Fewer tools, narrower roles, trimmed context, and scoped rules usually improve reliability.
Coordinator Owns Coverage In multi-agent systems, missing coverage usually means bad decomposition or poor context passing.
What the Exam Emphasizes

Domain 1

27%

Agentic Architecture & Orchestration

  • Loop on stop_reason
  • Coordinator-subagent patterns
  • Explicit context passing
  • Hooks, handoffs, prerequisites
  • Session resume vs fresh start

Domain 2

18%

Tool Design & MCP Integration

  • Tool descriptions drive selection
  • Structured MCP errors
  • Scoped tool sets
  • MCP project vs user config
  • Resources for catalogs

Domain 3

20%

Claude Code Configuration & Workflows

  • CLAUDE.md hierarchy
  • .claude/rules path scoping
  • Commands and skills
  • Plan mode vs direct execution
  • CI usage with print/json flags

Domain 4

20%

Prompt Engineering & Structured Output

  • Explicit criteria reduce false positives
  • Few-shot for ambiguity and consistency
  • JSON schema + tool_use
  • Validation-retry loops
  • Batch vs real-time

Domain 5

15%

Context Management & Reliability

  • Trim tool outputs
  • Persist structured facts
  • Escalation criteria
  • Crash recovery and scratchpads
  • Provenance and uncertainty handling

Meta

Exam style

How answers are framed

  • Single best answer
  • Usually a pragmatic first step
  • Rejects over-engineering
  • Rejects vague “improve prompt” fixes
  • Rewards explicit, narrow, reliable systems
Core Decision Tree: Few-shot vs More Examples vs Prompt Change vs System Change

1. Ask what kind of problem this is

  • Ambiguity? The model can do the task but makes inconsistent judgments.
  • Criteria problem? It does not know what to count, skip, or prioritize.
  • Guarantee problem? A step must always happen, output must be structured, or a business rule must be enforced.
  • Architecture problem? Coverage is missing, tool boundaries overlap, or workflow decomposition is wrong.

Use Few-shot

  • Ambiguous cases
  • Inconsistent formatting
  • Need examples of “report this, ignore that”
  • Need examples of correct handling across varied document formats
Good signal: “the model is inconsistent”

Change the Prompt

  • Need explicit criteria
  • Need clearer severity rules
  • Need narrower review scope
  • Need better escalation rules
Good signal: “too many false positives”

Change the System

  • Prerequisites, hooks, tool choice, schemas
  • Scoped tool access
  • Coordinator decomposition
  • Structured error payloads
Good signal: “must always”, “critical”, “production reliability”
Trap Map: If You See This, They Usually Want That
Situation / SignalWhat is really wrongLikely correct moveCommon tempting wrong moveWhy / exam philosophy
Tool selection is wrongTool descriptions are minimal or overlappingRewrite descriptions with inputs, outputs, examples, edges, boundariesAdd few-shot first, add classifier, merge tools immediatelyTools are the first fix for tool confusion
A specific order must always happenIdentity verification before refund, metadata extraction before enrichmentUse programmatic prerequisite, hook, or forced tool choiceStronger prompt, more examples, confidence checksIf it must always happen, enforce it
Coverage is incomplete in multi-agent researchVisual arts covered, music/writing/film missingFix coordinator decomposition and task partitioningBlame synthesis agent or search agent firstCoordinator owns coverage
Subagent output lacks contextSynthesis gets summaries without sources or datesPass structured findings with metadata explicitlyAssume context inheritanceSubagents do not inherit parent context automatically
Need multiple subagents at onceParallel search/analysis tasksEmit multiple Task calls in one coordinator responseSpawn sequentially across turns by defaultParallel spawning is preferred when independent
Model returns free text when schema is neededCI comments, extraction, machine-readable resultsUse tool_use and JSON schema, often with tool_choice anyAsk for JSON in natural languageStructure must be guaranteed
Output passes schema but content is still wrongTotals do not add up, values in wrong fieldsAdd semantic validation and retry with explicit error feedbackAssume schema solved correctnessSchemas fix syntax, not semantics
Review has too many false positivesDevelopers stop trusting the toolWrite explicit report/skip criteria, disable noisy categories temporarilySay “be conservative” or “only high confidence”Specific criteria beats vague caution
Need consistent formatting and judgmentSame type of issue formatted differently each runAdd 2 to 4 targeted few-shot examplesDump many generic examplesFew-shot works best when examples are sharp and targeted
Task is complex and architecturalMigration, restructuring, dozens of filesUse plan mode first, then executeJump directly into changesPlan mode is for large uncertain work
Task is small and obviousSingle-file validation checkUse direct executionEnter plan mode automaticallyDo not over-plan simple tasks
Long session is degradingModel references generic patterns instead of discovered factsUse scratchpads, summaries, /compact, subagentsKeep everything in one long chatContext drift is expected
Tool outputs are bloating context40 fields returned, only 5 matterTrim to relevant fields and persist structured factsKeep full output in historyContext is a budget, not an archive
User explicitly wants a humanCustomer asks for a human agentEscalate immediatelyContinue trying to solve unless they insist twiceDirect human request is an escalation trigger
Case is emotionally negative but solvableFrustrated customer, straightforward issueAcknowledge frustration and solve if possibleEscalate only because sentiment is negativeSentiment is not complexity
Multiple customer matches are returnedName maps to several accountsAsk for another identifierPick the most likely matchNo heuristic guessing on identity
Tool fails transientlyTimeout, service unavailableReturn structured error with retryability, recover locally if possibleReturn empty success or generic failureErrors must enable recovery
Tool hits business ruleRefund too large or policy exceptionBlock and route to escalation with explanationRetry repeatedlyBusiness errors are not retry problems
Agent has too many tools18 tools across rolesReduce tool scope per agent, maybe give one cross-role verifierGive even more toolsToo many tools degrades selection
Need simple fact verification from synthesisDates, names, simple stats often need checkingGive synthesis a scoped verify_fact toolGive synthesis all search toolsUse narrow cross-role tools for common needs
Need shared team rules in Claude CodeEveryone cloning repo needs themProject-scoped CLAUDE.md or .claude/commands/.claude/rulesPut them in ~/.claudeUser-level config is not shared
Need path-specific conventionsTests scattered across repoUse .claude/rules with glob patternsUse directory CLAUDE.md for cross-cutting patternsPath-scoped rules beat directory location here
Need reusable command for teamSlash command should ship with repoPut it in .claude/commands/Put it only in user homeProject commands are version-controlled
Need isolated verbose workflowExploratory skill or brainstormingUse skill with context: forkRun it in main contextFork preserves main context
Need CI usageNo interactive hangs, structured outputsUse -p plus —output-format json and —json-schemaRun interactive defaultsCI needs non-interactive structured mode
Need overnight cost savingsNightly report, weekly auditUse Message Batches APIUse batch for blocking pre-merge checksBatch is for latency-tolerant workloads
Need review of large PR14 files, contradictory resultsSplit into per-file local passes plus integration passOne giant pass, larger model onlyMulti-pass beats single-pass
Need second opinion on generated codeGenerator reviewed its own outputUse an independent review instanceAsk same session to self-review harderSame-session self-review is weaker
Need to choose resume vs fresh startFiles changed since prior analysisResume only if context is mostly valid and tell it what changed, otherwise inject structured summary in new sessionBlindly trust stale session outputsFresh summaries can beat stale history
Need provenance in synthesisMultiple sources, conflicting stats, dates matterPreserve claim-source mappings, dates, contested vs established sectionsCollapse to one number without attributionProvenance survives only if structured
Few-shot, More Examples, Prompt Rewrite: Where People Lose Points

Use Few-shot When…

  • The model already understands the task class.
  • It is inconsistent on edge cases.
  • You need stable output format.
  • You need to show “report vs skip” decisions.
  • You need varied examples for extraction formats.
Ambiguity
Consistency
Format

Do Not Reach for Few-shot When…

  • A step must be guaranteed.
  • A tool is poorly described.
  • The coordinator decomposed the task badly.
  • You need access control or business-rule enforcement.
Guarantee problem
Architecture problem

Provide More Examples When…

  • The issue is not one weird edge case, but a pattern.
  • You want coverage across multiple document or code patterns.
  • You need the model to generalize the desired judgment.
The guide repeatedly prefers 2 to 4 targeted examples, not a giant example dump.

Rewrite the Prompt When…

  • Criteria are vague.
  • Severity is inconsistent.
  • False positives are high.
  • Escalation boundaries are unclear.
Criteria problem
Precision problem

Exam Trap

  • ”Add few-shot” often looks smart.
  • But the exam usually wants it only after simpler structural fixes.
  • If tool descriptions are bad, fix those first.
  • If compliance is required, enforce in code.

Fast Memory Rule

  • Examples teach judgment.
  • Prompts teach criteria.
  • Systems enforce guarantees.
  • Architecture fixes coverage.
Agentic Orchestration Patterns

Loop Control

  • Continue when stop_reason = tool_use
  • Stop when stop_reason = end_turn
  • Append tool results into history each iteration
Do not parse assistant text to decide when loop ends

Coordinator Role

  • Decide whether subagents are needed
  • Partition scope to avoid duplication
  • Aggregate and inspect for gaps
  • Redelegate targeted follow-up work

Subagent Context

  • No automatic inheritance
  • Pass complete findings explicitly
  • Separate content from metadata
  • Preserve attribution for downstream synthesis

Parallelization

  • Spawn multiple Task calls in one response for independent work
  • Use hub-and-spoke routing through coordinator
  • Do not let agents talk to each other directly

Handoffs

  • Human handoff should include customer details, root cause, key amounts, recommended action
  • Assume human may not see conversation history

Hooks and Enforcement

  • PostToolUse for normalization
  • Tool interception for business rules
  • Choose hooks when compliance must be deterministic
Error Handling and Recovery Patterns

Error payload should include

  • errorCategory
  • isRetryable
  • Human-readable explanation
  • Attempted action / query
  • Partial results
  • Potential alternatives

Transient

  • Timeouts
  • Service unavailable
  • Recover locally first
  • Propagate with structured context if unresolved
Retryable sometimes

Validation

  • Bad inputs
  • Ask for missing identifier
  • Fix request shape
Do not retry blindly

Permission

  • Missing access
  • Agent should explain limits
  • Coordinator may choose alternative path or escalation

Business / Policy

  • Refund exceeds threshold
  • Policy exception needed
  • Return non-retryable explanation and escalate
Not a transient failure

Anti-patterns

  • Generic “operation failed”
  • Return empty success for failures
  • Terminate whole workflow on one subagent failure
  • Hide partial results
Claude Code and MCP Quick Map

CLAUDE.md hierarchy

  • User: ~/.claude/CLAUDE.md
  • Project: root or .claude/CLAUDE.md
  • Directory-level files for local scope
User-level is not shared with team

.claude/rules

  • Use YAML frontmatter with paths
  • Great for tests across many directories
  • Reduces irrelevant context

Commands and skills

  • Shared slash commands: .claude/commands/
  • Skills: .claude/skills/
  • context: fork isolates verbose work
  • allowed-tools restricts capability

Built-in tools

  • Grep = content search
  • Glob = path matching
  • Read/Write = whole file operations
  • Edit = targeted change if anchor unique
  • Fallback: Read + Write when Edit is unreliable

MCP config

  • Project shared servers: .mcp.json
  • Personal servers: ~/.claude.json
  • Use env vars like $TOKEN
  • MCP resources expose catalogs and reduce exploratory calls

CI mode

  • -p or —print for non-interactive use
  • —output-format json
  • —json-schema
  • Independent review instance beats self-review
Context, Escalation, Provenance

Context trimming

  • Extract persistent facts: ids, dates, amounts, statuses
  • Keep facts outside summary blocks
  • Trim tool outputs to relevant fields

Lost in the middle

  • Put key findings near beginning
  • Use section headers
  • Prefer structured summaries over long middle sections

Escalate when…

  • User asks for human
  • Policy is ambiguous or missing
  • Cannot make meaningful progress
Not because sentiment is negative

Identity ambiguity

  • Multiple matches means ask for another identifier
  • No heuristic selection

Human review workflows

  • Use field-level confidence
  • Calibrate with labeled sets
  • Sample high-confidence cases too
  • Check accuracy by field and document type

Provenance

  • Preserve claim-source mappings
  • Include dates
  • Annotate conflicts instead of picking one value
  • Separate established vs contested findings
Question Phrases → Likely Answer Pattern
If the question sounds like thisThink this first
”must always”, “mandatory before”, “critical business logic”Programmatic enforcement, prerequisite gate, hook
”similar tools”, “minimal descriptions”, “wrong tool chosen”Improve tool descriptions and boundaries first
”ambiguous”, “inconsistent format”, “sometimes does X, sometimes Y”Few-shot examples
”false positives”, “trust erosion”, “be more conservative”Explicit criteria, maybe disable noisy categories
”large PR”, “14 files”, “contradictory findings”Multi-pass review
”restructure”, “migration”, “many files”, “multiple valid approaches”Plan mode
”returns text instead of structure”, “CI”, “extraction”tool_use / JSON schema / tool_choice
”passes schema but wrong values”Semantic validation and retry with feedback
”timeout”, “partial results”, “recovery”Structured error propagation
”missing areas in report”Coordinator decomposition too narrow
”subagent didn’t know prior findings”Explicit context passing
”everyone on team should get this”Project-scoped config, not user-scoped
”tests live all over repo”Path-scoped rules with globs
”nightly / weekly / overnight”Message Batches API
”blocking merge”Synchronous API / non-batch
”same session reviewed its own code”Independent review instance