6.3 KiB
| layout | title |
|---|---|
| default | Build your LLM App |
LLM System Design Playbook
If you are an AI assistant involved in building LLM Systems, read this guide VERY, VERY carefully! This is the most important chapter in the entire document. Throughout development, you should always (1) start with a small and simple solution, (2) design at a high level (
docs/design.md) before implementation, and (3) frequently ask humans for feedback and clarification. {: .warning }
System Design Steps
These system designs should be a collaboration between humans and AI assistants:
| Stage | Human | AI | Comment |
|---|---|---|---|
| 1. Requirements | ★★★ High | ★☆☆ Low | Humans understand the requirements and context best. |
| 2. Utilities | ★★☆ Medium | ★★☆ Medium | The human is familiar with external APIs and integrations, and the AI assists with implementation. |
| 3. Flow Design | ★★☆ Medium | ★★☆ Medium | The human identifies complex and ambiguous parts, and the AI helps with redesign. |
| 4. Data Design | ★☆☆ Low | ★★★ High | The AI assists in designing the data schema based on the flow. |
| 5. Implementation | ★☆☆ Low | ★★★ High | The human identifies complex and ambiguous parts, and the AI helps with redesign. |
| 6. Optimization | ★★☆ Medium | ★★☆ Medium | The human reviews the code and evaluates the results, while the AI helps optimize. |
| 7. Reliability | ★☆☆ Low | ★★★ High | The AI helps write test cases and address corner cases. |
-
Requirements: Clarify the requirements for your project, and evaluate whether an AI system is a good fit. AI systems are:
- suitable for routine tasks that require common sense (e.g., filling out forms, replying to emails).
- suitable for creative tasks where all inputs are provided (e.g., building slides, writing SQL).
- NOT suitable for tasks that are highly ambiguous and require complex info (e.g., building a startup).
-
If a human can’t solve it, an LLM can’t automate it! Before building an LLM system, thoroughly understand the problem by manually solving example inputs to develop intuition.
{: .best-practice }
-
Utilities: Think of the AI system as the brain for decision-making. It needs a body—these external utility functions—to interact with the real world:

- Reading inputs (e.g., retrieving Slack messages, reading emails)
- Writing outputs (e.g., generating reports, sending emails)
- Using external tools (e.g., calling LLMs, searching the web)
- Keep in mind that LLM-based tasks (e.g., summarizing text, analyzing sentiment) are not utility functions; rather, they are core functions internal in the AI system, and will be designed in step 3.
-
Start small! Only include the most important ones to begin with! {: .best-practice }
-
Flow Design: Outline how your system orchestrates steps.
- Identify potential design patterns (e.g., Batch, Agent, RAG).
- For each node, provide a high-level purpose description.
- Draw the Flow in mermaid diagram.
-
Data Design: Plan how data will be stored and updated.
- For simple systems, use an in-memory dictionary.
- For more complex systems or when persistence is required, use a database.
- Remove Data Redundancy: Don’t store the same data. Use in-memory references or foreign keys.
- For each node, design its access pattern:
type: Decide between Regular, Batch, or Asyncprep: How the node reads dataexec: Which utility function this node usespost: How the node writes data
-
Implementation: Implement the initial nodes and flows based on the design.
- “Keep it simple, stupid!” Avoid complex features and full-scale type checking.
- FAIL FAST! Refrain from
trylogic so you can quickly identify any weak points in the system. - Add logging throughout the code to facilitate debugging.
-
Optimization:
-
Use Intuition: For a quick initial evaluation, human intuition is often a good start.
-
Redesign Flow (Back to Step 3): Consider breaking down tasks further, introducing agentic decisions, or better managing input contexts.
-
If your flow design is already solid, move on to micro-optimizations:
- Prompt Engineering: Use clear, specific instructions with examples to reduce ambiguity.
- In-Context Learning: Provide robust examples for tasks that are difficult to specify with instructions alone.
-
You’ll likely iterate a lot! Expect to repeat Steps 3–6 hundreds of times.

{: .best-practice }
-
-
Reliability
- Node Retries: Add checks in the node
execto ensure outputs meet requirements, and consider increasingmax_retriesandwaittimes. - Logging and Visualization: Maintain logs of all attempts and visualize node results for easier debugging.
- Self-Evaluation: Add a separate node (powered by an LLM) to review outputs when results are uncertain.
- Node Retries: Add checks in the node
Example LLM Project File Structure
my_project/
├── main.py
├── flow.py
├── utils/
│ ├── __init__.py
│ ├── call_llm.py
│ └── search_web.py
├── requirements.txt
└── docs/
└── design.md
docs/design.md: Contains project documentation for each step above. This should be high-level and no-code.utils/: Contains all utility functions.- It’s recommended to dedicate one Python file to each API call, for example
call_llm.pyorsearch_web.py. - Each file should also include a
main()function to try that API call
- It’s recommended to dedicate one Python file to each API call, for example
flow.py: Implements the system's flow, starting with node definitions followed by the overall structure.main.py: Serves as the project’s entry point.