refine docs

This commit is contained in:
zachary62 2024-12-28 04:40:44 +00:00
parent d7cc366d15
commit 382dfd4a63
2 changed files with 6 additions and 5 deletions

View File

@ -39,6 +39,7 @@ We model the LLM workflow as a **Nested Flow**:
- Map Reduce - Map Reduce
- RAG - RAG
- Structured Output - Structured Output
- Evaluation
## Example Use Cases ## Example Use Cases

View File

@ -4,10 +4,10 @@ title: "LLM Integration"
nav_order: 3 nav_order: 3
--- ---
# Call LLM # LLM Wrappers
For your LLM application, implement a function to call LLMs yourself. For your LLM app, implement a wrapper function to call LLMs yourself.
You can ask an assistant like ChatGPT or Claude to generate an example. You can ask an assistant like ChatGPT or Claude to implement it.
For instance, asking ChatGPT to "implement a `call_llm` function that takes a prompt and returns the LLM response" gives: For instance, asking ChatGPT to "implement a `call_llm` function that takes a prompt and returns the LLM response" gives:
```python ```python
@ -64,9 +64,9 @@ def call_llm(prompt):
``` ```
## Why not provide an LLM call function? ## Why not provide LLM Wrappers?
I believe it is a bad practice to provide LLM-specific implementations in a general framework: I believe it is a bad practice to provide LLM-specific implementations in a general framework:
- LLM APIs change frequently. Hardcoding them makes maintenance difficult. - LLMs change frequently. Hardcoding them makes maintenance difficult.
- You may need flexibility to switch vendors, use fine-tuned models, or deploy local LLMs. - You may need flexibility to switch vendors, use fine-tuned models, or deploy local LLMs.
- You may need optimizations like prompt caching, request batching, or response streaming. - You may need optimizations like prompt caching, request batching, or response streaming.