Update llm.md

This commit is contained in:
Zach 2025-04-05 10:30:16 -04:00 committed by GitHub
parent a159c649ae
commit 050eb2a402
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
1 changed files with 17 additions and 4 deletions

View File

@ -32,14 +32,14 @@ Here, we provide some minimal example implementations:
def call_llm(prompt): def call_llm(prompt):
from anthropic import Anthropic from anthropic import Anthropic
client = Anthropic(api_key="YOUR_API_KEY_HERE") client = Anthropic(api_key="YOUR_API_KEY_HERE")
response = client.messages.create( r = client.messages.create(
model="claude-3-7-sonnet-20250219", model="claude-3-7-sonnet-20250219",
max_tokens=3000, max_tokens=3000,
messages=[ messages=[
{"role": "user", "content": prompt} {"role": "user", "content": prompt}
] ]
) )
return response.content[0].text return r.content[0].text
``` ```
3. Google (Generative AI Studio / PaLM API) 3. Google (Generative AI Studio / PaLM API)
@ -47,11 +47,11 @@ Here, we provide some minimal example implementations:
def call_llm(prompt): def call_llm(prompt):
import google.generativeai as genai import google.generativeai as genai
genai.configure(api_key="YOUR_API_KEY_HERE") genai.configure(api_key="YOUR_API_KEY_HERE")
response = genai.generate_text( r = genai.generate_text(
model="models/text-bison-001", model="models/text-bison-001",
prompt=prompt prompt=prompt
) )
return response.result return r.result
``` ```
4. Azure (Azure OpenAI) 4. Azure (Azure OpenAI)
@ -81,6 +81,19 @@ Here, we provide some minimal example implementations:
return response.message.content return response.message.content
``` ```
6. DeepSeek
```python
def call_llm(prompt):
from openai import OpenAI
client = OpenAI(api_key="YOUR_DEEPSEEK_API_KEY", base_url="https://api.deepseek.com")
r = client.chat.completions.create(
model="deepseek-chat",
messages=[{"role": "user", "content": prompt}]
)
return r.choices[0].message.content
```
## Improvements ## Improvements
Feel free to enhance your `call_llm` function as needed. Here are examples: Feel free to enhance your `call_llm` function as needed. Here are examples: