You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: docs/source/en/reference/agents.md
+4-4
Original file line number
Diff line number
Diff line change
@@ -65,7 +65,7 @@ You could use any `model` callable for your agent, as long as:
65
65
1. It follows the [messages format](./chat_templating) (`List[Dict[str, str]]`) for its input `messages`, and it returns a `str`.
66
66
2. It stops generating outputs *before* the sequences passed in the argument `stop_sequences`
67
67
68
-
For defining your LLM, you can make a `custom_model` method which accepts a list of [messages](./chat_templating) and returns text. This callable also needs to accept a `stop_sequences` argument that indicates when to stop generating.
68
+
For defining your LLM, you can make a `custom_model` method which accepts a list of [messages](./chat_templating) and returns an object with a .content attribute containing the text. This callable also needs to accept a `stop_sequences` argument that indicates when to stop generating.
69
69
70
70
```python
71
71
from huggingface_hub import login, InferenceClient
0 commit comments