19
19
20
20
from neo4j_graphrag .exceptions import LLMGenerationError
21
21
from neo4j_graphrag .llm .base import LLMInterface
22
- from neo4j_graphrag .llm .types import LLMResponse , MessageList
22
+ from neo4j_graphrag .llm .types import LLMResponse , MessageList , BaseMessage
23
23
24
24
try :
25
25
from vertexai .generative_models import (
@@ -77,7 +77,7 @@ def __init__(
77
77
self .model_params = kwargs
78
78
79
79
def get_messages (
80
- self , input : str , message_history : Optional [list [Any ]] = None
80
+ self , input : str , message_history : Optional [list [BaseMessage ]] = None
81
81
) -> list [Content ]:
82
82
messages = []
83
83
if message_history :
@@ -87,16 +87,16 @@ def get_messages(
87
87
raise LLMGenerationError (e .errors ()) from e
88
88
89
89
for message in message_history :
90
- if message .get ( " role" ) == "user" :
90
+ if message .role == "user" :
91
91
messages .append (
92
92
Content (
93
- role = "user" , parts = [Part .from_text (message .get ( " content" ) )]
93
+ role = "user" , parts = [Part .from_text (message .content )]
94
94
)
95
95
)
96
- elif message .get ( " role" ) == "assistant" :
96
+ elif message .role == "assistant" :
97
97
messages .append (
98
98
Content (
99
- role = "model" , parts = [Part .from_text (message .get ( " content" ) )]
99
+ role = "model" , parts = [Part .from_text (message .content )]
100
100
)
101
101
)
102
102
@@ -106,7 +106,7 @@ def get_messages(
106
106
def invoke (
107
107
self ,
108
108
input : str ,
109
- message_history : Optional [list [Any ]] = None ,
109
+ message_history : Optional [list [BaseMessage ]] = None ,
110
110
system_instruction : Optional [str ] = None ,
111
111
) -> LLMResponse :
112
112
"""Sends text to the LLM and returns a response.
@@ -137,7 +137,7 @@ def invoke(
137
137
raise LLMGenerationError (e )
138
138
139
139
async def ainvoke (
140
- self , input : str , message_history : Optional [list [Any ]] = None
140
+ self , input : str , message_history : Optional [list [BaseMessage ]] = None
141
141
) -> LLMResponse :
142
142
"""Asynchronously sends text to the LLM and returns a response.
143
143
0 commit comments