Testing GraphRAG with other LLMs #321
Replies: 7 comments 2 replies
-
Yes I am also looking forward to see the performance with other models |
Beta Was this translation helpful? Give feedback.
-
I got the Get Started guide working entirely locally on Apple Silicon. SetupToolioI used https://github.com/OoriData/Toolio as my OpenAI-compatible server for text completions, because it uses Apple's MLX and supports JSON schemas. I started the server with: toolio_server --model=mlx-community/Hermes-2-Theta-Llama-3-8B-4bit I also used a model with a larger context to satisfy the community report of graphrag: toolio_server --model=mlx-community/Llama-3-8B-Instruct-1048k-4bit I changed the open-text-embeddingsFor embeddings, I used open-text-embeddings. I started the server with: PORT=8080 VERBOSE=1 MODEL=BAAI/bge-large-en python -m open.text.embeddings.server I changed the With this, I was able to get the indexing. I think it took around 2 hours. I was also able to run the global and local search queries, after resolving some tweaks which I describe next. IssuesEnhancement: JSON Schema defined informally instead of formallyToolio can work with JSON schemas and enforce them. I added the following schema: MAP_SYSTEM_JSON = """
{
"type": "object",
"properties": {
"points": {
"type": "array",
"contains": {
"type": "object",
"properties": {
"description": { "type": "string" },
"score": { "type": "number" },
},
"required": ["description", "score"],
},
"minContains": 5,
},
},
"required": ["points"],
}
""" And then added a I suspect whenever GraphRAG informally prompts for some JSON, the effectiveness with Toolio could be improved by passing the schema formally. Some bug with
|
Beta Was this translation helpful? Give feedback.
-
It's work on chatglm4-9b-chat with xinference |
Beta Was this translation helpful? Give feedback.
-
I tried deepseek as a lower-level alternative to gpt4o, and mentioned in the issue that many utf8 and json problems (before 0.2.1) will not be encountered in this model |
Beta Was this translation helpful? Give feedback.
-
I've been experimenting with @TheAiSingularity's repo here. This is not an endorsement, but it seems to work well leveraging ollama locally on Apple Silicon to try out different models. |
Beta Was this translation helpful? Give feedback.
-
I am using a custom vLLM server fork incl. openai compatible tool & function calling as well as an embeddings endpoint using funtionary-3.1-small (llama3.1-8B) AWQ quant on my RTX 3080 and indexing as well as global search work amazingly well. Both with Will try with bigger versions of the functionary model family in the near future... Local search does not work. But did not look to deeply, why that is the case. |
Beta Was this translation helpful? Give feedback.
-
I think it would be great if GraphRAG project was built on top of another Microsoft project, Autogen. This way, interoperability with various LLM providers would need to be implemented only in one place. Another approach is to use existing open-source LLM interoperability libraries like LiteLLM. |
Beta Was this translation helpful? Give feedback.
-
The team primarily built GraphRAG with the GPT4-family of models and current prompts have been tested to work well with GPT4-o.
We'd love to see how GraphRAG works with open models like phi3, llama, mistral, etc.
Share your ideas, experiments, and experiences here!
Beta Was this translation helpful? Give feedback.
All reactions