-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
❌ use vllm create_final_community_reports error #640
Comments
11:17:11,575 graphrag.index.run ERROR error running workflow create_final_community_reports
|
I was having similar issue, for me the solution was in changing the community_report.txt prompt. In my case the model was generating community reports in a broken json format because of double curly braces in the prompt. When I changed the prompt by putting single curly braces instead of double ones - the error got solved. Here are my error logs): and here is the updated prompt part:
|
Consolidating alternate model issues here: #657 |
I switched to lmdeploy. It's about 2x faster then vllm when serving qwen2-7b and won't be stuck during create_final_community_reports. |
I was actually able to use functionary vllm server with a llama-3.1 AWQ quant and indexing as well as both local and global search works like a charm. |
only when create_final_community_reports , display error almost when is finished 。
None
⠙ GraphRAG Indexer
├── Loading Input (InputFileType.text) - 1 files loaded (0 filtered) 100%
├── create_base_text_units
├── create_base_extracted_entities
├── create_summarized_entities
├── create_base_entity_graph
├── create_final_entities
├── create_final_nodes
├── create_final_communities
├── join_text_units_to_entity_ids
├── create_final_relationships
├── join_text_units_to_relationship_ids
└── create_final_community_reports
❌ Errors occurred during the pipeline run, see logs for more details.
03:52:10,989 graphrag.index.reporting.file_workflow_callbacks INFO Error Invoking LLM details={'input': '\nYou are a helpful assistant responsible for >
01:07:09,532 graphrag.index.graph.extractors.community_reports.community_reports_extractor ERROR error generating community report
01:07:10,885 graphrag.index.reporting.file_workflow_callbacks INFO Community Report Extraction Error details=None
01:07:10,885 graphrag.index.verbs.graph.report.strategies.graph_intelligence.run_graph_intelligence WARNING No report found for community: 15
01:07:10,930 datashaper.workflow.workflow INFO executing verb window
01:07:10,931 datashaper.workflow.workflow ERROR Error executing verb "window" in create_final_community_reports: 'community'
vllm serve /mnt/vdb1/model/Qwen2-72B-Instruct-GPTQ-Int4
--host 0.0.0.0
--port 8000
--served-model-name qwen2-72b
--max-model-len 12000
--gpu-memory-utilization 0.8
--dtype bfloat16
--enforce_eager
llm:
api_key: "xx"
type: openai_chat
model: qwen2-72b
api_base: http://10.x.x.x:8000/v1
model_supports_json: true
max_tokens: 4000
request_timeout: 180.0
another question:
only when create community_reports, GPU is at low load about 5% , query once only every 180s , so speed is very slow.
The text was updated successfully, but these errors were encountered: