Replies: 4 comments 1 reply
-
we meet the same promblem |
Beta Was this translation helpful? Give feedback.
-
Same any news? :) |
Beta Was this translation helpful? Give feedback.
-
I attempted to try the Mixtral 8x7B instruct GPTQ model and fine-tuned it to a custom dataset. I applied to the latest LlamaIndex Vllm call but met an OSError: does not appear to have a file named config.json. config.json is located in the base model so, see if there's any way to point to the tokenizer path into the Vlllm.... but not. It's interesting since the previous Vllm accepted the tokenizer path input. |
Beta Was this translation helpful? Give feedback.
-
I met this error during the call LLM with the Mixtral8x7B fine-tuned LoRA Merged Model. It worked well in torch 2.1.2 and vlllm 0.3.3 but torch 2,3 and vlllm 0.4.2 met this valueError. What is missing here? Should I use the different quantization method like awq or gptq than bitsandbytes?
Beta Was this translation helpful? Give feedback.
All reactions