-
Notifications
You must be signed in to change notification settings - Fork 320
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running Llama3 like a genius (that I am not) [SOLVED]: #1669
Comments
Finally was able to solve the issue. Here are the steps to run the new Llama3 model. Step 1 - GET FILESGet the model from here: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct Step 2 - PREPARE FILES
Step 3 - CONVERT TO CTRANSLATE2 FORMATConvert the model to the ctranslate2 format. You can do this one of two ways:
CLICK HERE FOR MY VERY EXCELLENT SCRIPT
If you correctly install Pyside6 and run the script you should see an easy-to-understand GUI that will convert any compatible model. You only need to select the folder containing the model files: Step 4 - FIX THE STUPID VALUE IN CONFIG.JSON
Step 5 - PROMPT FORMAT
Note, this does NOT entail a multi-turn conversation with memory. For that you'll need to consult the
Step 6 - PROPER TOKENIZATION and PROMPTINGI've complained before about the dearth of examples on how to use stereotypical "chat" models. The only helpful example from the "Docs" was for Falcon, which I adapted. Anyhow, I'll provide my full script below for the benefit of the community, but before that, it's helpful to understand a few things:
Step 7 - SAMPLE SCRIPTWithout further ado, here is a sample script. I have put in all caps placeholders for personal information as well as things that depend on your use case; MY VERY COOL AWESOME VERY DECENT SCRIPT USING LLAMA3
I'd love to hear from true experts on |
I push the MR #1671 to fix llama 3. There is only a problem about the |
got this error when running your conversion tool for meta-llama/Llama-3.2-3B-Instruct any ideas? Starting conversion for float32 with command: Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s] |
Yep, in the folder containing the model there's a folder named ".cache" that you need to delete and then it should work. I can't figure out how to exclude it from the download process in my script. I should probably have it manually deleted afterwards. Be aware that it may be hidden so you need to show hidden files. |
This issue was closed long time ago. I forgot about it, but you should try my converter on my repository instead it still experiences the same issue sometimes but as long as you know to delete the cache folder you should be good. |
excellent, will do! |
Deleting my initial message because it contained paths on my computer and my follow up post below addresses the error I was getting anyways!
The text was updated successfully, but these errors were encountered: