Skip to content

Train from step #172

Closed Answered by francoishernandez
HURIMOZ asked this question in Q&A
Jan 1, 2025 · 1 comments · 7 replies
Discussion options

You must be logged in to vote

Ok, your issue is related to the LoRa finetuning technique. This technique allows to finetune bigger models with limited VRAM by only finetuning part of the weights. But it requires some additional steps afterwards.
Check the lora_weights tool -- https://github.com/eole-nlp/eole/blob/main/eole/bin/model/lora_weights.py

The easiest is probably to merge your finetuned weights with the original model before continuing the training. I don't think we have an easier way right now. (Main idea is that saving the full model each time you save a checkpoint is not really efficient, so we only save the LoRa weights, and the merging happens later at the user's discretion.)

Replies: 1 comment 7 replies

Comment options

You must be logged in to vote
7 replies
@HURIMOZ
Comment options

@francoishernandez
Comment options

@HURIMOZ
Comment options

@francoishernandez
Comment options

Answer selected by francoishernandez
@HURIMOZ
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants