Skip to content

Conversation

@BenjaminBossan
Copy link
Member

@BenjaminBossan BenjaminBossan commented Oct 16, 2025

There are already GPU tests for P-tuning and prefix tuning, but none that involve actual training. This PR adds two training tests (analog to what we have with LoRA) using transformers Trainer.

This can be important because changes in Trainer, for instance how cache is handled, can impact the possibility of training with prompt learning methods.

A recent example for a change is huggingface/transformers#41585. Although it doesn't break anything, it shows the potential for breaking in the Trainer class that could impact these PEFT methods.

There are already GPU tests for P-tuning and prefix tuning, but none
that involve actual training. This PR adds two training tests (analog to
what we have with LoRA) using transformers Trainer.

This can be important because changes in Trainer, for instance how cache
is handled, can impact the possibility of training with prompt learning
methods.

A recent example for a change is
huggingface/transformers#41585. Although it
doesn't break anything, it shows the potential for breaking in the
Trainer class that could impact these PEFT methods.
@BenjaminBossan BenjaminBossan changed the title TST Add GPU training tests for p/prefix tuning TST Add GPU training tests for p-tuning & prefix tuning Oct 16, 2025
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants