Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about Custom Attention Mask Support #1454

Open
BeiningWu opened this issue Dec 20, 2024 · 1 comment
Open

Question about Custom Attention Mask Support #1454

BeiningWu opened this issue Dec 20, 2024 · 1 comment

Comments

@BeiningWu
Copy link

Hi, I am currently using the Unsloth framework for fine-tuning large models. I have a question regarding the support for custom attention masks.

As of now, I understand that the framework supports the causal mask. However, I would like to know if it's possible to implement or customize a bidirectional attention mask, or if this feature is supported out of the box.

Could you please clarify whether this functionality is available, and if not, provide guidance on how to implement a bidirectional mask for my use case?

Thank you!

@shimmyshimmer
Copy link
Collaborator

I'm pretty sure you'll be able to as I saw a user talking about managing to do it but it will require custom stuff. You can join our Discord and maybe ask there if you'd like: https://discord.com/invite/unsloth

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants