Skip to content

try linear attention #435

@jpata

Description

@jpata

This library https://github.com/fla-org/flash-linear-attention wraps various linear attention mechanisms.
Try some common ones from there, to see if we can get a better physics / computational tradeoff, suitable for CPU or legacy GPU inference that is useful for CERN.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions