Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Checkpoint Streaming? #167

Open
blester125 opened this issue Jan 27, 2023 · 1 comment
Open

Checkpoint Streaming? #167

blester125 opened this issue Jan 27, 2023 · 1 comment
Labels
Speed Up Working towards decreasing the latency of git theta commands.

Comments

@blester125
Copy link
Collaborator

One thing that will always slow down Git Theta is the fact we have to wait for the native framework to be loaded in it's entirety before we can start processing parameters.

Will it be possible to start processing parameters as they are loaded? It would require support from the downstream framework, it would require custom reader code that understands a framework's format and does custom reading.

A similar approach would be to do lazy loading for parameters where the structure of the model is pre-populated but the values are only loaded when needed, this would allow things to start quicker and for the readings to be interleaved with the rest of execution via async.

This is a very ambitious idea and should probably be on of the last speedups we try

@blester125 blester125 added the Speed Up Working towards decreasing the latency of git theta commands. label Jan 27, 2023
@blester125
Copy link
Collaborator Author

Some checkpoint plug-ins, like safetensors, support lazy loading which would probably be part of an implementation like this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Speed Up Working towards decreasing the latency of git theta commands.
Projects
None yet
Development

No branches or pull requests

1 participant