You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Some functions in Torchx.Backend create intermediate Iotas and similar tensors. These should be allocated in the same device as the function's input tensor.
The text was updated successfully, but these errors were encountered:
polvalente
changed the title
[Torchx] Allocate intermediate tensors in the same device as the input tensor
Allocate intermediate tensors in the same device as the input tensor
Mar 21, 2022
a lot of developers have Apple Silicon (and many has it provided by their companies)
due to unified memory Mac Book Pro with 30+ GB of memory is capable of ~22b model serving. I tested it with ollama and codestral 22b for example. It's not blazing fast, but my 3080 RTX cannot even run such models.
GPU setup with 30+ Gb is either more expensive than macbook (A100) or somewhat tricky (build a separate machine with several old Tesla GPUs, nvlink them, pray that you LLM rig will work).
Considering this, a huge audience who would like to experiment with LLMs will be forced to use Python or pay for hosted GPUs. :(
Some functions in Torchx.Backend create intermediate Iotas and similar tensors. These should be allocated in the same device as the function's input tensor.
The text was updated successfully, but these errors were encountered: