There's three ways that I can think of having distributed training:
- Use of Pytorch's Distributed Training infrastructure. Would require establishing communication protocols specific to the case of Deep RL. This would all be in Python (most likely) unless we find a way around.
- Use of Reverb
- Use TF based Datasets (@threewisemonkeys-as )
- Pytorch wrapper for the conversion of NumPy arrays, etc (that are received) (Short-term, up for grabs)