Skip to content
This repository was archived by the owner on Jul 22, 2024. It is now read-only.

Conversation

@ROGERDJQ
Copy link

Here may include a issue when multiGPUs are used.
Since the default self.batch_size=8 at L452, when multiGPUs are used and data shape is more than 8, I found the domain_batched at line452 has actually fewer dimensions than it should be, which leads to the zip error at line 460. (suppose for 3 GPU with train batch size=4, with all domains, line 452 only returns one element, while the other three at line 456 returns 2 elements: one with shape (8,), the other with shape (4,) ) It is noticed that domain_batched is not used afterward. So may be a straightforward way is to delete it.

@songfeng
Copy link
Contributor

@sivasankalpp could you take a look?

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants