Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training logic on powerset #46

Open
njallskarp opened this issue Jun 30, 2023 · 0 comments
Open

Training logic on powerset #46

njallskarp opened this issue Jun 30, 2023 · 0 comments
Labels
enhancement New feature or request

Comments

@njallskarp
Copy link
Owner

Training logic for powerset

We have already added a command line argument (main.py) where users can specify the domains or datasets. What we need to do is we need to have each domain load a different Dataset class. Then, during each iteration of for set of sources in the powerset, we need to use torch Dataset's concat method to concatinate the multiple domains to create a single Dataset. If we have N domains then we will end up creating 2^N - 1 dataset classes, one per iteration.

Where this could happen

It seems to me that this might happen inside the run training fuction. That is, around the for ... in range(epochs) there will be something like for domain_subset in powerset:

This means that we will need to pass the Dataset classes into this function, not the dataloaders.

We can schedule a meeting to discuss this in detail.

@njallskarp njallskarp added the enhancement New feature or request label Jun 30, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant