Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hyperparameter Configuration Guidelines for Fine-tuning on Custom Datasets #146

Open
JYDataScience opened this issue Nov 21, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@JYDataScience
Copy link

Is your feature request related to a problem? Please describe.
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]

I'm always frustrated when I try to fine-tune the model with my dataset. I'm currently using Moirai-R models for a feasibility study as part of a demand forecasting model enhancement project. I am finding it challenging to configure various parameters, such as Prediction Time (PDT), Context Length (CTX), Patch Size (PSZ), Batch Size (BSZ), and Test Set Length(TEST), or to set the 'offset', 'eval_length', 'prediction_lengths', 'context_lengths', and 'patch_sizes' parameters in the YAML file located under uni2ts/cli/conf/finetune/val_data. Each of these parameters has a significant impact on the model's performance, and optimizing them remains both challenging and time-consuming. I am particularly facing significant challenges due to the small size of my dataset, which is insufficient for effectively training and testing models. The limited size of the dataset makes it challenging to properly tune and optimize the model, often leading to suboptimal performance and reduced robustness in predictions.

Describe the solution you'd like
A clear and concise description of what you want to happen.

I wonder if it is possible to provide a comprehensive parameter configuration guideline tailored to the characteristics of the dataset, such as its size. I would like to see documentation that can guide users through the process of setting these parameters based on the characteristics of their specific datasets. It can recommend clearly defined parameter ranges and YAML-specific settings. Specific parameter setups could be provided for different scenarios, such as small datasets, large datasets, or models requiring high robustness in predictions. For example, the documentation might include guidance on reducing overfitting with small datasets or improving generalization with limited data points. It would be helpful to include a clear explanation of how each parameter affects the model's performance and predictions. It would be helpful to include a clear explanation of how each parameter affects the model's performance and predictions.

Describe alternatives you've considered
A clear and concise description of any alternative solutions or features you've considered.

As an alternative, I attempted to adjust the parameters manually through trial and error. However, this approach proved to be highly time-consuming and inefficient, as it required extensive iterations to identify optimal configurations, especially with a small dataset. I considered implementing grid search or random search to systematically explore parameter combinations. While these methods can find better configurations, they are computationally expensive and not ideal given the limited size of my dataset. Additionally, I followed the instructions provided in the README file of the GitHub repository to fine-tune a pre-trained model on my custom dataset. I created data configuration files and defined the ranges for the hyperparameters as part of the fine-tuning process. However, setting the appropriate ranges for them was difficult, so the performance did not meet my expectations.

Additional context
Add any other context or screenshots about the feature request here.

There are no additional details or screenshots related to this feature request.

@chenghaoliu89 chenghaoliu89 added the enhancement New feature or request label Dec 4, 2024
@chenghaoliu89
Copy link
Contributor

Hi @JYDataScience, thanks for your suggestion. Our early versions primarily focused on pre-training and evaluation, while the fine-tuning functionality lacked comprehensive experimental testing and detailed tutorials. I found there is a high demand for fine-tuning in the community. We will improve it in the future. Let me know if you have any other suggestion.

@chenghaoliu89 chenghaoliu89 changed the title Hyperparameter Configuration Guidelines for Custom Datasets Hyperparameter Configuration Guidelines for Fine-tuning on Custom Datasets Dec 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants