-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TreeGen: A Tree-Based Transformer Architecture for Code Generation #39
Comments
Hi! Maybe I can help. What are the suggested next steps to have this work @HiroakiMikami? |
Hi @HiroakiMikami apologies for the spam. Let me know if there is something specific that you suggest to look into to help with having pytorch TreeGen working! |
Hi @HiroakiMikami apologies for the spam! Let me know if you do want help. |
Sorry for the late reply. I do not value paper reproduction (such as TreeGen) now for 2 reasons. The first reason is that using a large model with NLP technique may be a more effective and simpler approach than AST-based program generation (e.g., http://arxiv.org/abs/2105.09938v1). And the second reason is that I found out that the quality and characteristics of dataset are more critical than model differences for many usecases. So, the status of these paper-reproduction issues is pending. Thank you for your comments! |
Hi Hiroaki, I appreciate your response. I was wondering if you could clarify what you meant by that. Do you mean that all models perform relatively similar across all data sets? I wasn't quite sure how to interpret your response. I am also curious, what type of model would you prefer then? NLP based ones? Grammar syntax based ones? Or something else? thanks for your time again! |
Another question, our of curiosity, if you do not value reproduction of papers anymore what do you value? Related, does this make your mlprogram repo obsolete for you (besides the fact it makes it easy for you to run experiments)? Thanks for your time! Its appreciated. |
That's almost right. I think all models show a similar performance if the used computational resource (e.g., model parameters, FLOPs, and training epochs) is almost the same. Also, I think that dataset quality (e.g. the number of annotation mistakes, the code quality) is important. The performance of program synthesis may be limited by the quality of the dataset, not DNN model structures.
I think NLP based one is enough for program synthesis. Grammer syntax based models reduce the syntax errors of the outputs. However, the inference procedure is very complex and cannot utilize GPU efficiently. So using NLP based models and filtering out the code with invalid syntax (like CodeXGLUE baselines) may be more efficient than using grammer syntax models.
I made and maintained this repository in order to make my experiments easy, so the purpose of mlprogram is not changed. But I should use transformers module as a model zoo. |
The text was updated successfully, but these errors were encountered: