-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The results in the inference mode is relatively inaccurate. #28
Comments
The above is the evaluation code I used, and I am not sure if it is correct. Below are my prediction results, and the differences are very significant.
|
I wonder if it is because I used the GT as the source for reference points and query embedding during the training and validation phases, but could only use the backbone output during inference. |
Hi Fusica, for which dataset are you testing this? And which pre-trained weights are you using? Best, |
i use lmo to run inference, the pretrained model is poet_lmo_maskrcnn.pth. |
You are not doing anything wrong. It was already reported that PoET on the LM-O dataset is not performing as expected. We have a fix for this, but we need to wait for the results of a review process to put out the code and the new model. I am sorry for that. Best, |
However, why can't I achieve at least consistent results with the validation set during inference? Because I found that during validation and inference, the prediction results for the same image are different. So I suspect it might be because the bbox_mode is different during validation and inference. |
That is 100% the reason. I kindly refer to the supplementary material of our paper: https://www.aau.at/wp-content/uploads/2022/09/jantos_poet.pdf We investigated how the object detector quality influences the pose estimators performance. Hence, the difference can be quite drastic when using actual network predictions instead of GT information. Best, |
I loaded the pre-trained weights you provided and used inference to test an image from the training data, but I found that the returned result has a significant gap compared to the ground truth in the dataset. I would like to know if this gap is normal or if there might be an issue somewhere, such as in my evaluation process. Do the authors have performance test results? I hope the authors can provide clarification. Thank you very much.
The text was updated successfully, but these errors were encountered: