-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ObjectNav results on HM3D (test-standard) #3
Comments
Yea, the ablation study in our paper is on 500 episodes from HM3D val set. |
Ah, so all the results in Table 1(include other methods, such as ProcTHOR) were evaluate on 500 episodes? |
No, that evaluation is using the standard test set on EvalAI: here |
The results in Table 1 were evaluated online? So the code in nav/collect.py is for ablation? I would like to know whether the code has an evaluation part for the HM3D test standard. |
On this website (https://github.com/matterport/habitat-matterport-3dresearch), I really don't find the test-standard dataset of HM3D. There are only train and val. Could you please tell me where to download the test split dataset? |
We cannot download the test dataset. It is intentionally kept hidden from the public so that people cannot optimize their agent's based on test set performance. This is common for many ML benchmarks. The only way to evaluate test set performance is by submitting your docker image to EvalAI. If you want to do that, a few steps need to be done first:
Note that you should ideally only evaluate the test set performance to report metrics for a paper, not for improving your agent. Use the train and val data to optimize your agent. |
Thanks for your reply! I'll follow your guidance to try to get the metrics! |
Today I tried to do this, but I found that the link to the test-standard summit page was not working.(orz). |
I think you need an account first. Can you make an EvalAI account and try again? |
Yes, I got that page after I logged in. Maybe you can get in because you have participated in that habitat2022. Can you browse the habitat2021 submit page? |
Yes, you can click on the "Participate" button to register a team. |
Hmm that's weird, I thought it was available at any time. You can try asking the Habitat people about that, they can probably help. If there's no solution, then I guess you'll have to use the val split. |
Good, I'll try. Thanks for your help! |
Hello! I'm so sorry to bother you again that I would like to know if the performance of PEANUT on HM3D is test on 500 episodes or on all 2000 episodes? Cause now I use my reproduced checkpoint to obtain a comparable performance when I run 500 episodes, but on all 2000 episodes its performance is lower than before.
The text was updated successfully, but these errors were encountered: