Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the switch between stubborn and peanut #7

Open
kellyiss opened this issue Feb 23, 2024 · 9 comments
Open

About the switch between stubborn and peanut #7

kellyiss opened this issue Feb 23, 2024 · 9 comments

Comments

@kellyiss
Copy link

Hi, I see there is implementation for switching between corner goal and prediction goal, but seems that the switch step is set to 0.
Does peanut just rely on the prediction model for deciding the long-term goal at a frequency?

Thanks for your interesting and inspiring work!

@ajzhai
Copy link
Owner

ajzhai commented Feb 23, 2024

That's right, it always uses the prediction. We thought that it might be interesting to switch between the two policies, but we found it doesn't really help. Maybe it is worth exploring further though!

@kellyiss
Copy link
Author

Thanks for the reply.
I have one more question, I run the evaluation and find that the evaluation can stop at a certain step (< 500 steps) for some unsuccessful episodes. How do you implement it?

@ajzhai
Copy link
Owner

ajzhai commented Feb 26, 2024

Hmm it should not stop. Can you show the error message?

@kellyiss
Copy link
Author

1709006883142

@ajzhai
Copy link
Owner

ajzhai commented Feb 27, 2024

Oh that's a normal failure case, it will happen if there is a false positive detection and the agent calls "stop" as it thinks it reached the target.

@kellyiss
Copy link
Author

Yeah, I see.
And for the prediction model, seems the map size for training is 960 * 960, but during inference, the map size is 720*720.
Am I right?
Moreover, I find that only the target prediction inside the local map is used for update the global goal. Is it reasonable to train the prediction model only with the local maps?
Thanks for your patience.

@ajzhai
Copy link
Owner

ajzhai commented Feb 28, 2024

Currently it is set that way, but maybe that is actually a mistake -- if you try 960 during inference, let me know what happens!

It is better to predict using global map than only using local map. See the ablation study in our paper for comparison

@kellyiss
Copy link
Author

I thought the performance might not be significantly impacted. But actually, a 1% decrease in success rate on the validation set when using the map size 960, which is perplexing to me.

@ajzhai
Copy link
Owner

ajzhai commented Mar 17, 2024

I believe only very few episodes need the agent to go outside of the 720 size map, so there is not much benefit to predicting out there. And then the new predictions causes some slightly different behavior sometimes which leads to 1% decrease in success (not really significant).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants