-
Notifications
You must be signed in to change notification settings - Fork 384
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issues with HoverAviary: Navigation and Plotting #227
Comments
as per #225 , action spaces that start with |
Dear @JacopoPan, I have already tried it for 10,000,000 steps. The drone does not reach there. Basically, I have shifted the desired location from (0,0,1) to some other point like (2,2,2). It never reached there even once even in 10,000,000 steps. Any suggestion? |
Have you tried to learn to stabilize in 0,0,1 with a 3D action space first? I would make sure that at least that is working for you first. |
Yes, it works for (0,0,1), however, it does not work for an arbitrary point like (2,2,1). |
Can you please attach the terminal outputs (training curves) of those 2 trainings and the sourcecode changes you made to go to (2,2,1)? |
I do not have those training curves. I cleared my work history. I am running the simulation again and I shall soon send you those results. For the time being the code is attached herewith. Could you please take a look what could possible go wrong? Thank you so much for your help. Truly grateful to you. |
can you share the code changes as a PR please? |
My Linux was broken and i just reinstalled it. I do not have git diff. Also, pull request was causing problem for me on you repo. So, I created a separate repo citing your main repo. Just to keep my own version. If you think it is inappropriate, please let me know and I shall take it down. I did not make big changes in the I am not sure but I guess there is a logical problem in
|
No worries with having your own version but it makes it a bit difficult for me to tell what might be fixed/changed from the original one. Can you past the modified reward here? |
Thank you so much @JacopoPan for your understanding and help. Here is the reward function:
Basically, trying to penalise the agent if it is moving away from the goal, moving too fast, going out of the working boudary (flight arena), tilting too much or taking too long time to reach to the goal. pretty standard things. I think the main hurdle is |
Hello @abdul-mannan-khan I would definitely try starting from the 3D ( |
Hello @JacopoPan , Thank you so much for your reply. I have tested |
A few points:
|
Thank you so much for highlighting it. I am particularly intersted in finding out more about the following:
Can you please put some more light on these points? I am sure that the solution is there in 2-layers and not in one layer. Should I write two RL algorithms one for low level control and one for high level control? For example, one control (MPC/RL) only able to track the given trajecctory and the second controller (RL-based) able to search for the trajectroy to reach the goal? Like one RL for trajectory planning and the second one for tracking it? Thank you so much for your response. |
In essence, if you "only" specify a reward based on a far away target position, you are asking the RL agent (e.g. PPO) to figure out what in, a conventional pipeline, would be a stabilizing controller (to hover in place), a position controller (to perturb the equilibrium towards a desired reference point), and a trajectory generator (to generate a feasible sequence of poses towards your goal), all of which are affected by the dynamical qualities of the quadcopter itself. |
Dear @JacopoPan, Thank you so much for your kind response. Following your comment, if I would like to change my action type to I have already its definition in
assuming that my action creating this motion will be followed by a local MPC. Where can I define this action? I have already defined it in
I am confused about defining it in I have made following changes in
where can I define it? I tried to work on Thank you so much for your kind response. |
|
Thank you so much for your response @JacopoPan. I believe that we do not need an RL algo for a dynamic control. MPC can do a pretty good job and there is no value addition in removing it but this is just my view. You can disagree with it obviously. I am tying to understand what you mean by
Do you mean that I need to restrict it in
If you could point me some directory which would help me solve the problem I am trying to do, it would be a great favour. Thank you so much for your time and efforts. |
yes, all the changes should be local to In practice, because you have 6 actions instead of over a continuous 3D |
Hi everyone,
I'm working with the HoverAviary in gym_pybullet_drones and facing two main issues:
Navigation to Target: The drone only moves vertically and doesn't navigate to the set (x, y, z) coordinates. Does anyone have tips on enabling horizontal movement toward a target? For reference, I have attached the video response here.
crazyflie_video22.mp4
Plotting Problems: I'm also struggling with plotting the simulation results. If anyone has suggestions on how to effectively set up plots in this environment particularly for the
main
branch, it would be greatly appreciated.Thanks for any help you can provide!
The text was updated successfully, but these errors were encountered: