-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bad training result after 90 epochs #15
Comments
@ArashJavan sorry for disturbing, if you did not know the reason, could you please give me your training hyper parameters. |
@rginjapan your sequence is not correct for an RNN-based architecture. The combination should be:
Also, I would first try with only one Network for example only IMU to see if everything works fine. Then switch to more complex networks with both lidar and imu fused! Another hint is about your grad-norm, if the above points did not help try to use gradient-clipping. |
@ArashJavan Thanks for your reply, I think "sequences" means images numbers, for example, sequences:3 represents 3 continuous frames I0, I1 and I2. I remember you explain in other close issue, "combinations [0,2]" will increase the relative motion compared with "combination [0,1]" , which is good for training. Am I correct ? "Another hint is about your grad-norm, if the above points did not help try to use gradient-clipping." Sorry, can you explain in more details, I did not understand. |
@ArashJavan Thanks for your quick reply. Right now I understand, I already used [0,1],[0,2],[1,2] to get good result in only lidar network, so then I thought it will work on lidar+imu network, but not... So, do you evaluate the difference between large sequence number like 5 in your config and samll number like 2? Since [0,1] and [4,5] is similar in training, right? |
@rginjapan plz look at the loss function here. As you can see in the loss function we already take care for global motion since the loss is calculated for local - e.g. [x_i, x_i+1] and global -e.g. [x0, x1],...,[x0, x_seq]. |
@rginjapan Can you share your config-file and your meta-data settings (lr, wd and so on). |
config.yaml datasets: kitti:
DeepLIO Networkdeeplio: Lidar Feature Netowrkslidar-feat-pointseg: # pointseg feature lidar-feat-flownet: lidar-feat-resnet: lidar-feat-simple-1: imu-feat-fc: # FC imu-feat-rnn: # RNN fusion-layer-soft: Odometry Feature Netowrksodometry feature network with fully connected layersodom-feat-fc: odometry feature network with rnn-layersodom-feat-rnn: Loss Configurationslosses: current-dataset: 'kitti' hyper params in train.py parser.add_argument('-j', '--workers', default=4, type=int, metavar='N', |
Could you please upload your trained model to let me try to get some good result, it is not to be as good as your paper's result. P.S. I increased the epochs to 150, the loss can not be decreased, still around 16. And also the result is still bad... Looking forward to your guidance. |
@huazai665 I cannot reproduce the result in the paper in the end. |
@rginjapan do you revise the liegroups library to use cuda? |
@rginjapan Hi, have you solve the worse result issue? I think I meet the same question with you. If you are doing this work or some fantastic works in this filed, meybe we can exchange contact information. |
Due to my GPU limitation, parameters are:
sequence-size: 3 # must be >= 1
combinations: [ [ 0, 1 ],[0,2],[1,2] ]
batch size:2
epochs:90
I trained the lidar with imu network, using all the training dataset in original config. I also plot the loss curve and test result in 2011-10-03: [27] for you to check.
What wrong with my implementation?
Looking forward to your reply, thanks ahead!
The text was updated successfully, but these errors were encountered: