-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What is the configuration of pertraining on hypersim ? #5
Comments
For NYUv2: python main.py \
--tasks semantic normal scene instance orientation \
--enable-panoptic \
--results-basepath /some/path \
--validation-skip 0.95 \
--checkpointing-skip 0.95 \
--checkpointing-metrics valid_semantic_miou bacc panoptic_deeplab_semantic_miou panoptic_all_deeplab_pq panoptic_all_with_gt_deeplab_pq \
--rgb-encoder-backbone resnet34 \
--rgb-encoder-backbone-block nonbottleneck1d \
--depth-encoder-backbone resnet34 \
--depth-encoder-backbone-block nonbottleneck1d \
--encoder-backbone-pretrained-weights-filepath /path/to/our/imagenet/checkpoint.pth \
--input-modalities rgb depth \
--tasks-weighting 1.0 0.25 0.25 2.0 0.0 \
--learning-rate 0.005 \
--dataset hypersim \
--subset-train 0.2 \
--instance-center-heatmap-top-k 128 For SUNRGB-D: python main.py \
--tasks semantic normal scene instance orientation \
--enable-panoptic \
--results-basepath /some/path \
--validation-skip 0.95 \
--checkpointing-skip 0.95 \
--checkpointing-metrics valid_semantic_miou bacc panoptic_deeplab_semantic_miou panoptic_all_deeplab_pq panoptic_all_with_gt_deeplab_pq \
--rgb-encoder-backbone resnet34 \
--rgb-encoder-backbone-block nonbottleneck1d \
--depth-encoder-backbone resnet34 \
--depth-encoder-backbone-block nonbottleneck1d \
--encoder-backbone-pretrained-weights-filepath /path/to/our/imagenet/checkpoint.pth \
--input-modalities rgb depth \
--tasks-weighting 1.0 0.25 0.25 2.0 0.0 \
--learning-rate 0.005 \
--dataset hypersim \
--subset-train 0.3 \
--instance-center-heatmap-top-k 128 |
Thank you so much ! The epoch number had not been set in the command, does this mean that the number of epoch is 500, as set in the args.py ? |
Yes. However, note that the actual number of iterations also depends on the specified subset parameter. Even with a random subset of 0.2 or 0.3 per epoch, training on an A100 will take around one week. |
Thank you ! 🙏 But there is an another problem. That is when I train EMSANet on nyuv2 with the the pretrained weights for the encoder backbone ResNet-34 NBt1D, using the command in the last of Readme file, the test miou is 0.5041. It is different from the paper——0.5097, though I repeated the training process three times. Did I do something wrong ? |
This should not happen. I will run a test training to double-check this. |
Ok, I did some test trainings and was able to almost reproduce the reported results in a more recent environment:
Note that the learning rate is slightly lower than the reported value in the paper: 0.04 (paper) vs 0.03 (here). However, as the environment is different, I enqueued runs with 0.02, 0.03, and 0.04. The best result is shown above. It was at epoch 447 based on Training was done on an A100 40GB with driver 470.63.01. Please find below additional details on the environment.
I hope this helps. |
Epoch, Task weighting or the other settings, can you show the training command on hypersim ?
The text was updated successfully, but these errors were encountered: