Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unable to reproduce the results of sparse conv on modelnet40 #38

Open
bollossom opened this issue Dec 9, 2024 · 3 comments
Open

Unable to reproduce the results of sparse conv on modelnet40 #38

bollossom opened this issue Dec 9, 2024 · 3 comments

Comments

@bollossom
Copy link

The result of training for 200 epochs using 8 A800 GPUs on the ModelNet40 dataset is only 80.06% accuracy. All other parameters are the same as those in the train.yaml configuration file.

@eliphatfs
Copy link
Collaborator

I think the result is reasonable as this is only differing by 3% to our checkpoint. There are still many random factors leaving behind, like data sampling. Also, I think we used fewer GPUs so the effective batch size differ, which might lead to this discrepancy. You will need to tune the batch size or the learning rate schedule for better performance at 8 GPUs. If you kept many checkpoints you can try a self-ensemble which reduces these effects.

@bollossom
Copy link
Author

bollossom commented Dec 14, 2024

Thanks for your reply. Could you tell me some other training hyper parameters, such as the batch size of the learning rate and whether to warm up? I think 1000 epochs is too much, and whether to experiment with cosine decay?

@bollossom
Copy link
Author

I think I have figured out the reason. I chose the 5.3M parameter version of the SparseConv model, which achieved an 80.4% result on ModelNet40. The original paper achieved an 83.4% zero-shot effect on ModelNet40, which likely used the model with 41.3M parameters.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants