Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Could you share distill log for each dataset ? #30

Open
JarveeLee opened this issue May 31, 2023 · 5 comments
Open

Could you share distill log for each dataset ? #30

JarveeLee opened this issue May 31, 2023 · 5 comments

Comments

@JarveeLee
Copy link

As title, for nuScenes, ScanNet, Matterport, it would be great if you can share the logs of distilling, which can help me to debug the training process ...

@pengsongyou
Copy link
Owner

Sorry I did not take the logs out from my internship, so I don't really have it anymore. Maybe you can specify what kind of issues you are encountering?

@JarveeLee
Copy link
Author

JarveeLee commented Jun 1, 2023

I am distilling on my own data.
Occasionally I get 'WARNING:tensorboardX.x2num:NaN or Inf found in input tensor.
' then the loss is Nan as well. Restart then NaN would disappear ...

I have no idea how to explore this bug ..... It would be great if you can suggest me to inspect something inside my data. I just replace the data_dir in config, while the other remains the same as nuScenes.
image

@pengsongyou
Copy link
Owner

pengsongyou commented Jun 1, 2023 via email

@Outlying3720
Copy link

I meet the same problem too when I run distill on Matterport or ScanNet datasets.

NaN or Inf found in input

2023-07-06_193228

Could anyone have some solution?

I try to add gradient clip to max 1 or lower the lr. But still no help.

What's more, I found it will be overfitting and val loss will be NaN. But the training process will not be interrupted.
2023-07-06_193457

I run code on Pytorch 2.0 with 4090 GPU. Maybe the Pytorch upgrade caused this problem?

@pengsongyou
Copy link
Owner

pengsongyou commented Jul 7, 2023

@wengmingao I will investigate this from my end and share the information to you later.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants