-
Notifications
You must be signed in to change notification settings - Fork 326
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] SyncDataCollector Crashes with Resources Leak During Data Collection #2644
Comments
Maybe there is a common/root cause between #2614 and this issue |
Maybe it's related to my disk storage being too small ? I'm storing stacked frames (4, 3, 84, 84) into my replay buffer which uses |
I have ran into a similar problem before. When I was using torchRL with IsaacLab, I would have training runs die midway through when using
|
Is this what you were running into @AlexandreBrown? rl/torchrl/collectors/collectors.py Lines 1083 to 1090 in f5a187d
|
I think either would work! There shouldn't be a case where one would use multiprocessed collectors with IsaacSim. |
I haven't tried the fix but I'm not opposed to having the option. |
Describe the bug
I've observed that lateset trainings crash after 180k steps with the following message :
To Reproduce
Expected behavior
No crash
Screenshots
If applicable, add screenshots to help explain your problem.
System info
output :
Checklist
The text was updated successfully, but these errors were encountered: