Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

VRAM is not freed on errors #303

Open
rofoto opened this issue Apr 10, 2024 · 3 comments
Open

VRAM is not freed on errors #303

rofoto opened this issue Apr 10, 2024 · 3 comments
Labels
bug Something isn't working downstream-fix Likely requires fixing another project downstream enhancement New feature or request

Comments

@rofoto
Copy link

rofoto commented Apr 10, 2024

When using musicgen the process completes and all files are created, but I have blocked network traffic out and this causes an error. If you enable multi band after this, there is a chance that there will not be enough vram for it.

Not sure this is intended behavior since there seems to be a delayed cleanup between runs.

This error does not effect the outputs but it also puts the GPU in a state where VRAM is not freed, forcing a restart.

This is not the only error that puts the GPU in this state. It appears that pretty much any error, including but not limited to;
torch.cuda.OutOfMemoryError and errors when trying to download models puts the GPU in this state.

@rsxdalv
Copy link
Owner

rsxdalv commented Apr 11, 2024 via email

@rofoto
Copy link
Author

rofoto commented Apr 12, 2024

this is the only other error I am seeing

"tts-6.0_webui\installer_files\env\lib\asyncio\proactor_events.py", line 165, in _call_connection_lost
    self._sock.shutdown(socket.SHUT_RDWR)
ConnectionResetError: [WinError 10054] An existing connection was forcibly closed by the remote host. 

That's why I thought telemetry was the potential issue.

I think some of them would free it on next load, but that's not ideal.
After looking, it would appear that in some situations the vram is freed on the next run but ideally it can be cleared at the end of generation, just in case.

@rsxdalv
Copy link
Owner

rsxdalv commented Apr 12, 2024 via email

@rsxdalv rsxdalv added bug Something isn't working enhancement New feature or request downstream-fix Likely requires fixing another project downstream labels Sep 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working downstream-fix Likely requires fixing another project downstream enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants