Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FAQ] Option to utilize CPU (or at least to control GPU RAM usage) and to offload finished models from GPU to CPU #162

Open
LordGedelicious opened this issue Nov 28, 2024 · 1 comment
Labels
question Further information is requested

Comments

@LordGedelicious
Copy link

I'm currently running a HPO on my binary classification model using randomized search CV. The first split, it takes 38.1 GB out of the available 40.1 GB from my A100 GPU (provided by Google Colab). However, when the code goes to train the model using the second split, I notice that the GPU RAM consumption never decreased so I was hit with a CUDA Out of Memory error.

I notice that there's no parameters to set the model to be trained/loaded into cpu, gpu, or cuda like XGBoost's library. Is CUDA the default? Is there any way to offload the model from CUDA/GPU to CPU?

@LordGedelicious LordGedelicious added the question Further information is requested label Nov 28, 2024
@AnFreTh
Copy link
Collaborator

AnFreTh commented Nov 28, 2024

Hi,
there is a trainer_kwargs argument in the fit method where one can specify the device the model is being trained on.
We will also release a new version in the coming days with built in HPO possibilities and a more efficient Mamba Version then the current pytorch one.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants