-
Notifications
You must be signed in to change notification settings - Fork 53
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference takes large GPU memory #1
Comments
Same problem.. I have been trying to run it on a 16 GB GPU with no luck. the models themselves aren't that large so I am very confused why they take that much space. |
@plusgrey We're looking into this and will get back to you soon. @AmrSheta22 could you share the necessary changes? I will update the repo instructions. Thanks! |
We suspect the reason is we are using default transformer layers without flash attention or some other memory-efficient implementations. The reason is that we were not able to compile these implementations, e.g., xFormers, under google's code base due to conflicts. This can be easily solved by using the original DINOv2 implementation. |
For me the same, it basically takes the whole GPU, in my case almost 11GB this is how I could reduce it
|
@paolovic Hi, can you tell me in more detail how you solved it? |
Hi,
But I was not able to enable GPU acceleration (that's why the 0% GPU-Util), although CUDA and cuDNN are installed and recognized. |
Hi,
Thanks for your great work. However, when I run your demo on an RTX 3090 (24GB) it takes about 22GB of GPU memory. Is this normal? Compared to LoFTR, which only consumes about 1.7GB, it is really surprising.
The text was updated successfully, but these errors were encountered: