-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error Running cogvlm model on Self-Hosted GPU Server with Roboflow Inference (Transformer Version) #355
Comments
Hi, Thanks for raising the issue. It is likely that you are right about the source of the bug. Will take a look next week to see if downgrade helps and if it does, we will fix the problem given it does not create another. |
I think the version of the |
ok, checked that this fix work on my end: We need to ship this with the next release, but for the time being you can build docker image on your end: git clone git@github.com:roboflow/inference.git
cd inference
docker build --build-arg="TARGETPLATFORM=linux/amd64" -t roboflow/roboflow-inference-server-gpu:dev -f docker/dockerfiles/Dockerfile.onnx.gpu . To run the server: docker run --gpus all roboflow/roboflow-inference-server-gpu:dev |
I have the same issue as @YoungjaeDev and @PawelPeczek-Roboflow suggestion above got me the same output, but with an additional error about WithFixedSizeCache.
|
Hi @BChip, thank you for running the test, I have pushed small change to this PR and then followed your test steps, I see no error and additionally I can confirm transformers version is now bound to |
@grzegorz-roboflow Awesome, just tried it and it works! When will this be mainstream fixed? Thank you!!!!! |
@BChip - that will be shipped to dockerhub with next release - which I believe would be done as soon as we close and test this PR: #343 which consumes big part of our time and capacity now. |
Search before asking
Bug
I'm encountering an issue while attempting to deploy the cogvlm model on my own GPU server using Roboflow inference code. The server setup seems to be correct, but when I try to run the model, I run into the following error:
Upon further investigation and based on this GitHub issue (THUDM/CogVLM#396), it's recommended to downgrade the transformers library to version 4.37 due to compatibility issues. However, the current deployment is using version 4.38. Could you please confirm if the transformers version could be the source of this issue and if downgrading would be appropriate? Any other insights or suggestions would also be greatly appreciated.
Thank you!
Environment
inference 0.9.20
inference-cli 0.9.20
inference-gpu 0.9.20
inference-sdk 0.9.20
x86-gpu(rtx3090)
Minimal Reproducible Example
Additional
No response
Are you willing to submit a PR?
The text was updated successfully, but these errors were encountered: