-
Notifications
You must be signed in to change notification settings - Fork 521
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Loading and using open_vision model #747
Comments
Please paste the output log of the terminal here. |
the output log of the terminal
Is this the problem? I have been trying to solve it for a long time but failed. Maybe it is because my professional ability is not solid. Before this, I looked at other open_vision related issues, which helped me a lot. Could it be that the operating environment is still not configured correctly? Please give me more advice |
A1: Please update the source code to v2.5.0 or higher.
A2: Please follow the instructions to install the correct onnxruntime version, ensuring it is compatible with your local CUDA version.
A3: Please refer to the original repository to install the necessary packages. Remember to compile the operator:
|
All right. You shouldn't execute
Happy labeling! 🚀 |
Thanks, I will continue to try. I strongly recommend that you create a more complete help document for the deployment of open_vision to help more people(or beginners). Thank you very much and wish you a happy life! |
Hey @FightingKai01, You'll need to handle the environment setup on your own. I suggest trying WSL2; it's generally easier for beginners. You can also research how to compile on Windows, and this might give you some guidance. |
Glad I didn't give up and I got it working in WSL. Thanks Thanks Thanks🚀🚀🚀 🚀My suggestion Hahahaha, communicating with you is really rewarding. |
Woo... 😃 I'm so glad to hear about your success and growth! It's wonderful to see how you've been able to learn and develop new skills. Your enthusiasm is truly infectious! 🚀 Regarding T-Rex2 - yes! I'm know about it. While it's indeed an impressive model with powerful capabilities for automatic annotation using visual and language prompts. Unfortuanatelly 😞 , it's currently only available through commercial API services rather than being open-source. I understand your interest in its flexible interactive features and the potential for online learning adaptation. That said, I share your hope for the open-source community to develop even more powerful foundation models. The collaborative nature of open-source development has already given us many remarkable tools, and I'm confident we'll continue to see innovative contributions in this space. Keep exploring and learning - your curiosity and engagement with these technologies is truly inspiring! Looking forward to hearing more about your discoveries and experiences. 🙏 |
I read the relevant documents again, but I am not yet capable enough to study the underlying code of open_vision. I have some questions. Of course, you can answer selectively. I am a beginner and I am not sure whether I should ask you questions.
|
Hey there! @FightingKai01: Thank you for waiting! Let me answer your questions one by one. 😊
A1: You can refer to this guide for fine-tuning
A2: You can think of
A3: Essentially,
A4: No, Hope it helps, if you need more clarification feel free to ask! 🚀 |
About A3: I'm in the process of moving from application to research code and hope to continue communicating. |
|
Search before asking
Question
I have correctly loaded the open_vision model, but found a problem during the actual reasoning process.
The model cannot infer other similar objects under visual prompt, which is very different from the video demonstration you gave.
By reading the relevant documents, I may know the problem, but due to my limited professional ability. Observing the open_vision.yaml , it can be seen that the problem may be that the CountGD model is not loaded.
Could you provide a detailed reference document to help me or others complete the Text-Visual Prompting Grounding project?
Looking forward to your guidance.
Additional
No response
The text was updated successfully, but these errors were encountered: