Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for Video with Qwen2-VL #75

Open
tmoroney opened this issue Oct 2, 2024 · 9 comments · May be fixed by #97
Open

Support for Video with Qwen2-VL #75

tmoroney opened this issue Oct 2, 2024 · 9 comments · May be fixed by #97

Comments

@tmoroney
Copy link

tmoroney commented Oct 2, 2024

It would be really great if support for Video could be added to Qwen2-VL as it seems to only support images at the moment. I am working on a project that would seriously benefit from it.

@Blaizzy
Copy link
Owner

Blaizzy commented Oct 2, 2024

Hey @tmoroney

Indeed, it's a cool feature of Qwen2-VL. I will work on adding it after #41 and Molmo port.

Could you share a little bit more about your project?

@Blaizzy
Copy link
Owner

Blaizzy commented Oct 3, 2024

I got multi-image support working which is pretty close to video.

From here to video it’s pretty close but I want to make the API seamless for all models, that’s why it will take a bit longer.

https://x.com/Prince_Canuma/status/1841634911825858978

@tmoroney
Copy link
Author

tmoroney commented Oct 3, 2024

Hey @tmoroney

Indeed, it's a cool feature of Qwen2-VL. I will work on adding it after #41 and Molmo port.

Could you share a little bit more about your project?

Essentially the project is Copilot for video editors. As part of my masters in computer science I am researching how to create an AI video editing assistant that will suggest the next shot as you edit based on the context of the story so far and the emotional tone etc, along with providing inspiration to remove creative blockers. I want to analyse all of the project's footage on device using small and efficient ML models, along with algorithms (Computer vision, sentiment analyses etc) whenever possible in order to reduce compute.

@tmoroney
Copy link
Author

tmoroney commented Oct 3, 2024

I got multi-image support working which is pretty close to video.

From here to video it’s pretty close but I want to make the API seamless for all models, that’s why it will take a bit longer.

https://x.com/Prince_Canuma/status/1841634911825858978

Amazing! Thanks for all your hard work :)

This was referenced Oct 16, 2024
@anishjain123
Copy link

@Blaizzy any update on the video support? Love the work you've been doing!

@Blaizzy
Copy link
Owner

Blaizzy commented Nov 22, 2024

Thanks guys!

There is a PR for video support #97 it works but needs a bit of polishing

I will do that and merge it over the weekend.

@anishjain123
Copy link

you're a g @Blaizzy , its insane how much the memory usage gets on this since it parses it as an array of images. Im wondering if this is the right architecture for on device video processing

@Blaizzy
Copy link
Owner

Blaizzy commented Nov 27, 2024

Could you elaborate?

@andimarafioti
Copy link
Contributor

SmolVLM should really help with that!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

4 participants