You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! I noticed that you’re working with LongViLa-LLama3-1024Frames. I’m also trying to run inference with long context but am encountering issues with multi-GPU usage—my model only runs on a single GPU. Have you found a way to successfully utilize multiple GPUs for long-context inference? Any insights or suggestions would be greatly appreciated!
LongViLa-LLama3-1024Frames output is often repetitive. Why does this happen, and are there any suggestions to reduce the repetition?
The text was updated successfully, but these errors were encountered: