-
Notifications
You must be signed in to change notification settings - Fork 20
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Different image inferences with same result #16
Comments
I was testing in a jupyter notebook and encountered similar issuse as yours before. I found that's becauset my GPU CUDA memory was not enough. Note my image size is about 1024p. When I was using larger image like 4032p it giave me results with all 0.5. |
I did not test on an image beyond 2K. But I guess you need to resize the input to avoid too large resolution, e.g., beyond 2K. The main reason, I guess, is that CLIP is only trained on 224x224 and too large image resolutions lead to diverse receptive fields for the network, which should affect the performance of the network. |
Hi IceClear, |
We resized the images in SPAQ. You can find it in our paper. #10 |
Hi,
Im doing different tests with
demo/clipiqa_single_image_demo.py
and theattribute_list = ['Quality', 'Brightness', 'Sharpness', 'Noisiness', 'Colorfulness', 'Contrast']
.First, I’ve seen that fitting a good size to the input image is essential because in some cases the result is NaN. Is there a fixed size that should be adjusted in the inbound image?
In the tests I’m doing, regardless of the resize, I see that it doesn’t matter the input image the result is always the same, you know what might be?
Example:
The text was updated successfully, but these errors were encountered: