Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CLIP score improvements #2978

Open
wants to merge 29 commits into
base: master
Choose a base branch
from
Open

CLIP score improvements #2978

wants to merge 29 commits into from

Conversation

SkafteNicki
Copy link
Member

@SkafteNicki SkafteNicki commented Feb 27, 2025

What does this PR do?

Fixes #2935
Fixes #2906
Fixes #2883

Add the possibility for more multimodel models that supports longer captions than the original clip model.

Before submitting
  • Was this discussed/agreed via a Github issue? (no need for typos and docs improvements)
  • Did you read the contributor guideline, Pull Request section?
  • Did you make sure to update the docs?
  • Did you write any new necessary tests?
PR review

Anyone in the community is free to review the PR once the tests have passed.
If we didn't discuss your PR in Github issues there's a high chance it will not be merged.

Did you have fun?

Make sure you had fun coding 🙃


📚 Documentation preview 📚: https://torchmetrics--2978.org.readthedocs.build/en/2978/

@SkafteNicki SkafteNicki added the enhancement New feature or request label Feb 27, 2025
@SkafteNicki SkafteNicki added this to the v1.6.0 milestone Feb 27, 2025
@SkafteNicki SkafteNicki self-assigned this Feb 27, 2025
@SkafteNicki SkafteNicki modified the milestones: v1.6.0, v1.7.0 Feb 28, 2025
@SkafteNicki SkafteNicki marked this pull request as ready for review March 3, 2025 12:34
@mergify mergify bot added the has conflicts label Mar 3, 2025
Copy link

codecov bot commented Mar 4, 2025

Codecov Report

Attention: Patch coverage is 21.05263% with 30 lines in your changes missing coverage. Please review.

Project coverage is 69%. Comparing base (34fca16) to head (34bfe8a).

Additional details and impacted files
@@          Coverage Diff           @@
##           master   #2978   +/-   ##
======================================
- Coverage      69%     69%   -0%     
======================================
  Files         338     338           
  Lines       19436   19464   +28     
======================================
+ Hits        13480   13486    +6     
- Misses       5956    5978   +22     
🚀 New features to boost your workflow:
  • Test Analytics: Detect flaky tests, report on failures, and find test suite problems.

@mergify mergify bot added the ready label Mar 4, 2025
@mergify mergify bot removed the ready label Mar 5, 2025
@Borda
Copy link
Member

Borda commented Mar 5, 2025

seems like just the docs are failing with warnings?

@mergify mergify bot added the ready label Mar 5, 2025
@SkafteNicki
Copy link
Member Author

@Borda no there are also failing gpu tests (the test for clip score only runs on gpu, else it would take too long on cpu)

@Borda
Copy link
Member

Borda commented Mar 5, 2025

no there are also failing gpu tests (the test for clip score only runs on gpu, else it would take too long on cpu)

GPU testing is fine :)

>           raise ValueError(f"pic should not have > 4 channels. Got {pic.shape[-1]} channels.")
E           ValueError: pic should not have > 4 channels. Got 64 channels.

Let's try to pass it as PIL image instead?
Or it has reversed order of dimensions width x height x channels?

@SkafteNicki
Copy link
Member Author

@Borda is there a way to get more info from azure? it is hard to debug without what is going wrong

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
3 participants