CLIP Embeddings
Medium150 pts0 solves
OpenAI's CLIP was trained on 400M image-text pairs using contrastive learning: matching images with their correct captions.
What training approach does CLIP use?
Flag format: CONGRESS{training_approach_in_snake_case}
Hint
Push matching image-text pairs together, push non-matching pairs apart.