CLIP Embeddings

Medium150 pts0 solves
OpenAI's CLIP was trained on 400M image-text pairs using contrastive learning: matching images with their correct captions. What training approach does CLIP use? Flag format: CONGRESS{training_approach_in_snake_case}
Hint
Push matching image-text pairs together, push non-matching pairs apart.