Skip to content

CLIP Models¤

Coming Soon

This example is planned for a future release. Check back for updates on CLIP model implementations.

Overview¤

This example will demonstrate:

  • Contrastive Language-Image Pre-training (CLIP)
  • Zero-shot image classification
  • Text-image similarity computation
  • Fine-tuning CLIP for custom domains

Planned Features¤

  • Pre-trained CLIP model loading
  • Custom CLIP training from scratch
  • Image and text encoder architectures
  • Contrastive loss implementation

References¤

  • Radford et al., "Learning Transferable Visual Models From Natural Language Supervision" (2021)