For downloads and more information, please view on a desktop device.
Description
The CLIP (Contrastive Language-Image Pretraining) model combines vision and language using contrastive learning. It understands images and text together, enabling tasks like image classification and object detection.
Publisher
OpenAI
Latest Version
latest
Modified
August 8, 2023
Size
0 B
AI models generate responses and outputs based on complex algorithms and machine learning techniques, and those responses or outputs may be inaccurate or indecent. By using this Playground, you assume the risk of any harm caused by any response or output of the model.