openai/
The CLIP model was developed by OpenAI to investigate the robustness of computer vision models. It uses a Vision Transformer architecture and was trained on a large dataset of image-caption pairs. The model shows promise in various computer vision tasks but also has limitations, including difficulties with fine-grained classification and potential biases in certain applications.
You need to login to use this model
LoginSettings
Please upload an image file
dog (0.90)
cat (0.10)
Run models at scale with our fully managed GPU infrastructure, delivering enterprise-grade uptime at the industry's best rates.