openai/clip-vit-base-patch32 cover image

openai/clip-vit-base-patch32

The CLIP model was developed by OpenAI to investigate the robustness of computer vision models. It uses a Vision Transformer architecture and was trained on a large dataset of image-caption pairs. The model shows promise in various computer vision tasks but also has limitations, including difficulties with fine-grained classification and potential biases in certain applications.

The CLIP model was developed by OpenAI to investigate the robustness of computer vision models. It uses a Vision Transformer architecture and was trained on a large dataset of image-caption pairs. The model shows promise in various computer vision tasks but also has limitations, including difficulties with fine-grained classification and potential biases in certain applications.

Public
$0.0005/sec

Input

Please upload an image file

list of labels to guess from. Please separate items by comma

You need to login to use this model

Output

dog (0.90)

cat (0.10)

 


© 2023 Deep Infra. All rights reserved.

Discord Logo