Results for "zero-shot-image-classification"

100 matches found.

openai

openai/clip-vit-base-patch32

The CLIP model was developed by researchers at OpenAI to learn about what contributes to robustness in computer vision tasks. The model was ...

🎯 zero-shot-image-classification 20,032,440
openai

openai/clip-vit-large-patch14

The CLIP model was developed by researchers at OpenAI to learn about what contributes to robustness in computer vision tasks. The model was ...

🎯 zero-shot-image-classification 6,880,423
openai

openai/clip-vit-large-patch14-336

More information needed...

🎯 zero-shot-image-classification 5,190,264
laion

laion/CLIP-ViT-B-32-DataComp.XL-s13B-b90K

No description available.

🎯 zero-shot-image-classification 4,127,343
laion

laion/CLIP-ViT-B-32-laion2B-s34B-b79K

No description available.

🎯 zero-shot-image-classification 2,953,106
patrickjohncyh

patrickjohncyh/fashion-clip

UPDATE (10/03/23): We have updated the model! We found that laion/CLIP-ViT-B-32-laion2B-s34B-b79K checkpoint (thanks Bin!) worked better tha...

🎯 zero-shot-image-classification 2,818,619
openai

openai/clip-vit-base-patch16

The CLIP model was developed by researchers at OpenAI to learn about what contributes to robustness in computer vision tasks. The model was ...

🎯 zero-shot-image-classification 2,646,498
google

google/siglip-so400m-patch14-384

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 2,622,346
google

google/siglip-base-patch16-224

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 1,191,101
google

google/siglip2-so400m-patch16-naflex

No description available.

🎯 zero-shot-image-classification 856,896
google

google/siglip2-base-patch16-naflex

No description available.

🎯 zero-shot-image-classification 835,970
Marqo

Marqo/marqo-fashionSigLIP

No description available.

🎯 zero-shot-image-classification 773,326
laion

laion/CLIP-ViT-H-14-laion2B-s32B-b79K

No description available.

🎯 zero-shot-image-classification 692,291
laion

laion/CLIP-ViT-L-14-laion2B-s32B-b82K

No description available.

🎯 zero-shot-image-classification 675,197
laion

laion/CLIP-convnext_base_w-laion2B-s13B-b82K-augreg

No description available.

🎯 zero-shot-image-classification 654,740
microsoft

microsoft/BiomedCLIP-PubMedBERT_256-vit_base_patch16_224

- clip - biology - medical libraryname: openclip - src: https://huggingface.co/microsoft/BiomedCLIP-PubMedBERT256-vitbasepatch16224/resolve/...

🎯 zero-shot-image-classification 606,374
google

google/siglip2-so400m-patch14-384

No description available.

🎯 zero-shot-image-classification 534,866
facebook

facebook/PE-Core-L14-336

[\[πŸ“ƒ Tech Report\]](https://arxiv.org/abs/2504.13181) [\[πŸ“‚ Github\]](https://github.com/facebookresearch/perceptionmodels/) Perception Encod...

🎯 zero-shot-image-classification 444,447
google

google/siglip2-base-patch16-224

No description available.

🎯 zero-shot-image-classification 436,683
timm

timm/ViT-B-16-SigLIP-i18n-256

Model Type: Contrastive Image-Text, Zero-Shot Image Classification. - Original: https://github.com/google-research/bigvision - Dataset: WebL...

🎯 zero-shot-image-classification 371,588
timm

timm/MobileCLIP2-S3-OpenCLIP

These weights and model card are adapted from the original Apple model at https://huggingface.co/apple/MobileCLIP2-S3. This version uses can...

🎯 zero-shot-image-classification 354,898
laion

laion/CLIP-convnext_large_d_320.laion2B-s29B-b131K-ft-soup

No description available.

🎯 zero-shot-image-classification 297,506
wkcn

wkcn/TinyCLIP-ViT-8M-16-Text-3M-YFCC15M

No description available.

🎯 zero-shot-image-classification 274,126
yuvalkirstain

yuvalkirstain/PickScore_v1

No description available.

🎯 zero-shot-image-classification 235,324
q-future

q-future/one-align

No description available.

🎯 zero-shot-image-classification 193,877
laion

laion/CLIP-ViT-B-16-laion2B-s34B-b88K

No description available.

🎯 zero-shot-image-classification 186,024
google

google/siglip2-so400m-patch16-384

No description available.

🎯 zero-shot-image-classification 184,160
timm

timm/vit_base_patch16_plus_clip_240.laion400m_e31

dataset: LAION-400M...

🎯 zero-shot-image-classification 178,371
timm

timm/ViT-B-16-SigLIP2-256

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 170,734
google

google/siglip-base-patch16-256

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 157,840
google

google/siglip2-so400m-patch14-224

No description available.

🎯 zero-shot-image-classification 156,810
BAAI

BAAI/AltCLIP

No description available.

🎯 zero-shot-image-classification 146,390
apple

apple/MobileCLIP-S2-OpenCLIP

No description available.

🎯 zero-shot-image-classification 138,169
timm

timm/vit_base_patch32_clip_224.laion400m_e32

dataset: LAION-400M...

🎯 zero-shot-image-classification 127,651
imageomics

imageomics/bioclip

No description available.

🎯 zero-shot-image-classification 98,882
google

google/siglip2-giant-opt-patch16-384

No description available.

🎯 zero-shot-image-classification 89,035
Xenova

Xenova/clip-vit-base-patch32

No description available.

🎯 zero-shot-image-classification 87,414
vinid

vinid/plip

No description available.

🎯 zero-shot-image-classification 85,595
google

google/siglip2-base-patch16-512

No description available.

🎯 zero-shot-image-classification 79,622
timm

timm/ViT-SO400M-14-SigLIP-384

Model Type: Contrastive Image-Text, Zero-Shot Image Classification. - Original: https://github.com/google-research/bigvision - Dataset: WebL...

🎯 zero-shot-image-classification 77,056
google

google/siglip-base-patch16-384

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 76,866
OFA-Sys

OFA-Sys/chinese-clip-vit-base-patch16

No description available.

🎯 zero-shot-image-classification 72,974
flaviagiammarino

flaviagiammarino/pubmed-clip-vit-base-patch32

PubMedCLIP was trained on the Radiology Objects in COntext (ROCO) dataset, a large-scale multimodal medical imaging dataset. The ROCO datase...

🎯 zero-shot-image-classification 72,354
Marqo

Marqo/marqo-fashionCLIP

No description available.

🎯 zero-shot-image-classification 56,709
google

google/siglip2-base-patch16-256

No description available.

🎯 zero-shot-image-classification 55,923
laion

laion/CLIP-ViT-bigG-14-laion2B-39B-b160k

No description available.

🎯 zero-shot-image-classification 54,351
timm

timm/vit_large_patch14_clip_336.openai

The CLIP model was developed by researchers at OpenAI to learn about what contributes to robustness in computer vision tasks. The model was ...

🎯 zero-shot-image-classification 53,236
google

google/siglip2-large-patch16-512

No description available.

🎯 zero-shot-image-classification 49,675
google

google/siglip2-large-patch16-384

No description available.

🎯 zero-shot-image-classification 47,808
google

google/siglip2-large-patch16-256

No description available.

🎯 zero-shot-image-classification 47,802
facebook

facebook/metaclip-b32-400m

The Demystifying CLIP Data paper aims to reveal CLIP’s method around training data curation. OpenAI never open-sourced code regarding their ...

🎯 zero-shot-image-classification 47,489
timm

timm/ViT-SO400M-14-SigLIP

Model Type: Contrastive Image-Text, Zero-Shot Image Classification. - Original: https://github.com/google-research/bigvision - Dataset: WebL...

🎯 zero-shot-image-classification 47,279
timm

timm/ViT-B-32-SigLIP2-256

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 46,682
timm

timm/vit_large_patch14_clip_224.metaclip_2pt5b

dataset: MetaCLIP-2.5B...

🎯 zero-shot-image-classification 46,366
google

google/siglip2-base-patch16-384

No description available.

🎯 zero-shot-image-classification 44,905
laion

laion/CLIP-ViT-L-14-DataComp.XL-s13B-b90K

No description available.

🎯 zero-shot-image-classification 44,029
google

google/siglip2-so400m-patch16-512

No description available.

🎯 zero-shot-image-classification 41,332
laion

laion/CLIP-ViT-L-14-CommonPool.XL-s13B-b90K

No description available.

🎯 zero-shot-image-classification 36,806
timm

timm/resnet50_clip.openai

No description available.

🎯 zero-shot-image-classification 34,931
timm

timm/vit_base_patch32_clip_224.laion400m_e31

dataset: LAION-400M...

🎯 zero-shot-image-classification 33,728
timm

timm/ViT-B-16-SigLIP2-512

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 33,249
LanguageBind

LanguageBind/LanguageBind_Image

【ICLR 2024 πŸ”₯】LanguageBind: Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment If you like our project, ...

🎯 zero-shot-image-classification 29,787
google

google/medsiglip-448

No description available.

🎯 zero-shot-image-classification 29,747
google

google/siglip2-so400m-patch16-256

No description available.

🎯 zero-shot-image-classification 29,287
laion

laion/CLIP-convnext_xxlarge-laion2B-s34B-b82K-augreg-soup

No description available.

🎯 zero-shot-image-classification 29,120
UCSC-VLAA

UCSC-VLAA/ViT-L-16-HTxt-Recap-CLIP

Model Type: Contrastive Image-Text, Zero-Shot Image Classification. - Original: https://github.com/UCSC-VLAA/Recap-DataComp-1B - Dataset: ht...

🎯 zero-shot-image-classification 23,314
LanguageBind

LanguageBind/LanguageBind_Video_merge

【ICLR 2024 πŸ”₯】LanguageBind: Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment If you like our project, ...

🎯 zero-shot-image-classification 23,161
facebook

facebook/PE-Core-G14-448

[\[πŸ“ƒ Tech Report\]](https://arxiv.org/abs/2504.13181) [\[πŸ“‚ Github\]](https://github.com/facebookresearch/perceptionmodels/) Perception Encod...

🎯 zero-shot-image-classification 22,756
timm

timm/ViT-B-16-SigLIP-256

Model Type: Contrastive Image-Text, Zero-Shot Image Classification. - Original: https://github.com/google-research/bigvision - Dataset: WebL...

🎯 zero-shot-image-classification 21,334
imageomics

imageomics/bioclip-2

No description available.

🎯 zero-shot-image-classification 19,469
google

google/siglip-large-patch16-384

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 18,297
timm

timm/ViT-SO400M-16-SigLIP2-512

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 17,983
zer0int

zer0int/CLIP-GmP-ViT-L-14

πŸ€– New and greatly improved version of the model, check out: - πŸŒ‘ https://huggingface.co/zer0int/CLIP-KO-LITE-TypoAttack-Attn-Dropout-ViT-L-14...

🎯 zero-shot-image-classification 17,600
timm

timm/ViT-SO400M-16-SigLIP2-384

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 17,586
timm

timm/ViT-SO400M-14-SigLIP2

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 17,119
facebook

facebook/metaclip-2-worldwide-huge-quickgelu

No description available.

🎯 zero-shot-image-classification 16,883
timm

timm/eva02_enormous_patch14_plus_clip_224.laion2b_s9b_b144k

No description available.

🎯 zero-shot-image-classification 14,875
Xenova

Xenova/clip-vit-base-patch16

No description available.

🎯 zero-shot-image-classification 14,428
Salesforce

Salesforce/blip2-itm-vit-g

No description available.

🎯 zero-shot-image-classification 14,374
timm

timm/ViT-B-16-SigLIP

Model Type: Contrastive Image-Text, Zero-Shot Image Classification. - Original: https://github.com/google-research/bigvision - Dataset: WebL...

🎯 zero-shot-image-classification 14,349
timm

timm/ViT-B-16-SigLIP2

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 14,106
qihoo360

qihoo360/fg-clip2-so400m

No description available.

🎯 zero-shot-image-classification 13,190
google

google/siglip-base-patch16-512

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 13,120
timm

timm/PE-Core-L-14-336

This is an OpenCLIP (image + text) remaped version of the the original...

🎯 zero-shot-image-classification 12,045
timm

timm/ViT-gopt-16-SigLIP2-384

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 10,789
wkcn

wkcn/TinyCLIP-ViT-61M-32-Text-29M-LAION400M

No description available.

🎯 zero-shot-image-classification 10,624
google

google/siglip-so400m-patch14-224

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 10,277
google

google/siglip-base-patch16-256-multilingual

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 10,006
timm

timm/eva02_base_patch16_clip_224.merged2b_s8b_b131k

No description available.

🎯 zero-shot-image-classification 9,411
timm

timm/eva02_large_patch14_clip_224.merged2b_s4b_b131k

No description available.

🎯 zero-shot-image-classification 9,337
yujiepan

yujiepan/clip-vit-tiny-random-patch14-336

No description available.

🎯 zero-shot-image-classification 8,556
laion

laion/CLIP-ViT-B-16-DataComp.XL-s13B-b90K

No description available.

🎯 zero-shot-image-classification 8,124
kakaobrain

kakaobrain/align-base

No description available.

🎯 zero-shot-image-classification 7,835
timm

timm/ViT-SO400M-16-SigLIP2-256

A SigLIP 2 Vision-Lanuage model trained on WebLI. This model has been converted for use in OpenCLIP from the original JAX checkpoints in Big...

🎯 zero-shot-image-classification 7,492
laion

laion/CLIP-ViT-B-32-256x256-DataComp-s34B-b86K

No description available.

🎯 zero-shot-image-classification 7,491
facebook

facebook/metaclip-h14-fullcc2.5b

The Demystifying CLIP Data paper aims to reveal CLIP’s method around training data curation. OpenAI never open-sourced code regarding their ...

🎯 zero-shot-image-classification 7,488
laion

laion/CLIP-ViT-g-14-laion2B-s34B-b88K

No description available.

🎯 zero-shot-image-classification 7,009
laion

laion/CLIP-convnext_base_w-laion2B-s13B-b82K

No description available.

🎯 zero-shot-image-classification 6,905
wisdomik

wisdomik/QuiltNet-B-32

- zero-shot-image-classification - clip - vision - language - histopathology - histology - medical librarytag: openclip - src: >- https://qu...

🎯 zero-shot-image-classification 6,727
google

google/siglip-large-patch16-256

SigLIP is CLIP, a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a...

🎯 zero-shot-image-classification 6,640