Results for "video-classification"

51 matches found.

microsoft

microsoft/xclip-base-patch32

X-CLIP is a minimal extension of CLIP for general video-language understanding. The model is trained in a contrastive way on (video, text) p...

📼 video-classification 185,308
google

google/videoprism-base-f16r288

We release the following model variants: | Model Name | Configuration Name | Model Type | Backbone | #Params | File Size | Checkpoint | | --...

📼 video-classification 137,543
ai-forever

ai-forever/kandinsky-videomae-large-camera-motion

VideoMAE model(`large`) variant that has been finetuned for multi-label video classification (a video can belong to multiple classes simulta...

📼 video-classification 129,242
google

google/videoprism-lvt-base-f16r288

We release the following model variants: | Model Name | Configuration Name | Model Type | Backbone | #Params | File Size | Checkpoint | | --...

📼 video-classification 93,906
facebook

facebook/vjepa2-vitg-fpc64-256

No description available.

📼 video-classification 89,398
facebook

facebook/timesformer-base-finetuned-k400

No description available.

📼 video-classification 88,131
facebook

facebook/vjepa2-vitl-fpc64-256

No description available.

📼 video-classification 59,832
MCG-NJU

MCG-NJU/videomae-base

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 52,242
google

google/vivit-b-16x2-kinetics400

ViViT is an extension of the Vision Transformer (ViT) to video. We refer to the paper for details....

📼 video-classification 38,447
MCG-NJU

MCG-NJU/videomae-base-finetuned-kinetics

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 20,356
facebook

facebook/vjepa2-vitg-fpc64-384-ssv2

No description available.

📼 video-classification 19,330
microsoft

microsoft/xclip-base-patch16-16-frames

X-CLIP is a minimal extension of CLIP for general video-language understanding. The model is trained in a contrastive way on (video, text) p...

📼 video-classification 15,374
facebook

facebook/vjepa2-vitg-fpc64-384

No description available.

📼 video-classification 15,121
facebook

facebook/vjepa2-vitl-fpc16-256-ssv2

No description available.

📼 video-classification 12,395
OpenGVLab

OpenGVLab/VideoMAEv2-Base

No description available.

📼 video-classification 12,284
microsoft

microsoft/xclip-base-patch32-16-frames

X-CLIP is a minimal extension of CLIP for general video-language understanding. The model is trained in a contrastive way on (video, text) p...

📼 video-classification 8,444
google

google/vivit-b-16x2

ViViT is an extension of the Vision Transformer (ViT) to video. We refer to the paper for details....

📼 video-classification 7,279
google

google/videoprism-lvt-large-f8r288

We release the following model variants: | Model Name | Configuration Name | Model Type | Backbone | #Params | File Size | Checkpoint | | --...

📼 video-classification 6,786
OpenGVLab

OpenGVLab/VideoMAEv2-Large

No description available.

📼 video-classification 5,801
microsoft

microsoft/xclip-large-patch14

X-CLIP is a minimal extension of CLIP for general video-language understanding. The model is trained in a contrastive way on (video, text) p...

📼 video-classification 5,785
microsoft

microsoft/xclip-base-patch16

X-CLIP is a minimal extension of CLIP for general video-language understanding. The model is trained in a contrastive way on (video, text) p...

📼 video-classification 4,864
facebook

facebook/timesformer-base-finetuned-k600

No description available.

📼 video-classification 4,010
microsoft

microsoft/xclip-base-patch16-zero-shot

X-CLIP is a minimal extension of CLIP for general video-language understanding. The model is trained in a contrastive way on (video, text) p...

📼 video-classification 3,751
facebook

facebook/vjepa2-vith-fpc64-256

No description available.

📼 video-classification 3,416
MCG-NJU

MCG-NJU/videomae-large

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 3,396
OpenGVLab

OpenGVLab/VideoMAEv2-Huge

No description available.

📼 video-classification 2,458
OpenGVLab

OpenGVLab/InternVideo2-Stage2_6B

No description available.

📼 video-classification 2,230
MCG-NJU

MCG-NJU/videomae-small-finetuned-kinetics

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 2,164
MCG-NJU

MCG-NJU/videomae-large-finetuned-kinetics

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 1,812
MCG-NJU

MCG-NJU/videomae-base-finetuned-ssv2

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 1,488
MCG-NJU

MCG-NJU/videomae-huge-finetuned-kinetics

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 1,173
google

google/videoprism-large-f8r288

We release the following model variants: | Model Name | Configuration Name | Model Type | Backbone | #Params | File Size | Checkpoint | | --...

📼 video-classification 986
facebook

facebook/timesformer-base-finetuned-ssv2

No description available.

📼 video-classification 947
qubvel-hf

qubvel-hf/vjepa2-vitl-fpc16-256-ssv2

No description available.

📼 video-classification 929
OpenGVLab

OpenGVLab/VideoMAEv2-giant

No description available.

📼 video-classification 833
MCG-NJU

MCG-NJU/videomae-small-finetuned-ssv2

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 592
facebook

facebook/vjepa2-vitl-fpc32-256-diving48

No description available.

📼 video-classification 557
ttyh

ttyh/videomae-base-finetuned-ucf101-subset

More information needed...

📼 video-classification 496
mitegvg

mitegvg/videomae-tiny-92-kinetics-binary-finetuned-xd-violence

More information needed...

📼 video-classification 478
MCG-NJU

MCG-NJU/videomae-base-ssv2

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 378
MCG-NJU

MCG-NJU/videomae-base-short

VideoMAE is an extension of Masked Autoencoders (MAE) to video. The architecture of the model is very similar to that of a standard Vision T...

📼 video-classification 284
Nikeytas

Nikeytas/videomae-crime-detector-ultra-v1

No description available.

📼 video-classification 275
KhoiBui

KhoiBui/tiktok-video-safety-classifier

Base Model: VideoMAE (MCG-NJU/videomae-base-finetuned-kinetics) - Task: Binary classification (safe/harmful) - Input: 16 frames, 224x224...

📼 video-classification 251
qualcomm

qualcomm/ResNet-Mixed-Convolution

Model Type: Modelusecase.videoclassification Model Stats: - Model checkpoint: Kinetics-400 - Input resolution: 112x112 - Number of parameter...

📼 video-classification 223
DanJoshua

DanJoshua/videomae-base-finetuned-rwf2000-subset

More information needed...

📼 video-classification 218
Naman712

Naman712/Deep-fake-detection

No description available.

📼 video-classification 199
Shawon16

Shawon16/timesformer_wlasl_100_200ep_coR_

More information needed...

📼 video-classification 199
nateraw

nateraw/videomae-base-finetuned-ucf101-subset

More information needed...

📼 video-classification 154
Ammar2k

Ammar2k/videomae-base-finetuned-deepfake-subset

No description available....

📼 video-classification 148
muneeb1812

muneeb1812/videomae-base-fake-video-classification

More information needed...

📼 video-classification 142
TanAlexanderlz

TanAlexanderlz/ALL_RGBCROP_ori16F-8B16F-GACWD1

More information needed...

📼 video-classification 136