vivit
There are 11 repositories under vivit topic.
mx-mark/VideoTransformer-pytorch
PyTorch implementation of a collections of scalable Video Transformer Benchmarks.
AliAmini93/ViViT-Medical-Video-Classification
Developed the ViViT model for medical video classification, enhancing 3D organ image analysis using transformer-based architectures.
sanket-poojary-03/Fine-tuning-ViVit
Python script to fine tune Open source Video Vision Transformer (ViVit) using HuggingFace Trainer Library
TomStog/Infrared-SpO2
The dataset used for the "A non-contact SpO2 estimation using video magnification and infrared data" publication
moharamfatema/graduation-project
Video vision transformers for hierarchical anomaly detection in video scenes.
quanghuy0497/MM-Fit_Excercise_Recognition
Some incomplete works with 2D action recognition on MM-Fit dataset using ViT, ViViT, and MLP-Mixer Topics Resources
amitparag/Attention-Classification
Slip detection with Franka Emika and GelSight Sensors
brianhill11/ViViT-tensorflow
Unofficial Tensorflow implementation of the ViViT model architecture
denpalrius/sports_action_recognition
A comparative study of ViViT, CNN-GRU sequence models for video action recognition using the UCF101 dataset
KimRass/ViViT
PyTorch implementation of 'ViViT' (Arnab et al., 2021) from scratch.
JakobGr/ADA
This repository contains code for training and evaluating transformer-based models like TimeSformer and VideoMAE for sign language recognition on the WLASL dataset. The project includes frame sampling techniques, preprocessing pipelines, fine-tuning strategies, and performance evaluation using metrics like top-1, top-5, and top-10 accuracy.