yangkang779
I am a computer vision researcher, working in the field of object detection and object tracking.
yangkang779's Stars
facebookresearch/segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
open-mmlab/mmdetection
OpenMMLab Detection Toolbox and Benchmark
openai/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
mlfoundations/open_clip
An open source implementation of CLIP.
gaomingqi/Track-Anything
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
fundamentalvision/BEVFormer
[ECCV 2022] This is the official implementation of BEVFormer, a camera-only framework for autonomous driving perception, e.g., 3D object detection and semantic map segmentation.
mit-han-lab/bevfusion
[ICRA'23] BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation
KaiyangZhou/CoOp
Prompt Learning for Vision-Language Models (IJCV'22, CVPR'22)
anuragxel/salt
Segment Anything Labelling Tool
kennymckormick/pyskl
A toolbox for skeleton-based action recognition.
megvii-research/mdistiller
The official implementation of [CVPR2022] Decoupled Knowledge Distillation https://arxiv.org/abs/2203.08679 and [ICCV2023] DOT: A Distillation-Oriented Trainer https://openaccess.thecvf.com/content/ICCV2023/papers/Zhao_DOT_A_Distillation-Oriented_Trainer_ICCV_2023_paper.pdf
microsoft/RegionCLIP
[CVPR 2022] Official code for "RegionCLIP: Region-based Language-Image Pretraining"
zhenyuw16/UniDetector
Code release for our CVPR 2023 paper "Detecting Everything in the Open World: Towards Universal Object Detection".
megvii-research/MOTRv2
[CVPR2023] MOTRv2: Bootstrapping End-to-End Multi-Object Tracking by Pretrained Object Detectors
cwlroda/falldetection_openpifpaf
Fall Detection using OpenPifPaf's Human Pose Estimation model
Syliz517/CLIP-ReID
Official implementation for "CLIP-ReID: Exploiting Vision-Language Model for Image Re-identification without Concrete Text Labels" (AAAI 2023)
DirtyHarryLYL/Transferable-Interactiveness-Network
Code for Transferable Interactiveness Knowledge for Human-Object Interaction Detection. (CVPR'19, TPAMI'21)
DirtyHarryLYL/HAKE-Action-Torch
HAKE-Action in PyTorch
anosorae/IRRA
Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person Retrieval (CVPR 2023)
wangxiyang2022/DeepFusionMOT
Code for RA-L journal and IROS 2022 paper "DeepFusionMOT: A 3D Multi-Object Tracking Framework Based on Camera-LiDAR Fusion with Deep Association".
WaterScenes/WaterScenes
Official repository for WaterScenes dataset
pengzhiliang/G2SD
Bigtuo/YOLOv7-Pose-Bytetrack-STGCN
YOLOv7-POSE was used for key point detection, Bytetrack for tracking, and Stgan for fall and other behavior recognition
Y-B-Class-Projects/Human-Fall-Detection
Human Falling Detection
zhangyp15/SimMOD
Implementation of SimMOD: A Simple Baseline for Multi-Camera 3D Object Detection
Bigtuo/YOLO-POSE-Bytetrack-STGCN
YOLO-POSE was used for key point detection, Bytetrack for tracking, and STGCN for fall and other behavior recognition
nvnnghia/Online-Realtime-Action-Recognition-based-on-OpenPose
A skeleton-based real-time online action recognition project, classifying and recognizing base on framewise joints, which can be used for safety surveilence.
triple-Mu/mmyolo
OpenMMLab YOLO series toolbox and benchmark
Zhihaibi/2019_SEU_FaceAi
yangkang779/YOLO-POSE-Bytetrack-STGCN
YOLO-POSE was used for key point detection, Bytetrack for tracking, and STGCN for fall and other behavior recognition