Tchuanm
A PhD student. Working on computer vision, especially for visual tracking and diffusion models.
UCAS; CVC-UABSpain
Tchuanm's Stars
QwenLM/Qwen2.5
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
datawhalechina/so-large-lm
大模型基础: 一文了解大模型基础知识
ydhcg-BoBo/STCMOT
Tchuanm/AViTMP
The official implement of "Exploiting Image-Related Inductive Biases in Single-Branch Visual Tracking".
roboflow/supervision
We write your reusable computer vision tools. 💜
lllyasviel/sd-forge-layerdiffuse
[WIP] Layer Diffusion for WebUI (via Forge)
nubsym/STTrack
Joint Spatial-temporal Modeling for Visual tracking
Event-AHU/FELT_SOT_Benchmark
The First Frame-Event Long-Term Single Object Tracking Benchmark
lllyasviel/ControlNet
Let us control diffusion models!
datawhalechina/daily-interview
Datawhale成员整理的面经,内容包括机器学习,CV,NLP,推荐,开发等,欢迎大家star
guoxiaowhu/UCAS_Thesis
The newest template of thesis for UCAS
TL-System/plato
A federated learning framework to support scalable and reproducible research
lcylmhlcy/Awesome-algorithm-interview
算法工程师(人工智能CV方向)面试问题及相关资料
google/diffseg
DiffSeg is an unsupervised zero-shot segmentation method using attention information from a stable-diffusion model. This repo implements the main DiffSeg algorithm and additionally includes an experimental feature to add semantic labels to the masks based on a generated caption.
adobe-research/custom-diffusion
Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)
IDEA-Research/Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
google/inspiration_tree
Event-AHU/EventVOT_Benchmark
[CVPR-2024] The First High Definition (HD) Event based Visual Object Tracking Benchmark Dataset
Tchuanm/IterInv
The official implement of "IterInv: Iterative Inversion for Pixel-Level T2I Models".
google/break-a-scene
Official implementation for "Break-A-Scene: Extracting Multiple Concepts from a Single Image" [SIGGRAPH Asia 2023]
cure-lab/PnPInversion
[ICLR2024] Official repo for paper "PnP Inversion: Boosting Diffusion-based Editing with 3 Lines of Code"
deep-floyd/IF
rese1f/StableVideo
[ICCV 2023] StableVideo: Text-driven Consistency-aware Diffusion Video Editing
Picsart-AI-Research/Text2Video-Zero
[ICCV 2023 Oral] Text-to-Image Diffusion Models are Zero-Shot Video Generators
Tchuanm/COESOT
A large-scale benchmark dataset for color-event based visual tracking
omerbt/Text2LIVE
Official Pytorch Implementation for "Text2LIVE: Text-Driven Layered Image and Video Editing" (ECCV 2022 Oral)
qiuyu96/CoDeF
[CVPR 2024 Highlight] Official PyTorch implementation of CoDeF: Content Deformation Fields for Temporally Consistent Video Processing
thu-ml/controlvideo
Official implementation for "ControlVideo: Adding Conditional Control for One Shot Text-to-Video Editing"
omerbt/TokenFlow
Official Pytorch Implementation for "TokenFlow: Consistent Diffusion Features for Consistent Video Editing" presenting "TokenFlow" (ICLR 2024)