zhangcong2711's Stars
MedicineToken/Medical-SAM2
Medical SAM 2: Segment Medical Images As Video Via Segment Anything Model 2
YichiZhang98/SAM4MIS
SAM & SAM 2 for Medical Image Segmentation: Open-Source Project Summary
richard-peng-xia/CARES
[arXiv'24 & ICMLW'24] CARES: A Comprehensive Benchmark of Trustworthiness in Medical Vision Language Models
andrewekhalel/MLQuestions
Machine Learning and Computer Vision Engineer - Technical Interview Questions
iioSnail/chinese_medical_ner
中文医疗领域的命名实体识别
FreedomIntelligence/HuatuoGPT-II
HuatuoGPT2, One-stage Training for Medical Adaption of LLMs. (An Open Medical GPT)
uf-hobi-informatics-lab/ClinicalTransformerRelationExtraction
A project for developing transformer-based models for clinical relation extraction
uf-hobi-informatics-lab/ClinicalTransformerNER
a library for named entity recognition developed by UF HOBI NLP lab featuring SOTA algorithms
huyuanxin/CMeKGCrawler
Medical Graph for Neo4j
king-yyf/CMeKG_tools
zhao-zy15/PMC-Patients
PMC-Patients
linhandev/dataset
医学影像数据集列表 『An Index for Medical Imaging Datasets』
explodinggradients/ragas
Evaluation framework for your Retrieval Augmented Generation (RAG) pipelines
bcmi/libcom
Image composition toolbox: everything you want to know about image composition or object insertion
bcmi/Awesome-Image-Harmonization
A curated list of papers, code and resources pertaining to image harmonization.
TigerResearch/TigerBot
TigerBot: A multi-language multi-task LLM
mbzuai-oryx/Video-LLaVA
PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models
mbzuai-oryx/Video-ChatGPT
[ACL 2024 🔥] Video-ChatGPT is a video conversation model capable of generating meaningful conversation about videos. It combines the capabilities of LLMs with a pretrained visual encoder adapted for spatiotemporal video representation. We also introduce a rigorous 'Quantitative Evaluation Benchmarking' for video-based conversational models.
PKU-YuanGroup/Video-LLaVA
Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
PKU-YuanGroup/Chat-UniVi
[CVPR 2024 Highlight🔥] Chat-UniVi: Unified Visual Representation Empowers Large Language Models with Image and Video Understanding
THUDM/CogVLM
a state-of-the-art-level open visual language model | 多模态预训练模型
Alpha-VLLM/LLaMA2-Accessory
An Open-source Toolkit for LLM Development
SALT-NLP/LLaVAR
Code/Data for the paper: "LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image Understanding"
mlfoundations/open_clip
An open source implementation of CLIP.
fudan-zvg/Semantic-Segment-Anything
Automated dense category annotation engine that serves as the initial semantic labeling for the Segment Anything dataset (SA-1B).
TXH-mercury/VALOR
Codes and Models for VALOR: Vision-Audio-Language Omni-Perception Pretraining Model and Dataset
salesforce/LAVIS
LAVIS - A One-stop Library for Language-Vision Intelligence
DAMO-NLP-SG/Video-LLaMA
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
shruti-jadon/Video-Summarization-using-Keyframe-Extraction-and-Video-Skimming
Experimenting with different Summarizing techniques on SumMe Dataset
KaiyangZhou/vsumm-reinforce
AAAI 2018 - Unsupervised video summarization with deep reinforcement learning (Theano)