kennymckormick
Ph.D. candidate @ CUHK MMLAB, working on video understanding.
Shanghai AI LaboratoryShanghai
kennymckormick's Stars
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
InternLM/InternLM
Official release of InternLM2.5 base and chat models. 1M context support
OpenGVLab/InternVL
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
google-deepmind/alphageometry
open-compass/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
InternLM/xtuner
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
LC1332/Luotuo-Chinese-LLM
骆驼(Luotuo): Open Sourced Chinese Language Models. Developed by 陈启源 @ 华中师范大学 & 李鲁鲁 @ 商汤科技 & 冷子昂 @ 商汤科技
InternLM/InternLM-XComposer
InternLM-XComposer-2.5: A Versatile Large Vision Language Model Supporting Long-Contextual Input and Output
baaivision/Emu
Emu Series: Generative Multimodal Models from BAAI
InternLM/HuixiangDou
HuixiangDou: Overcoming Group Chat Scenarios with LLM-based Technical Assistance
open-compass/VLMEvalKit
Open-source evaluation toolkit of large vision-language models (LVLMs), support 160+ VLMs, 50+ benchmarks
open-compass/MixtralKit
A toolkit for inference and evaluation of 'mixtral-8x7b-32kseqlen' from Mistral AI
Leymore/ruozhiba
ali-vilab/FlashFace
InternLM/Agent-FLAN
[ACL2024 Findings] Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for Large Language Models
InternLM/InternEvo
InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.
OpenRobotLab/HIMLoco
Learning-based locomotion control from OpenRobotLab, including Hybrid Internal Model & H-Infinity Locomotion Control
open-compass/LawBench
Benchmarking Legal Knowledge of Large Language Models
open-compass/MMBench
Official Repo of "MMBench: Is Your Multi-modal Model an All-around Player?"
JourneyDB/JourneyDB
MMStar-Benchmark/MMStar
[NeurIPS 2024] This repo contains evaluation code for the paper "Are We on the Right Way for Evaluating Large Vision-Language Models"
open-compass/BotChat
Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.
open-compass/MathBench
[ACL 2024 Findings] MathBench: A Comprehensive Multi-Level Difficulty Mathematics Evaluation Dataset
xverse-ai/XVERSE-V-13B
CUHK-ARISE/EmotionBench
Benchmarking LLMs' Emotional Alignment with Humans
open-compass/Ada-LEval
The official implementation of "Ada-LEval: Evaluating long-context LLMs with length-adaptable benchmarks"
alonj/Same-Task-More-Tokens
The code for the paper: "Same Task, More Tokens: the Impact of Input Length on the Reasoning Performance of Large Language Models"
JF-D/Proteus
yysijie/chatgpt-dingtalk-robot