hsaest's Stars
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
NielsRogge/Transformers-Tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
jessevig/bertviz
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
OpenGVLab/InternVL
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
pytorch/captum
Model interpretability and understanding for PyTorch
QwenLM/Qwen2-VL
Qwen2-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
datawhalechina/learn-nlp-with-transformers
we want to create a repo to illustrate usage of transformers in chinese
lucidrains/gigagan-pytorch
Implementation of GigaGAN, new SOTA GAN out of Adobe. Culmination of nearly a decade of research into GANs
cambrian-mllm/cambrian
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
MrYxJ/calculate-flops.pytorch
The calflops is designed to calculate FLOPs、MACs and Parameters in all various neural networks, such as Linear、 CNN、 RNN、 GCN、Transformer(Bert、LlaMA etc Large Language Model)
QwenLM/Qwen2-Math
A series of math-specific large language models of our Qwen2 series.
LMD0311/Awesome-World-Model
Collect some World Models for Autonomous Driving papers.
ChenLiu-1996/CitationMap
A simple pip-installable Python tool to generate your own HTML citation world map from your Google Scholar ID.
enoche/MMRec
A Toolbox for MultiModal Recommendation. Integrating 10+ Models...
google-research/self-organising-systems
westlake-repl/Recommendation-Systems-without-Explicit-ID-Features-A-Literature-Review
Paper List of Pre-trained Foundation Recommender Models
lupantech/MathVista
MathVista: data, code, and evaluation for Mathematical Reasoning in Visual Contexts
mathllm/MathCoder
Family of LLMs for mathematical reasoning.
OSU-NLP-Group/GrokkedTransformer
Code for the paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'
JFan1997/Awesome_PhD_Opportunities
This repository is used for advertising PhD recruitment opportunities. Contributions are welcome!
Ber666/RAP
Reasoning with Language Model is Planning with World Model
ZrrSkywalker/MathVerse
[ECCV 2024] Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?
kohjingyu/search-agents
Code for the paper 🌳 Tree Search for Language Model Agents
xufangzhi/ENVISIONS
A Neural-Symbolic Self-Training Framework
THUDM/VisualAgentBench
Towards Large Multimodal Models as Visual Foundation Agents
siyuyuan/evoagent
Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"
HZQ950419/Math-LLaVA
Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models
PathMMU-Benchmark/PathMMU