jianguoz's Stars
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
BerriAI/litellm
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
ShishirPatil/gorilla
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
OpenBMB/XAgent
An Autonomous LLM Agent for Complex Task Solving
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
yizhongw/self-instruct
Aligning pretrained language models with instruction data generated by themselves.
xlang-ai/OpenAgents
[COLM 2024] OpenAgents: An Open Platform for Language Agents in the Wild
MeetKai/functionary
Chat language model that can use tools and interpret the results
THUDM/AgentTuning
AgentTuning: Enabling Generalized Agent Abilities for LLMs
Xwin-LM/Xwin-LM
Xwin-LM: Powerful, Stable, and Reproducible LLM Alignment
billxbf/ReWOO
Decoupling Reasoning from Observations for Efficient Augmented Language Models
ruixiangcui/AGIEval
microsoft/CodeT
hkust-nlp/deita
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
SalesforceAIResearch/AgentLite
allenai/lumos
Code and data for "Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs"
nexusflowai/NexusRaven-V2
InternLM/Agent-FLAN
[ACL2024 Findings] Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for Large Language Models
nexusflowai/NexusRaven
NexusRaven-13B, a new SOTA Open-Source LLM for function calling. This repo contains everything for reproducing our evaluation on NexusRaven-13B and baselines.
SalesforceAIResearch/xLAM
hkust-nlp/AgentBoard
An Analytical Evaluation Board of Multi-turn LLM Agents
Ber666/ToolkenGPT
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings - NeurIPS 2023 (oral)
snap-stanford/MLAgentBench
copilot-us/chatgpt-plugins
Official ChatGPT Plugins🧩
open-compass/T-Eval
[ACL2024] T-Eval: Evaluating Tool Utilization Capability of Large Language Models Step by Step
Data-Provenance-Initiative/Data-Provenance-Collection
CASIA-LM/MoDS
xingyaoww/mint-bench
Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Zihan Wang*, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng and Heng Ji.
gpt4life/alpagasus
Unofficial implementation of AlpaGasus
Junjie-Ye/ToolEyes
ToolEyes: Fine-Grained Evaluation for Tool Learning Capabilities of Large Language Models in Real-world Scenarios