Pinned Repositories
d3rlpy
An offline deep reinforcement learning library
junming-yang.github.io
Personal Website: Junming Yang (杨骏铭)
mbpo
Code for the paper "When to Trust Your Model: Model-Based Policy Optimization"
MBPO-pytorch
Model-based Policy Optimization re-implement by pytorch
MOAN
Model-based Offline Policy Optimization with Adversarial Network
mopo
Model-based Offline Policy Optimization re-implement all by pytorch
offline_rl
Offline RL codebase for Unstable Baselines
VLMEvalKit
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 30+ HF models, 15+ benchmarks
VLMEvalKit
Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 40+ benchmarks
WorldModelPapers
Paper collections of the continuous effort start from World Models.
junming-yang's Repositories
junming-yang/mopo
Model-based Offline Policy Optimization re-implement all by pytorch
junming-yang/MBPO-pytorch
Model-based Policy Optimization re-implement by pytorch
junming-yang/MOAN
Model-based Offline Policy Optimization with Adversarial Network
junming-yang/junming-yang.github.io
Personal Website: Junming Yang (杨骏铭)
junming-yang/offline_rl
Offline RL codebase for Unstable Baselines
junming-yang/VLMEvalKit
Open-source evaluation toolkit of large vision-language models (LVLMs), support GPT-4v, Gemini, QwenVLPlus, 30+ HF models, 15+ benchmarks
junming-yang/d3rlpy
An offline deep reinforcement learning library
junming-yang/mbpo
Code for the paper "When to Trust Your Model: Model-Based Policy Optimization"
junming-yang/mopo-pytorch
re-implementation of the offline model-based RL algorithm MOPO in pytorch
junming-yang/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
junming-yang/WorldModelPapers
Paper collections of the continuous effort start from World Models.