Prof-pengyin
Assistant Professor City University of Hong Kong Coding and designing for general artificial Intelligence.
City University of Hong KongHong Kong, China
Prof-pengyin's Stars
openai/whisper
Robust Speech Recognition via Large-Scale Weak Supervision
facebookresearch/segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
bmaltais/kohya_ss
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
google-deepmind/mujoco
Multi-Joint dynamics with Contact. A general purpose physics simulator.
py-why/dowhy
DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks.
huggingface/lerobot
🤗 LeRobot: Making AI for Robotics more accessible with end-to-end learning
andrewyng/translation-agent
agiresearch/AIOS
AIOS: LLM Agent Operating System
facebookresearch/jepa
PyTorch code and models for V-JEPA self-supervised learning from video.
agiresearch/OpenAGI
OpenAGI: When LLM Meets Domain Experts
BAAI-Agents/Cradle
The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.
google-deepmind/mujoco_menagerie
A collection of high-quality models for the MuJoCo physics engine, curated by Google DeepMind.
OSU-NLP-Group/HippoRAG
HippoRAG is a novel RAG framework inspired by human long-term memory that enables LLMs to continuously integrate knowledge across external documents. RAG + Knowledge Graphs + Personalized PageRank.
openvla/openvla
OpenVLA: An open-source vision-language-action model for robotic manipulation.
google-deepmind/mujoco_mpc
Real-time behaviour synthesis with MuJoCo, using Predictive Control
google-deepmind/open_x_embodiment
CYHSM/awesome-neuro-ai-papers
Papers from the intersection of deep learning and neuroscience
Improbable-AI/VisionProTeleop
VisionOS App + Python Library to stream head / wrist / finger tracking data from Vision Pro to any robots.
1x-technologies/1xgpt
world modeling challenge for humanoid robots
dingo-actual/infini-transformer
PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" (https://arxiv.org/abs/2404.07143)
danijar/daydreamer
DayDreamer: World Models for Physical Robot Learning
Srameo/LE3D
HDR 3D Scene Editing!
Timothyxxx/WorldModelPapers
Paper collections of the continuous effort start from World Models.
kvablack/susie
Code for subgoal synthesis via image editing
kodie-artner/AR-RViz
Unity Project for visualization and control of ROS systems in augmented reality
guiglass/Mocap_Fusion_Gloves
rail-berkeley/soar
Code release for paper "Autonomous Improvement of Instruction Following Skills via Foundation Models" | CoRL 2024
johnrso/spawnnet
SpawnNet: Learning Generalizable Visuomotor Skills from Pre-trained Networks
NUS-LinS-Lab/ManiFM
Official Site for ManiFoundation Model