gicheonkang's Stars
qpwodlsqp/CSEConv
The official implementation of "Continuous SO(3) Equivariant Convolution for 3D Point Cloud Analysis" [ECCV 24]
gicheonkang/clip-rt
📎 + 🦾 CLIP-RT: Learning Language-Conditioned Robotic Policies from Natural Language Supervision
alirezadir/Machine-Learning-Interviews
This repo is meant to serve as a guide for Machine Learning/AI technical interviews.
openvla/openvla
OpenVLA: An open-source vision-language-action model for robotic manipulation.
iwhwang/Fine-Grained-Causal-RL
Fine-Grained Causal Dynamics Learning with Quantization for Improving Robustness in Reinforcement Learning (ICML 2024)
JHKim-snu/PGA
[IROS 2024] PGA: Personalizing Grasping Agents with Single Human-Robot Interaction
gicheonkang/prograsp
🦾 PyTorch Implementation for the ICRA'24 Paper, "PROGrasp: Pragmatic Human-Robot Communication for Object Grasping"
kyegomez/RT-2
Democratization of RT-2 "RT-2: New model translates vision and language into action"
google-research/robotics_transformer
JHKim-snu/GVCCI
[IROS 2023] GVCCI: Lifelong Learning of Visual Grounding for Language-Guided Robotic Manipulation
opendilab/awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
OFA-Sys/OFA
Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
gicheonkang/gst-visdial
:speech_balloon: Official PyTorch Implementation for CVPR'23 Paper, "The Dialog Must Go On: Improving Visual Dialog via Generative Self-Training"
iwhwang/SelecMix
SelecMix: Debiased Learning by Contradicting-pair Sampling (NeurIPS 2022)
cshizhe/VLN-HAMT
Official implementation of History Aware Multimodal Transformer for Vision-and-Language Navigation (NeurIPS'21).
YicongHong/Discrete-Continuous-VLN
Code and Data of the CVPR 2022 paper: Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language Navigation
jacobkrantz/IVLN-CE
Official Implementation of IVLN-CE: Iterative Vision-and-Language Navigation in Continuous Environments
LeapLabTHU/Pseudo-Q
[CVPR 2022] Pseudo-Q: Generating Pseudo Language Queries for Visual Grounding
sudharsan13296/Awesome-Meta-Learning
A curated list of Meta Learning papers, code, books, blogs, videos, datasets and other resources.
jamixlee/placenet
Implementation of PlaceNet
facebookresearch/simmc2
Code for SIMMC 2.0: A Task-oriented Dialog Dataset for Immersive Multimodal Conversations
dialogtekgeek/AVSD-DSTC10_Official
Audio Visual Scene-Aware Dialog (AVSD) Challenge at the 10th Dialog System Technology Challenge (DSTC)
ChunyuanLI/Optimus
Optimus: the first large-scale pre-trained VAE language model
j-min/VL-T5
PyTorch code for "Unifying Vision-and-Language Tasks via Text Generation" (ICML 2021)
vmurahari3/visdial-bert
Implementation for "Large-scale Pretraining for Visual Dialog" https://arxiv.org/abs/1912.02379
gicheonkang/sglkt-visdial
🌈 PyTorch Implementation for EMNLP'21 Findings "Reasoning Visual Dialog with Sparse Graph Learning and Knowledge Transfer"
kibeomKim/GACE-GDAN
yuewang-cuhk/awesome-vision-language-pretraining-papers
Recent Advances in Vision and Language PreTrained Models (VL-PTMs)
google-research-datasets/conceptual-12m
Conceptual 12M is a dataset containing (image-URL, caption) pairs collected for vision-and-language pre-training.
facebookresearch/ParlAI
A framework for training and evaluating AI models on a variety of openly available dialogue datasets.