YueFan1014
Research Engineer at BIGAI. Interested in Embodied AI, video understanding, robotics.
Beijing, China
YueFan1014's Stars
langgenius/dify
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
vnpy/vnpy
基于Python的开源量化交易平台开发框架
IDEA-Research/Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
edeng23/binance-trade-bot
Automated cryptocurrency trading bot
threestudio-project/threestudio
A unified framework for 3D content generation.
OpenGVLab/Ask-Anything
[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
snap-stanford/ogb
Benchmark datasets, data loaders, and evaluators for graph machine learning
PaddlePaddle/PGL
Paddle Graph Learning (PGL) is an efficient and flexible graph learning framework based on PaddlePaddle
zhengjingwei/machine-learning-interview
算法工程师-机器学习面试题总结
siyuanliii/masa
Official Implementation of CVPR24 highligt paper: Matching Anything by Segmenting Anything
robodhruv/visualnav-transformer
Official code and checkpoint release for mobile robot foundation models: GNM, ViNT, and NoMaD.
Cheems-Seminar/grounded-segment-any-parts
Grounded Segment Anything: From Objects to Parts
zjukg/NeuralKG
[Tool] For Knowledge Graph Representation Learning
facebookresearch/Ego4d
Ego4d dataset repository. Download the dataset, visualize, extract features & example usage of the dataset
allenai/ScienceWorld
ScienceWorld is a text-based virtual environment centered around accomplishing tasks from the standardized elementary science curriculum.
WaldJohannaU/3RScan
3RScan Toolkit
scene-verse/SceneVerse
Official implementation of ECCV24 paper "SceneVerse: Scaling 3D Vision-Language Learning for Grounded Scene Understanding"
wz0919/ScaleVLN
[ICCV 2023 Oral]: Scaling Data Generation in Vision-and-Language Navigation
sharinka0715/semantic-gaussians
Official implemetation of the paper "Semantic Gaussians: Open-Vocabulary Scene Understanding with 3D Gaussian Splatting".
real-stanford/cow
[CVPR 2023] CoWs on Pasture: Baselines and Benchmarks for Language-Driven Zero-Shot Object Navigation
YueFan1014/VideoAgent
This is the official code of VideoAgent: A Memory-augmented Multimodal Agent for Video Understanding (ECCV 2024)
facebookresearch/EgoVLPv2
Code release for "EgoVLPv2: Egocentric Video-Language Pre-training with Fusion in the Backbone" [ICCV, 2023]
Ram81/habitat-web
Habitat-Web is a web application to collect human demonstrations for embodied tasks on Amazon Mechanical Turk (AMT) using the Habitat simulator.
facebookresearch/EgoTV
EgoTV Egocentric Task Verification from Natural Language Task Descriptions
clova-tool/CLOVA-tool
houzhijian/GroundNLQ
The champion solution for Ego4D Natural Language Queries Challenge in CVPR 2023
MM-FIRE/FIRE
ant-research/Parameter_Inference_Efficient_PIE
MIRALab-USTC/KDDCup2021_WikiKG90M_GraphMIRAcles
YueFan1014/3DCEMA
This is the code of paper ‘Gene Regulatory Network Inference using 3D Convolutional Neural Network’.