Issues
- 0
- 1
- 14
Request for Sharing QA Data Files
#24 opened by Cloud-li1 - 4
about nuscenes data convert
#15 opened by AMzhanghan - 3
- 0
- 0
关于visual encoder
#28 opened by Whale-ice - 2
Can upload the parameters to modelscope, thx
#12 opened by zizaisuiyuan - 0
Question about E2E part
#25 opened by Cloud-li1 - 2
Number of waypoints
#21 opened by missTL - 5
execute sh senna_nusc_converter.sh error
#17 opened by StrongTsai - 2
How to convert our model to the ONNX
#22 opened by chi0612 - 2
What is the DriveX dataset?
#2 opened by b5strbal - 1
Some weights of LlavaLlamaForCausalLM were not initialized from the model checkpoint
#23 opened by MIKE-GUO233 - 1
Using vicuna-7b-v1.5 model.
#20 opened by comflife - 1
Raw data获取过程询问
#19 opened by fjq-tongji - 4
senna_nusc_data_converter.py error
#18 opened by wolf943134497 - 1
- 5
[E2E module to LLM]
#10 opened by Lewis-Lu - 2
Requires for details of python environment
#9 opened by xiaodongww - 3
- 4
eval_plan_qa.json
#13 opened by PG-Wang - 1
- 1
If we want to achieve full scene understanding, what is the approximate cost of the vlm annotation solution
#7 opened by hunkyu - 1
Model Deployment
#6 opened by chensiweiTHU - 2
Code Release
#5 opened by abhigoku10 - 2
question about meta-action encoder
#4 opened by TmacTmac1992 - 1
Is it possible to achieve a large-scale indoor navigation function by providing a model with a video of an indoor space that can be analyzed based on the indoor layout, and then inputting a picture and the destination I want to reach, so that it can recognize my location and perform path planning and navigation? How many VRAMs are needed and how long will they be open sourced? I'm looking forward to it
#1 opened by libai-lab - 1
stage 2 traning error
#14 opened by PG-Wang - 4
Required time for training
#3 opened by kemaloksuz