GithubHelp home page GithubHelp logo

opendrivelab / elm Goto Github PK

View Code? Open in Web Editor NEW
119.0 8.0 7.0 5.45 MB

[ECCV 2024] Embodied Understanding of Driving Scenarios

Python 99.96% Shell 0.04%
autonomous-driving vision-language-model end-to-end-driving

elm's People

Contributors

devlinyan avatar eltociear avatar faikit avatar retsuh-bqw avatar zhouyunsong-sjtu avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar

elm's Issues

TypeError: default_drivelm() missing 1 required positional argument: 'ann_paths'

Below attaching the full log:

Using downloaded and verified file: /home/pravaig-20/ELM/lavis/.././data/drivelm/train.json
Using downloaded and verified file: /home/pravaig-20/ELM/lavis/.././data/drivelm/val.json
Using downloaded and verified file: /home/pravaig-20/ELM/lavis/.././data/drivelm/val.json
2024-04-03 10:44:16,614 [INFO] Building datasets...
Traceback (most recent call last):
  File "scripts/train.py", line 112, in <module>
    main()
  File "scripts/train.py", line 104, in main
    datasets = task.build_datasets(cfg)
  File "/home/pravaig-20/ELM/lavis/tasks/vqa.py", line 74, in build_datasets
    datasets = super().build_datasets(cfg)
  File "/home/pravaig-20/ELM/lavis/tasks/base_task.py", line 57, in build_datasets
    dataset = builder.build_datasets()
  File "/home/pravaig-20/ELM/lavis/datasets/builders/base_dataset_builder.py", line 58, in build_datasets
    datasets = self.build()  # dataset['train'/'val'/'test']
  File "/home/pravaig-20/ELM/lavis/datasets/builders/base_dataset_builder.py", line 228, in build
    datasets[split] = dataset_cls(
  File "/home/pravaig-20/ELM/lavis/datasets/datasets/elm_datasets.py", line 84, in __init__
    self.default_drivelm()
TypeError: default_drivelm() missing 1 required positional argument: 'ann_paths'

Train detail and GPU hours

Great work!
I have some questions to ask you.
I would like to know your training details during the pre-train phase, such as the amount of training data and training configuration.
Can you tell me the GPU hours for pre-train and fine-tuned?
It seems that you are only using the front view video and not the multi view. I don't know if I understand correctly

How to get Pr@k metric?

How to obtain the Pr@k indicators proposed in the paper? I can't seem to find the relevant code.

Typo in video_clip_processor.py

one typo in tools/video_clip_processor

from lavis.datasets.data_utils load_pickle

to

from lavis.datasets.data_utils import load_pickle

How to get the QAs about surrounding narration, Action & Decision?

I found that the responses to keyword I found that the responses to keyword A were all invalid:
were all invalid in drivelm_train.json:
like: Apologies, but I do not have information about the surrounding objects at this time.

What do I do if I want to compare these metrics about surrounding narration, Action & Decision?

Question about preprocess of nuscenes dataset

In lavis/datasets/datasets/elm_datasets.py and lavis/projects/blip2/train/advqa_t5_elm.yaml, I found some necessary files that hasn't been mentioned in README.

  • How to get bevdetv2-nuscenes_infos_trainval.pkl?
  • I've seen drivelm_train.json in this link , but how can I get 'data/drivelm/val.json' in advqa_t5_elm.yaml? Will this file be available in the future?

What is the temporal length for box qa?

I try to reproduce the code but I find the model use temporal image lists in default, which I think is the settings of default_box_qa but not default_drivelm. Nowhere says the temporal length issue, however.

About planning eval on waymo

Hi, it's a nice work!
I notice that you test the planning on waymo dataset. It's structure is differrent from nuscenes, so can you provide the scripts to get waymo trajectory gt and eval ? thanks so much

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.