GithubHelp home page GithubHelp logo

wayveai / driving-with-llms Goto Github PK

View Code? Open in Web Editor NEW
399.0 399.0 36.0 317.27 MB

PyTorch implementation for the paper "Driving with LLMs: Fusing Object-Level Vector Modality for Explainable Autonomous Driving"

License: Apache License 2.0

Python 100.00%

driving-with-llms's People

Contributors

melights avatar olegsinavski avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

driving-with-llms's Issues

does anyone meet 'Parameter' object has no attribute 'CB'?

Hi, I tried to install the requirement file as the author listed but it first raised the error of bitsandbytes not supporting GPU, I solved this by installing bitsandbytes==0.43.3. After that there is an error of 'Parameter' object has no attribute 'CB'. Does anyone have solution?

View Results

thanks for your great work!
I want to view the results after evaluation, where to find the WandB project "llm-driver"?

Thanks for your work.I have a question. Since the prompt in the model's input already describes the content represented by the vectors, why is it necessary to align the vectors with the LLM during the pre-training process? Are the vectors used to help the model further understand the driving scenario based on the prompt? Are the labels in the pre-training process the prompts generated by lanGen? What is the purpose of the 100k question-answer pairs in the pre-training process?

Could you explain what the model's input and the corresponding labels are during the pre-training stage to help me better understand this process?

real vehicles

Hi,Thanks for your great work! I would like to inquire, based on your experience and research, do you believe it is feasible to deploy this work onto real vehicles? Particularly, considering our current computational resources are 2 A800 GPUs (80G). In this scenario, how long do you think it might take us to achieve this goal? Are there any key technical challenges or issues that need to be addressed?

decapoda-research/llama-7b-hf

decapoda-research/llama-7b-hf is no longer accessible on HuggingFace。
I have a bug with “baffo32/decapoda-research-llama-7B-hf”:
ValueError: The device_map provided does not give any device for the following parameters: base_model.model.weighted_mask

decapoda-research/llama-7b-hf/ Not Found

Thank you for your very nice work!
I was trying to run the program and realized that the pre-trained model is not downloadable from hugging face
decapoda-research/llama-7b-hf/
Can you provide an alternative link for that model?

The environment

I'm grateful for your contributing work. Could you please introduce your setup like CUDA and cudnn? I have tried many times to run the code 'pip install -r requirements.txt.lock' but there are many different problems and errors.

Dataset and LLM

Hello, your work is very exciting. First of all, I would like to know if your model contains a control signal output. Does the dataset contain a control signal? Second, if I want to replace the LLM with the pre-trained model of the first stage, how can I do it?

Training with real world datasets

Hello, congrats on your work!

I was curious to know if you thought about including the real-world datasets as part of fine-tuning the LLM for decision making?

Paper citations

It is very related to your paper being ICRA 2024 cited, I hope you can have more and better results, but I don't seem to find the citations of this paper on the Internet, all of them are arxiv, do you have a citation in the bib file format of the ICRA2024,Thanks again!

The inference generation is very slow

The inference process is currently quite slow. Are there any methods available to accelerate it?
For action task, it costs about 9s for a sample.

Are Driving with LLMs and LingoQA related?

Hello! Amazing work, really.
I was wondering whether this work and the consecutive one (LingoQA) are comparable and if it a real improvement of this work or just another approach.
Since it seems more of a driving scenario commentator and less as a control model that predicts future control signals.
Thank you!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.