bpan's picture
Update README.md
b031bff verified
metadata
license: mit
tags:
  - LLM Agent
  - vision-language navigation
  - synthetic data

LangNav: Language as a Perceptual Representation for Navigation

About LangNav

LangNav is an LLM-based navigation agent which performs multi-step navigation end-to-end via textual descriptions of the scene. The language-based perceptual representation makes LangNav more data efficient compared to VL models. With only a few language-based trajectories from a R2R environment, we use GPT-4 to efficiently generate a huge amount of synthetic training data. A smaller language model (LLaMA2-7B) can then be trained on these synthetic data and do the task. In this repo, we provide the inference code, the model, and the training dataset we used for the paper:

LangNav: Language as a Perceptual Representation for Navigation

Bowen Pan, Rameswar Panda, SouYoung Jin, Rogerio Feris, Aude Oliva, Phillip Isola, Yoon Kim

NAACL 2024 (Findings)

[Paper][GitHub][MIT News]

Prerequisites

We don't have to install the Matterport3D Simulator as we have pre-extracted the caption of each viewpoint.

But we still need to prepare the data in directories

Install the Pytorch-Transformers.

Multi-step Navigation with Language-based Representation

Evaluate our LangNav-Sim2k-Llama2 model on the R2R datasets.

sh eval_scripts/eval_langnav_2k_synthetic_100_real.sh

We will also release the synthetic training dataset and the other models. Stay tuned!

Citation

If you use or discuss our LangNav, please cite our paper:

@article{pan2023langnav,
  title={Langnav: Language as a perceptual representation for navigation},
  author={Pan, Bowen and Panda, Rameswar and Jin, SouYoung and Feris, Rogerio and Oliva, Aude and Isola, Phillip and Kim, Yoon},
  journal={arXiv preprint arXiv:2310.07889},
  year={2023}
}