Skip to content

CrystalSixone/SEAT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Title

Enhanced Language-guided Robot Navigation with Panoramic Semantic Depth Perception and Cross-modal Fusion

Steps

1. Requirements and Installation

  1. Please install MatterPort3D simulator from here.
  2. Install requirements:

Please ensure to use the specified version, as discrepancies between versions can result in errors.

pip install -r requirements.txt
  1. Download Datasets, features, and models:

    1. Annotations from here.
    2. Features and trained weights (for both pre-trained and fine-tuned) from here.
    3. Download METER(Optional, only if you want to pre-train SEAT based on METER)from here. Our used meter model is meter_clip16_224_roberta_pretrain.ckpt.
    4. Download EnvEdit weights from here.
    5. For some reason, it may not be possible to access huggingface's model directly, especially when calling roberta's tokenizer. In this case, I recommend going directly to huggingface's official website to download the required file to the local like datasets/pretrained/roberta.

    The structure should be as follows (using R2R as a detailed example):

    datasets
    ├── R2R
    │   ├── annotations
    │   ├── features
    │   ├── navigator
    │   ├── pretrain
    │   └── id_paths.json
    ├── REVERIE
    ├── SOON
    ├── RQ
    └── EnvEdit
    

2. Pre-train

Use the following command to pre-train:

cd rq_train
bash run_rq.sh

cd pretrain_src
bash run_r2r_seat.sh

3. Fine-tune

Use the following command to fine-tune:

cd map_nav_src
bash scripts/run_r2r.sh

4. Valid

Use the following command to valid:

cd map_nav_src
bash scripts/run_r2r_valid.sh

Acknowledge

We thank to the authors for their awesome work and sharing:

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published