Efficient Occlusion-Aware Navigation for Air-Ground Robot in Dynamic Environments via State Space Model
- OMEGA (RA-L 2024.12): The First AGR-Tailored Dynamic Navigation System.
- HE-Nav (RA-L 2024.09): The First AGR-Tailored ESDF-Free Navigation System.
- AGRNav (ICRA 2024.01): The First AGR-Tailored Occlusion-Aware Navigation System.
- AMOV Lab Research Scholarship -- 2024.11: 5000 RMB
- AMOV Lab Research Scholarship -- 2024.10: 5000 RMB
- [03/07/2024]: OMEGA's simulation logs are available for download:
- [01/07/2024]: OccMamba's test and evaluation logs are available for download:
OccMamba Results | Experiment Log |
---|---|
OccMamba on the SemanticKITTI hidden official test dataset | link |
OccMamba test log | link |
OccMamba evaluation log | link |
- [28/06/2024]: The pre-trained model can be downloaded at OneDrive
- [25/06/2024]: We have released the code for OccMamba, a key component of OMEGA!
OMEGA emerges as the pioneering navigation system tailored for AGRs in dynamic settings, with a focus on ensuring occlusion-free mapping and pathfinding. It incorporates OccMamba, a module designed to process point clouds and perpetually update local maps, thereby preemptively identifying obstacles within occluded areas. Complementing this, AGR-Planner utilizes up-to-date maps to facilitate efficient and effective route planning, seamlessly navigating through dynamic environments.
@article{wang2025omega,
author={Wang, Junming and Guan, Xiuxian and Sun, Zekai and Shen, Tianxiang and Huang, Dong and Liu, Fangming and Cui, Heming},
journal={IEEE Robotics and Automation Letters},
title={OMEGA: Efficient Occlusion-Aware Navigation for Air-Ground Robots in Dynamic Environments via State Space Model},
year={2025},
volume={10},
number={2},
pages={1066-1073},
publisher={IEEE}
}
Please kindly star ⭐️ this project if it helps you. We take great efforts to develop and maintain it 😁.
Hardware | Link |
---|---|
AMOV Lab P600 UAV | link |
AMOV Lab Allapark1-Jetson Xavier NX | link |
Wheeltec R550 ROS Car | link |
Intel RealSense D435i | link |
Intel RealSense T265 | link |
TFmini Plus | link |
❗ Considering that visual positioning is prone to drift in the Z-axis direction, we added TFmini Plus for height measurement. Additionally, GNSS-RTK positioning is recommended for better localization accuracy.
🤑 Our customized Aerial-Ground Robot cost about RMB 70,000.
conda create -n occmamba python=3.10 -y
conda install pytorch==2.1.0 torchvision==0.16.0 torchaudio==2.1.0 pytorch-cuda=12.1 -c pytorch -c nvidia -y
pip install spconv-cu120
pip install tensorboardX
pip install dropblock
pip install pyg_lib torch_scatter torch_sparse torch_cluster torch_spline_conv -f https://data.pyg.org/whl/torch-2.1.0+cu121.html
pip install -U openmim
mim install mmcv-full
pip install mmcls==0.25.0
Note
Please refer to Vision-Mamba for more installation information.
- SemanticKITTI
Please download the Semantic Scene Completion dataset (v1.1) from the SemanticKITTI website and extract it.
Or you can use voxelizer to generate ground truths of semantic scene completion.
The dataset folder should be organized as follows.
SemanticKITTI
├── dataset
│ ├── sequences
│ │ ├── 00
│ │ │ ├── labels
│ │ │ ├── velodyne
│ │ │ ├── voxels
│ │ │ ├── [OTHER FILES OR FOLDERS]
│ │ ├── 01
│ │ ├── ... ...
Clone the repository:
git clone https://github.com/jmwang0117/Occ-Mamba.git
$ cd <root dir of this repo>
$ bash run_train.sh
$ cd <root dir of this repo>
$ bash run_val.sh
Since SemantiKITTI contains a hidden test set, we provide test routine to save predicted output in same format of SemantiKITTI, which can be compressed and uploaded to the SemanticKITTI Semantic Scene Completion Benchmark. You can provide which checkpoints you want to use for testing. We used the ones that performed best on the validation set during training. For testing, you can use the following command.
$ cd <root dir of this repo>
$ bash run_test.sh
Many thanks to these excellent open source projects: