Skip to content

A General Simultaneous Localization and Mapping Framework which supports feature based or direct method and different sensors including monocular camera, RGB-D sensors or any other input types can be handled.

License

Notifications You must be signed in to change notification settings

zdzhaoyong/GSLAM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

60 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GSLAM (A General SLAM Framework and BenchMark)

Build Status License Version

1. Introduction

If you use this code for your research, please cite our paper GSLAM: A General SLAM Framework and Benchmark:

@inproceedings{gslamICCV2019,
	title={GSLAM: A General SLAM Framework and Benchmark},
	author={Yong Zhao, Shibiao Xu, Shuhui Bu, Hongkai Jiang, Pengcheng Han},
	booktitle={Proceedings of the IEEE International Conference on Computer Vision},
	year={2019},
	organization={IEEE}
}
@article{gslam2019,
	title={GSLAM: A General SLAM Framework and Benchmark},
	author={Yong Zhao, Shibiao Xu, Shuhui Bu, Hongkai Jiang, Pengcheng Han},
	journal={arXiv:1902.07995 },
	year={2019}
}

1.1. What is GSLAM?

  1. For SLAM developers : Everyone can develop their own SLAM implementation based on GSLAM and publish it as a plugin with open-source or not.
  2. For SLAM users : Applications are able to use different SLAM plugins with the same API without recompilation and implementations are loaded at runtime.

2. Documentation

Documentation: https://zdzhaoyong.github.io/GSLAM/

3. Start with GSLAM

3.1. Basic usages

Test modules with google test:

gslam tests --gtest_filter=*

Run a slam system with datasets:

gslam qviz orbslam play -dataset dataset_path

3.2. Supported Datasets

GSLAM now implemented serveral plugins for public available datasets. It is very easy to play different datasets with parameter "Dataset" setted:

# Play kitti with monocular mode
gslam qviz play -dataset <dataset_path>/odomentry/color/00/mono.kitti

# Play kitti with stereo mode
gslam qviz play -dataset <dataset_path>/odomentry/color/00/stereo.kitti

# Play TUM RGBD Dataset (associate.txt file prepared)
gslam qviz play -dataset <dataset_path>/rgbd_dataset_freiburg1_360/.tumrgbd

# Play TUM Monocular (images unziped)
gslam qviz play -dataset <dataset_path>/calib_narrowGamma_scene1/.tummono

# Play EuRoC Dataset with IMU frames
gslam qviz play -dataset <dataset_path>/EuRoC/MH_01_easy/mav0/.euroc

# Play NPU DroneMap Dataset
gslam qviz play -dataset <dataset_path>/DroneMap/phantom3-village/phantom3-village-kfs/.npudronemap
gslam qviz play -dataset <dataset_path>/DroneMap/phantom3-village/phantom3-village-unified/.npudronemap

The datasets are default to be played on realtime, and the play speed can be controled with "playspeed":

gslam qviz play -dataset <dataset_path>/odomentry/color/00/mono.kitti -playspeed 2.

The following dataset plugins are now implemented:

Name     Channels Description
KITTI Stereo,Pose
TUMMono Monocular
TUMRGBD RGBD,Pose
EuRoc IMU,Stereo
NPUDroneMap GPS,Monocular
CVMono Monocular Online camera or video dataset using opencv.

Users can also implement dataset plugins by own.

3.3. Implemented SLAM plugins

Name       ScreenShot Description
DSO code
ORBSLAM code
SVO code
TheiaSfM code

4. Contacts

YongZhao: [email protected]

ShuhuiBu: [email protected]

ShibiaoXu: [email protected]

5. License

The GSLAM library is licensed under the BSD license. Note that this text refers only to the license for GSLAM itself, independent of its optional dependencies, which are separately licensed. Building GSLAM with these optional dependencies may affect the resulting GSLAM license.

Copyright (c) 2018 Northwestern Polytechnical University, Yong Zhao. All rights reserved.

This software was developed by the Yong Zhao at Northwestern Polytechnical University.

All advertising materials mentioning features or use of this software must display
the following acknowledgement: This product includes software developed by Northwestern Polytechnical University, PILAB.

Redistribution and use in source and binary forms, with or without modification,
are permitted provided that the following conditions are met:

1. Redistributions of source code must retain the above copyright notice, this list
of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright notice, this
list of conditions and the following disclaimer in the documentation and/or
other materials provided with the distribution.
3. All advertising materials mentioning features or use of this software must
display the following acknowledgement: This product includes software developed
by Northwestern Polytechnical University and its contributors.
4. Neither the name of the University nor the names of its contributors may be
used to endorse or promote products derived from this software without specific
prior written permission.

THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

About

A General Simultaneous Localization and Mapping Framework which supports feature based or direct method and different sensors including monocular camera, RGB-D sensors or any other input types can be handled.

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •