R-VIO
xivo
R-VIO | xivo | |
---|---|---|
1 | 2 | |
718 | 828 | |
1.1% | 0.0% | |
2.2 | 0.0 | |
about 1 year ago | about 1 year ago | |
C++ | C++ | |
GNU General Public License v3.0 only | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
R-VIO
-
Summary of excellent laboratories in the field of SLAM(3)
Huai Z, Huang G. Robocentric visual-inertial odometry[C]//2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2018: 6319-6326. (Github: https://github.com/rpng/R-VIO) Guoquan (Paul) Huang: Home page
xivo
-
Unsupervised Depth Completion from Visual Inertial Odometry
Hey there, interested in camera and range sensor fusion for point cloud (depth) completion?
Here is an extended version of our [talk](https://www.youtube.com/watch?v=oBCKO4TH5y0) at ICRA 2020 where we do a step by step walkthrough of our paper Unsupervised Depth Completion from Visual Inertial Odometry (joint work with Fei Xiaohan, Stephanie Tsuei, and Stefano Soatto).
In this talk, we present an unsupervised method (no need for human supervision/annotations) for learning to recover dense point clouds from images, captured by cameras, and sparse point clouds, produced by lidar or tracked by visual inertial odometry (VIO) systems. To illustrate what I mean, here is an [example](https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry/blob/master/figures/void_teaser.gif?raw=true) of the point clouds produced by our method.
Our method is light-weight (so you can run it on your computer!) and is built on top of [XIVO] (https://github.com/ucla-vision/xivo) our VIO system.
For those interested here are links to the [paper](https://arxiv.org/pdf/1905.08616.pdf), [code](https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry) and the [dataset](https://github.com/alexklwong/void-dataset) we collected.
-
[N][R] ICRA 2020 extended talk for Unsupervised Depth Completion from Visual Inertial Odometry
Our method is light-weight (so you can run it on your computer!) and is built on top of XIVO our VIO system.
What are some alternatives?
msckf_vio - Robust Stereo Visual Inertial Odometry for Fast Autonomous Flight
open_vins - An open source platform for visual-inertial navigation research.
rtabmap - RTAB-Map library and standalone application
VINS-Mono - A Robust and Versatile Monocular Visual-Inertial State Estimator
unsupervised-depth-completion-visual-inertial-odometry - Tensorflow and PyTorch implementation of Unsupervised Depth Completion from Visual Inertial Odometry (in RA-L January 2020 & ICRA 2020)
void-dataset - Visual Odometry with Inertial and Depth (VOID) dataset
r3live - A Robust, Real-time, RGB-colored, LiDAR-Inertial-Visual tightly-coupled state Estimation and mapping package
Open3D - Open3D: A Modern Library for 3D Data Processing
SuperPoint_SLAM - SuperPoint + ORB_SLAM2
vortex-auv - Software for guidance, navigation and control for the Vortex AUVs. Purpose built for competing in AUV/ROV competitions.