unsupervised-depth-completion-visual-inertial-odometry VS xivo

Compare unsupervised-depth-completion-visual-inertial-odometry vs xivo and see what are their differences.

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
unsupervised-depth-completion-visual-inertial-odometry xivo
2 2
183 828
- 0.0%
5.0 0.0
10 months ago about 1 year ago
Python C++
GNU General Public License v3.0 or later GNU General Public License v3.0 or later
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

unsupervised-depth-completion-visual-inertial-odometry

Posts with mentions or reviews of unsupervised-depth-completion-visual-inertial-odometry. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2021-08-30.
  • Unsupervised Depth Completion from Visual Inertial Odometry
    3 projects | news.ycombinator.com | 30 Aug 2021
    Hey there, interested in camera and range sensor fusion for point cloud (depth) completion?

    Here is an extended version of our [talk](https://www.youtube.com/watch?v=oBCKO4TH5y0) at ICRA 2020 where we do a step by step walkthrough of our paper Unsupervised Depth Completion from Visual Inertial Odometry (joint work with Fei Xiaohan, Stephanie Tsuei, and Stefano Soatto).

    In this talk, we present an unsupervised method (no need for human supervision/annotations) for learning to recover dense point clouds from images, captured by cameras, and sparse point clouds, produced by lidar or tracked by visual inertial odometry (VIO) systems. To illustrate what I mean, here is an [example](https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry/blob/master/figures/void_teaser.gif?raw=true) of the point clouds produced by our method.

    Our method is light-weight (so you can run it on your computer!) and is built on top of [XIVO] (https://github.com/ucla-vision/xivo) our VIO system.

    For those interested here are links to the [paper](https://arxiv.org/pdf/1905.08616.pdf), [code](https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry) and the [dataset](https://github.com/alexklwong/void-dataset) we collected.

  • [N][R] ICRA 2020 extended talk for Unsupervised Depth Completion from Visual Inertial Odometry
    4 projects | /r/MachineLearning | 30 Aug 2021
    In this talk, we present an unsupervised method (no need for human supervision/annotations) for learning to recover dense point clouds from images, captured by cameras, and sparse point clouds, produced by lidar or tracked by visual inertial odometry (VIO) systems. To illustrate what I mean, you can visit our github page for examples (gifs) of point clouds produced by our method.

xivo

Posts with mentions or reviews of xivo. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2021-08-30.
  • Unsupervised Depth Completion from Visual Inertial Odometry
    3 projects | news.ycombinator.com | 30 Aug 2021
    Hey there, interested in camera and range sensor fusion for point cloud (depth) completion?

    Here is an extended version of our [talk](https://www.youtube.com/watch?v=oBCKO4TH5y0) at ICRA 2020 where we do a step by step walkthrough of our paper Unsupervised Depth Completion from Visual Inertial Odometry (joint work with Fei Xiaohan, Stephanie Tsuei, and Stefano Soatto).

    In this talk, we present an unsupervised method (no need for human supervision/annotations) for learning to recover dense point clouds from images, captured by cameras, and sparse point clouds, produced by lidar or tracked by visual inertial odometry (VIO) systems. To illustrate what I mean, here is an [example](https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry/blob/master/figures/void_teaser.gif?raw=true) of the point clouds produced by our method.

    Our method is light-weight (so you can run it on your computer!) and is built on top of [XIVO] (https://github.com/ucla-vision/xivo) our VIO system.

    For those interested here are links to the [paper](https://arxiv.org/pdf/1905.08616.pdf), [code](https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry) and the [dataset](https://github.com/alexklwong/void-dataset) we collected.

  • [N][R] ICRA 2020 extended talk for Unsupervised Depth Completion from Visual Inertial Odometry
    4 projects | /r/MachineLearning | 30 Aug 2021
    Our method is light-weight (so you can run it on your computer!) and is built on top of XIVO our VIO system.

What are some alternatives?

When comparing unsupervised-depth-completion-visual-inertial-odometry and xivo you can also consider the following projects:

instant-ngp - Instant neural graphics primitives: lightning fast NeRF and more

open_vins - An open source platform for visual-inertial navigation research.

dino - PyTorch code for Vision Transformers training with the Self-Supervised learning method DINO

rtabmap - RTAB-Map library and standalone application

calibrated-backprojection-network - PyTorch Implementation of Unsupervised Depth Completion with Calibrated Backprojection Layers (ORAL, ICCV 2021)

void-dataset - Visual Odometry with Inertial and Depth (VOID) dataset

simclr - SimCLRv2 - Big Self-Supervised Models are Strong Semi-Supervised Learners

r3live - A Robust, Real-time, RGB-colored, LiDAR-Inertial-Visual tightly-coupled state Estimation and mapping package

Open3D - Open3D: A Modern Library for 3D Data Processing

bpycv - Computer vision utils for Blender (generate instance annoatation, depth and 6D pose by one line code)

SuperPoint_SLAM - SuperPoint + ORB_SLAM2