OpenChisel
colmap
OpenChisel | colmap | |
---|---|---|
1 | 28 | |
454 | 6,825 | |
0.4% | 3.1% | |
0.0 | 9.2 | |
about 1 month ago | 7 days ago | |
C++ | C++ | |
- | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
OpenChisel
-
Questions for SLAM/SfM for Dense 3D Reconstruction (DSO vs ORB, Monofusion etc.)
For instance you could go with : https://github.com/ov2slam/ov2slam , add some processing on the keyframes for depth maps computation and then fuse the depth maps in a TSDF using https://github.com/personalrobotics/OpenChisel or https://github.com/ethz-asl/voxblox
colmap
- Magic123: One Image to High-Quality 3D Object Generation
-
Drone mapping is pretty dang cool
Not saying its easy to use, but there is an application gui and it is free: https://github.com/colmap/colmap
-
Import many photogrammetry software's scenes into Blender
Colmap (Model folders (BIN and TXT), dense workspaces, NVM, PLY)
- Best options for monocular reconstruction?
-
improving camera pose estimation using multiple aruco markers
See colmap for example https://colmap.github.io/
-
2D images to 3D Object reconstruction
You're looking into a problem called photogrammetry, and a well-studied one at that. I'd recommend looking into "shape from motion" (sfm); specifically techniques that do "dense reconstruction." I'd recommend COLMAP to start with. It does pose estimation from images (e.g. you point it at a bunch of images and it will figure out the relative poses of the cameras that took them), as well as sparse and dense reconstcution.
-
Framework generate 3d meshes from camera images
COLMAP builds dense meshes from a collection of cameras https://colmap.github.io/
- Nerfstudio: A collaboration friendly studio for NeRFs
-
Neural Radiance Fields and input shape
I’ve seen references to using COLMAP (https://colmap.github.io/) to estimate camera position/pose, e.g. here
-
3D reconstruction of an object from videos/few images
Classical photogrammetry, where I agree with u/tdgros that the way to go is https://colmap.github.io/. There are actually better variants in literature but nothing is more reliable and user-friendly than COLMAP. This will give you a very precise point cloud, that can be meshed if needed.
What are some alternatives?
pixel-perfect-sfm - Pixel-Perfect Structure-from-Motion with Featuremetric Refinement (ICCV 2021, Best Student Paper Award)
Meshroom - 3D Reconstruction Software
ov2slam - OV²SLAM is a Fully Online and Versatile Visual SLAM for Real-Time Applications
OpenMVG (open Multiple View Geometry) - open Multiple View Geometry library. Basis for 3D computer vision and Structure from Motion.
voxblox - A library for flexible voxel-based mapping, mainly focusing on truncated and Euclidean signed distance fields.
Hierarchical-Localization - Visual localization made easy with hloc
pyslam - pySLAM contains a monocular Visual Odometry (VO) pipeline in Python. It supports many modern local features based on Deep Learning.
nerf - Code release for NeRF (Neural Radiance Fields)
MonoRec - Official implementation of the paper: MonoRec: Semi-Supervised Dense Reconstruction in Dynamic Environments from a Single Moving Camera (CVPR 2021)
openMVS - open Multi-View Stereo reconstruction library
instant-ngp - Instant neural graphics primitives: lightning fast NeRF and more
OpenSfM - Open source Structure-from-Motion pipeline