monodepth2
torchdyn
Our great sponsors
monodepth2 | torchdyn | |
---|---|---|
6 | 1 | |
3,974 | 1,272 | |
1.5% | 3.4% | |
0.0 | 5.2 | |
7 months ago | about 1 month ago | |
Jupyter Notebook | Jupyter Notebook | |
GNU General Public License v3.0 or later | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
monodepth2
- Calculation of an absolute depth map from multiple images or videos.
-
Easy to train a monocular (self) supervised depth estimation model?
I've used monodepth2 before and it's great: https://github.com/nianticlabs/monodepth2
-
Sources: Pixel 6 Pro was supposed to launch with face unlock
How can a single camera do that? My experience with computer vision is fairly limited so I'm curious how that would work. My understanding is you need to be able to generate a point map, stereo vision, or some non-CV related method , e.g. radar like the pixel 4. 2D depth estimation can be done with a single camera in somewhat useful way but it's not a secure way (https://github.com/nianticlabs/monodepth2 -- now somewhat similar functionality in OpenCV). Can you expand on what AI the single camera is being combined with that provides security guarantees?
- Can anyone explain the following github code to me. Also it’s my first time using GitHub so I’m completely lost.
-
Estimating camera height, orientation and field of view from a single monocular image.
I suspect you may have the best success by using monocular depth approaches (for example something like this: https://github.com/nianticlabs/monodepth2).
-
Looking for a fast monocular depth estimation library to use in a Rust project.
After that I have to do the same for Python I think, and then I have to find out how to figure out how to use a library like https://github.com/ialhashim/DenseDepth or https://github.com/nianticlabs/monodepth2 for that GStreamer plugin (or element, still trying to grasp the terminology here)
torchdyn
What are some alternatives?
DenseDepth - High Quality Monocular Depth Estimation via Transfer Learning
torchsde - Differentiable SDE solvers with GPU support and efficient sensitivity analysis.
packnet-sfm - TRI-ML Monocular Depth Estimation Repository
NeuralCDE - Code for "Neural Controlled Differential Equations for Irregular Time Series" (Neurips 2020 Spotlight)
cs231n - Note and Assignments for CS231n: Convolutional Neural Networks for Visual Recognition
handwritten-multi-digit-number-recognition - Recognize handwritten multi-digit numbers using a CRNN model trained with synthetic data.
ZoeDepth - Metric depth estimation from a single image
deep-learning-v2-pytorch - Projects and exercises for the latest Deep Learning ND program https://www.udacity.com/course/deep-learning-nanodegree--nd101
glasses - High-quality Neural Networks for Computer Vision 😎
hyperlearn - 2-2000x faster ML algos, 50% less memory usage, works on all hardware - new and old.
depth-estimate-gui - Depth Estimate GUI - Windows, Mac, Linux
BigDL - Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.