torchdyn
monodepth2
torchdyn | monodepth2 | |
---|---|---|
1 | 6 | |
1,277 | 3,974 | |
3.8% | 1.5% | |
5.2 | 0.0 | |
about 1 month ago | 7 months ago | |
Jupyter Notebook | Jupyter Notebook | |
Apache License 2.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
torchdyn
monodepth2
- Calculation of an absolute depth map from multiple images or videos.
-
Easy to train a monocular (self) supervised depth estimation model?
I've used monodepth2 before and it's great: https://github.com/nianticlabs/monodepth2
-
Sources: Pixel 6 Pro was supposed to launch with face unlock
How can a single camera do that? My experience with computer vision is fairly limited so I'm curious how that would work. My understanding is you need to be able to generate a point map, stereo vision, or some non-CV related method , e.g. radar like the pixel 4. 2D depth estimation can be done with a single camera in somewhat useful way but it's not a secure way (https://github.com/nianticlabs/monodepth2 -- now somewhat similar functionality in OpenCV). Can you expand on what AI the single camera is being combined with that provides security guarantees?
- Can anyone explain the following github code to me. Also it’s my first time using GitHub so I’m completely lost.
-
Estimating camera height, orientation and field of view from a single monocular image.
I suspect you may have the best success by using monocular depth approaches (for example something like this: https://github.com/nianticlabs/monodepth2).
-
Looking for a fast monocular depth estimation library to use in a Rust project.
After that I have to do the same for Python I think, and then I have to find out how to figure out how to use a library like https://github.com/ialhashim/DenseDepth or https://github.com/nianticlabs/monodepth2 for that GStreamer plugin (or element, still trying to grasp the terminology here)
What are some alternatives?
torchsde - Differentiable SDE solvers with GPU support and efficient sensitivity analysis.
DenseDepth - High Quality Monocular Depth Estimation via Transfer Learning
NeuralCDE - Code for "Neural Controlled Differential Equations for Irregular Time Series" (Neurips 2020 Spotlight)
packnet-sfm - TRI-ML Monocular Depth Estimation Repository
handwritten-multi-digit-number-recognition - Recognize handwritten multi-digit numbers using a CRNN model trained with synthetic data.
cs231n - Note and Assignments for CS231n: Convolutional Neural Networks for Visual Recognition
deep-learning-v2-pytorch - Projects and exercises for the latest Deep Learning ND program https://www.udacity.com/course/deep-learning-nanodegree--nd101
ZoeDepth - Metric depth estimation from a single image
hyperlearn - 2-2000x faster ML algos, 50% less memory usage, works on all hardware - new and old.
glasses - High-quality Neural Networks for Computer Vision 😎
BigDL - Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, etc.
depth-estimate-gui - Depth Estimate GUI - Windows, Mac, Linux