Skip to content

Transform 2D images into 3D worlds using deep learning. This small project explores monocular depth estimation and 3D point cloud generation from a single image.

Notifications You must be signed in to change notification settings

Joelcic/DL-DepthEstimation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Depth estimization

This project demonstrates how to estimate depth maps from 2D RGB images and transform them into 3D point clouds using modern deep learning models. It leverages pretrained models for monocular depth estimation (e.g., MiDaS, DPT, Depth-Anything) to infer spatial geometry from single images.

Overview

The pipeline takes a 2D image as input and performs the following steps:

  1. Depth Estimation — A pretrained deep learning model predicts a dense depth map.
  2. Depth Visualization — The predicted depth is normalized and colorized for analysis.
  3. 3D Point Cloud Generation — The depth map is projected into 3D space to reconstruct the scene geometry.
  4. (Optional) Mesh reconstruction and visualization with Open3D.

Examples:

As illustrated the deep learning model predicts a dense depth map okay.

Left: Original RGB image
Right: Estimated depth map

Transformed into a 3D pointclod:

Left: Original point cloud
Right: Colored point cloud

Files

  • utils.py: holds subfunctions.
  • computer_vision.py: holds functions for transforming from 2D to 3D.
  • main.py: holds the pipeline.

About

Transform 2D images into 3D worlds using deep learning. This small project explores monocular depth estimation and 3D point cloud generation from a single image.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages