: Open Source Project Class, Dongguk University, Seoul
- Team Leader : JunKyu Park
- Team Members : Chanyoung Lee, Eunseo Kim, Junghyun Kim
This project leverages robotics, deep learning-based crack detection, and 3D mapping technologies to automate indoor inspections and present results through an intuitive, interactive, web-based 3D map viewer. The robot collects data using high-resolution cameras and sensors, while a deep learning model ensures fast and accurate detection of cracks and defects. The collected data is processed into a high-quality 3D point cloud map using COLMAP, Gaussian splatting, and 3D rendering techniques. These results are visualized on a web platform, enabling real-time inspection and efficient maintenance planning. This technology is also applicable to various industries, including construction, facility management, and disaster response.
Please set up an environment identical to the one provided by the two GitHub pages below.
input
- Output of Colmap from the image stream robot takes
- Output of 2D Crack Detection model ( JSON format )
Click to toggle contents of detail JSON Format
{
"0.85": [
{
"filled_coordinates": [
[331, 434],
[331, 435],
[331, 436],
[331, 437],
[331, 438],
[331, 439],
[331, 440],
[331, 441],
[331, 442]
]
}
]
}
Image stream from handheld Camera ( iphone 16-pro )
By implementing the transformations between world-to-camera and camera-to-image spaces, I gained a clearer understanding of how 3D points are projected onto a 2D image using camera parameters.
While modifying the Gaussian Splatting code, I could observe how these transformations directly impact the rendering results, bridging the gap between theory and practice.
This iterative process of code modification and result validation significantly improved my intuitive understanding of the principles and their implementation.