SLAM (Simultaneous Localization and Mapping) It is one of the important techniques used in robotics and computer vision. It helps machines understand where they are and create a map of their environment. Motion blurred images face difficulties in dense visual SLAM systems for two reasons: 1) Inaccurate pose estimation during tracking: current dense and photorealistic image BANG The algorithms rely on clear images to estimate camera positions ensuring consistent brightness across all views. This affects the mapping process and results in inconsistent multi-view geometry. 2) Inconsistent multi-view geometry in mapping: Poor multi-view image quality can result in incorrect features, causing errors in 3D geometry and a low-quality reconstruction of the 3D map. Combining these two factors, existing dense virtual SLAM systems would typically perform poorly in handling motion-blurred images.
Traditional sparse SLAM methods use sparse point clouds for map reconstruction. Recent learning-based dense SLAM systems focus on generating dense maps important for downstream tasks. Neural Radiation Fields (NeRF) and 3D Gaussian Splash (3DGS) They have been used with SLAM systems to create realistic 3D scenes, improving map quality and texture. However, existing methods rely heavily on sharp, high-quality images. RGB-D inputs, which pose challenges when dealing with motion blurred frames, which are often encountered in low light or long exposure conditions, resulting in low localization and mapping accuracy and efficiency in various methods.
To solve these problems, a group of researchers from China conducted detailed research and proposed MBA-SLAMa dense, photorealistic RGB-D SLAM channel designed to handle motion blur inputs effectively. This approach integrates the physical motion blur imaging process into the tracking and mapping stages. The main goal of this framework is to reconstruct dense, high-quality 3D scenes and accurately measure camera motion trajectories, which was achieved by integrating two key components: a motion blur tracker and a blur mapper. adjusted per package based on NERF either 3D Gaussian Splash.
The method used a continuous motion model to track camera movement during the exposure. The system considered the initial and final camera positions for each motion-blurred image. At tracking, a sharp reference image was rendered, blurred to match the current image, and compared to improve motion estimation. Camera trajectories and 3D scenes were optimized in mapping to reduce image matching errors. Two scene representations were explored: implicit neural radiation fields (NeRF) and explicit 3D Gaussian splatters (3D-GS). NERF achieved higher frame rates but lower render quality, while 3D-GS It offered better quality at the cost of lower frame rates.
The method showed a measured reduction in tracking errors, with the ScanNet data set yielding an ATE RMSE of 0.053surpassing ORB-SLAM3 (0.081) and LDS-SLAM (0.071). in it TUM RGB-D dataset, MBA-SLAM achieved an ATE RMSE of 0.062showing its superior tracking accuracy. In image reconstruction, MBA-SLAM stood out with a PSNR of 31.2dB on the ArchViz dataset and an SSIM of 0.96 in ScanNetoutperforming methods such as ORB-SLAM3 and DSO in terms of quality. He LPIPS score of MBA-SLAM It is also reported that it is 0.18which reflects a better perceptual quality. Radiation fields and Gaussian spatter improved image quality, while CUDA The acceleration enabled real-time processing, making it 5 times faster than others. MBA SLAM provided improved tracking accuracy, better image quality and speed compared to others, and seemed to promise application in SLAM scenarios with motion blur due to dynamism in the environment.
In summary, the proposed framework MBA-SLAM effectively addresses the problems in the BANG system. With its physical motion blur imaging model, highly CUDAWith an optimized blur-aware tracker and a blur mapper, the MBA-SLAM tracked precise trajectories of camera motion within the exposure time and reconstructed a sharp, photorealistic map for the given video sequence input. It performed much better than previous methods on existing and real-world data sets. This work marks a significant development in the field of SLAM systems and can be used as a basis for future advances and research!
Verify he Paper and GitHub. All credit for this research goes to the researchers of this project. Also, don't forget to follow us on <a target="_blank" href="https://twitter.com/Marktechpost”>twitter and join our Telegram channel and LinkedIn Grabove. If you like our work, you will love our information sheet.. Don't forget to join our SubReddit over 55,000ml.
Divyesh is a Consulting Intern at Marktechpost. He is pursuing a BTech in Agricultural and Food Engineering from the Indian Institute of technology Kharagpur. He is a data science and machine learning enthusiast who wants to integrate these leading technologies in agriculture and solve challenges.
<script async src="//platform.twitter.com/widgets.js” charset=”utf-8″>