{ "id": "2403.11577", "version": "v2", "published": "2024-03-18T08:53:03.000Z", "updated": "2024-09-18T08:22:57.000Z", "title": "3DGS-Calib: 3D Gaussian Splatting for Multimodal SpatioTemporal Calibration", "authors": [ "Quentin Herau", "Moussab Bennehar", "Arthur Moreau", "Nathan Piasco", "Luis Roldao", "Dzmitry Tsishkou", "Cyrille Migniot", "Pascal Vasseur", "Cédric Demonceaux" ], "comment": "Accepted at IROS 2024 (Oral presentation). Project page: https://qherau.github.io/3DGS-Calib/", "categories": [ "cs.CV", "cs.RO" ], "abstract": "Reliable multimodal sensor fusion algorithms require accurate spatiotemporal calibration. Recently, targetless calibration techniques based on implicit neural representations have proven to provide precise and robust results. Nevertheless, such methods are inherently slow to train given the high computational overhead caused by the large number of sampled points required for volume rendering. With the recent introduction of 3D Gaussian Splatting as a faster alternative to implicit representation methods, we propose to leverage this new rendering approach to achieve faster multi-sensor calibration. We introduce 3DGS-Calib, a new calibration method that relies on the speed and rendering accuracy of 3D Gaussian Splatting to achieve multimodal spatiotemporal calibration that is accurate, robust, and with a substantial speed-up compared to methods relying on implicit neural representations. We demonstrate the superiority of our proposal with experimental results on sequences from KITTI-360, a widely used driving dataset.", "revisions": [ { "version": "v2", "updated": "2024-09-18T08:22:57.000Z" } ], "analyses": { "keywords": [ "3d gaussian splatting", "implicit neural representations", "3dgs-calib", "achieve faster multi-sensor calibration", "reliable multimodal sensor fusion algorithms" ], "tags": [ "github project" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }