IIIT-Delhi Institutional Repository

Multi-modal sensor fusion for geometric and semantic scene understanding with graphs

Show simple item record

dc.contributor.author Rai, Prakhar
dc.contributor.author Iyer, Anirudh R
dc.contributor.author Anand, Saket (Advisor)
dc.contributor.author Kaul, Sanjit Krishnan (Advisor)
dc.date.accessioned 2024-05-16T12:38:58Z
dc.date.available 2024-05-16T12:38:58Z
dc.date.issued 2023-11-29
dc.identifier.uri http://repository.iiitd.edu.in/xmlui/handle/123456789/1496
dc.description.abstract Multimodal sensor fusion is a technique which combines data from multiple sensors for applying the combined data in an effective manner which would provide a better understandning of the environment. In some cases such as autonomous driving, sensor fusion of multiple cameras and Lidar could give us semantic and geometric context for applications like localisation of the ego vehicle. Graph registration and smoothing of Multimodal sensors can also be used for tracking semantic and geometric objects around an ego vehicle which could give us more information of environment and of the ego vehicle, for example: we can deduce the velocity of the ego vehicle relative to stationary planes which are inferred through the fused and registered graphs. In this project, we propose a new approach using graph-based representation of semantic and geometric instances combined with Kalman filtering over time series of graphs. en_US
dc.language.iso en_US en_US
dc.publisher IIIT-Delhi en_US
dc.subject Multi-modal sensor fusion en_US
dc.subject Graph registration en_US
dc.subject Scene understanding en_US
dc.subject Autonomous driving en_US
dc.title Multi-modal sensor fusion for geometric and semantic scene understanding with graphs en_US
dc.type Other en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Repository


Advanced Search

Browse

My Account