dc.contributor.author |
Rai, Prakhar |
|
dc.contributor.author |
Iyer, Anirudh R |
|
dc.contributor.author |
Anand, Saket (Advisor) |
|
dc.contributor.author |
Kaul, Sanjit Krishnan (Advisor) |
|
dc.date.accessioned |
2024-05-16T12:38:58Z |
|
dc.date.available |
2024-05-16T12:38:58Z |
|
dc.date.issued |
2023-11-29 |
|
dc.identifier.uri |
http://repository.iiitd.edu.in/xmlui/handle/123456789/1496 |
|
dc.description.abstract |
Multimodal sensor fusion is a technique which combines data from multiple sensors for applying the combined data in an effective manner which would provide a better understandning of the environment. In some cases such as autonomous driving, sensor fusion of multiple cameras and Lidar could give us semantic and geometric context for applications like localisation of the ego vehicle. Graph registration and smoothing of Multimodal sensors can also be used for tracking semantic and geometric objects around an ego vehicle which could give us more information of environment and of the ego vehicle, for example: we can deduce the velocity of the ego vehicle relative to stationary planes which are inferred through the fused and registered graphs. In this project, we propose a new approach using graph-based representation of semantic and geometric instances combined with Kalman filtering over time series of graphs. |
en_US |
dc.language.iso |
en_US |
en_US |
dc.publisher |
IIIT-Delhi |
en_US |
dc.subject |
Multi-modal sensor fusion |
en_US |
dc.subject |
Graph registration |
en_US |
dc.subject |
Scene understanding |
en_US |
dc.subject |
Autonomous driving |
en_US |
dc.title |
Multi-modal sensor fusion for geometric and semantic scene understanding with graphs |
en_US |
dc.type |
Other |
en_US |