Overview
Robots equipped with rich sensor suites can localize reliably in partially-observable environments, but powering every sensor continuously is wasteful and often infeasible. Belief-space planners address this by propagating pose-belief covariance through analytic models and switching sensors heuristically–a brittle, runtime-expensive approach. Data-driven approaches–including diffusion models–learn multi-modal trajectories from demonstrations, but presuppose an accurate, always-on state estimate.
Research Goals
- Choose a minimal subset of sensors
 - Create semantic maps for any available sensor
 - Generate a sensor agnostic map representation
 - Localize and Navigate with minimal sensing capabilities
 
Related Publications
- Puthumanaillam, G., Penumarti, A., Vora, M., Padrao, P., Fuentes, J., Bobadilla, L., Shin, J., & Ornik, M., “Belief-Conditioned One-Step Diffusion: Real-Time Trajectory Planning with Just-Enough Sensing,” Accepted to 9th Conference on Robot Learning (CoRL) as Oral Presentation
 
Research Team
UF Team: Aditya Penumarti, Jane Shin External Collaborators: Gokul Puthumanaillam, Manav Vora, Paulo Padrao, Jose Fuentes, Leonardo Bobadilla, Melkior Ornik
Acknowledgement
This work is completed in collaboration with University of Illinois - Urbana Champaign, Florida International University and Providence University.
          
          
    

