|
|
|
|
|
|
|
|
|
|
At each step, the agent will receive a 360 panorama or 90 degree egocentric observation. Here in the video, we show a running example of our model exploring a novel scene using a frontier-based approach. A frontier is a boundary between free and unknown space, as denoted by the green pixels in the video. Yellow pixels are the selected frontier. |
|
|
|
|
For the map based model, We use a U-net architecture where the input to the U-Net model is the currently observed occupancy and semantic map. For the view based model, the input to the ResNet-18 model is the egocentric depth and semantic observation. |
|
|
@inproceedings{li2023learning, title={Learning-augmented model-based planning for visual exploration}, author={Li, Yimeng and Debnath, Arnab and Stein, Gregory J and Ko{\v{s}}eck{\'a}, Jana}, booktitle={2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)}, pages={5165--5171}, year={2023}, organization={IEEE} } |
AcknowledgementsWe thank members of the GMU Vision and Robotics Lab and RAIL.This webpage template was borrowed from some colorful folks. |