Indoor robots are becoming increasingly prevalent across a range of sectors, but the challenge of navigating multi-level structures through elevators remains largely uncharted. For a robot to operate successfully, it's pivotal to have an accurate perception of elevator states. This paper presents a robust robotic system, tailored to interact adeptly with elevators by discerning their status, actuating buttons, and boarding seamlessly. Given the inherent issues of class imbalance and limited data, we utilize the YOLOv7 model and adopt specific strategies to counteract the potential decline in object detection performance. Our method effectively confronts the class imbalance and label dependency observed in real-world datasets, Our method effectively confronts the class imbalance and label dependency observed in real-world datasets, offering a promising approach to improve indoor robotic navigation systems.
GAEMI is a sophisticated mobile manipulator equipped with a 5DoF robotic arm and a ZED camera. Its non-holonomic base features a 2D LiDAR sensor for obstacle detection and localization within a mapped environment. Additionally, GAEMI has a forward-facing RGB camera that serves as the primary sensor for our elevator perception system. This setup enables GAEMI to navigate complex indoor environments, detect elevator states, and interact with control panels effectively.
Category | Parameters |
---|---|
Elevator Door | Opened, Moving, Closed |
Current Robot Floor | B6, B5, ..., B1, 1, ..., 63 |
Current Elevator Floor (Outside/Inside) | B6, B5, ..., B1, 1, ..., 63 |
Current Elevator Direction (Outside/Inside) | Up, Down, None |
Elevator Button (Outside/Inside) | Up, Down, B6, B5, ..., B1, 1, ..., 63 |
The primary objective of our perception system is to ascertain the elevator's status, including door state, current floor, and location. We defined a comprehensive label superset covering all possible scenarios across diverse sites, enabling the robot to make decisions and navigate intricate environments effectively. This is essential for seamless navigation and interaction with elevator systems.
We developed two complementary datasets for our elevator interaction system. The (a) Indicator dataset is tailored to capture the basic status of an elevator, including door state, current floor indicators, and directional signals. This object detection dataset provides the fundamental situational awareness needed for decision-making.
The (b) Button dataset is an instance segmentation dataset designed to identify precise points of interaction between the robot and the elevator, facilitating successful task execution. This dataset includes pixel-level masks for elevator buttons, enabling the robot to accurately locate and interact with control panels. Together, these datasets provide the comprehensive perception capabilities required for autonomous elevator navigation.
Dataset | mAP@0.5 | mAP@0.95 |
---|---|---|
COCO-mini (base) | 0.014 | 0.007 |
COCO-blur | 0.018 | 0.009 |
COCO-cutout | 0.012 | 0.006 |
Method | mAP@0.5 | Status Accuracy |
---|---|---|
YOLOv7 | 0.730 | 0.813 |
YOLOv7 + patch | 0.784 | 0.878 |
YOLOv7 + patch + blur | 0.779 | 0.879 |
Task | Success Rate |
---|---|
GOTO BUTTON POSE | 10/10 |
BUTTON PUSHING | 9/10 |
ELEVATOR BOARDING | 3/10 |
Our experimental results validate the effectiveness of our approach in addressing label dependency and small object detection challenges. The COCO-mini evaluation demonstrates that our blur augmentation technique outperforms both the baseline and cutout methods. On the Indicator dataset, our patch augmentation significantly improved accuracy (mAP@0.5: 0.784, Status Accuracy: 0.878) compared to the baseline, while adding blur augmentation maintained high accuracy while addressing class imbalance.
In real-world robot operations conducted at Korea University, we evaluated three critical tasks with varying success rates: 100% for navigation, 90% for button pushing, and 30% for elevator boarding over ten trials each. The lower success rate in elevator boarding indicates areas for future improvement, particularly in handling potential obstructions at the elevator door.
Our final demonstration showcases the GAEMI robot navigating and interacting with elevators in the Woojung Hall of Informatics at Korea University. The system integrates all components - perception, planning, and control - to enable autonomous multi-floor navigation. The robot successfully detects elevator status, navigates to the appropriate position, interacts with buttons, and boards/exits elevators with high reliability. This real-world demonstration validates the practical applicability of our approach for indoor service robots operating in multi-level environments.
@inproceedings{shin2023robust,
title={Robust Detection for Autonomous Elevator Boarding using a Mobile Manipulator},
author={Shin, Seungyoun and Lee, Joonhyung and Noh, Junhyug and Choi, Sungjoon},
booktitle={Link Springer},
year={2023}
}