In scenarios involving the grasping of multiple targets, the learning of
stacking relationships between objects is fundamental for robots to execute
safely and efficiently. However, current methods lack subdivision for the
hierarchy of stacking relationship types. In scenes where objects are mostly
stacked in an orderly manner, they are incapable of performing human-like and
high-efficient grasping decisions. This paper proposes a perception-planning
method to distinguish different stacking types between objects and generate
prioritized manipulation order decisions based on given target designations. We
utilize a Hierarchical Stacking Relationship Network (HSRN) to discriminate the
hierarchy of stacking and generate a refined Stacking Relationship Tree (SRT)
for relationship description. Considering that objects with high stacking
stability can be grasped together if necessary, we introduce an elaborate
decision-making planner based on the Partially Observable Markov Decision
Process (POMDP), which leverages observations and generates the least
grasp-consuming decision chain with robustness and is suitable for
simultaneously specifying multiple targets. To verify our work, we set the
scene to the dining table and augment the REGRAD dataset with a set of common
tableware models for network training. Experiments show that our method
effectively generates grasping decisions that conform to human requirements,
and improves the implementation efficiency compared with existing methods on
the basis of guaranteeing the success rate.Comment: 8 pages, 8 figure