Bimanual manipulation tasks, such as unfolding a box or opening scissors, are prevalent in daily life and demand approaches that account for collaboration between both hands. These tasks, however, present substantial difficulties due to the need for large amounts of training data. Additionally, when manipulating novel object categories, the variation in geometric characteristics and physical properties across categories complicates generalization. To address these challenges, we present Bi-Adapt, a novel framework designed for efficient learning of bimanual manipulation for novel categories. Bi-Adapt leverages semantic correspondence from diffusion models, known for their strong generalization abilities, to transfer affordance to new categories. Furthermore, it introduces an efficient few-shot adaptation strategy that fine-tunes actions with minimal data, improving performance on novel object categories. Our experimental results demonstrate the high efficiency of Bi-Adapt, achieving high success rates in complex bimanual manipulation tasks with restricted data.
BibTex Code Here