OakInk: A Large-Scale Knowledge Repository for Understanding Hand-Object Interaction

Abstract

Learning how humans manipulate objects requires machines to acquire knowledge from two perspectives: one for understanding object affordances and the other for learning human's interactions based on the affordances. Even though these two knowledge bases are crucial, we find that current databases lack a comprehensive awareness of them. In this work, we propose a multi-modal and rich-annotated knowledge repository, OakInk, for visual and cognitive understanding of hand-object interactions. We start to collect 1,800 common household objects and annotate their affordances to construct the first knowledge base: Oak. Given the affordance, we record rich human interactions with 100 selected objects in Oak. Finally, we transfer the interactions on the 100 recorded objects to their virtual counterparts through a novel method: Tink. The recorded and transferred hand-object interactions constitute the second knowledge base: Ink. As a result, OakInk contains 50,000 distinct affordance-aware and intent-oriented hand-object interactions. We benchmark OakInk on pose estimation and grasp generation tasks. Moreover, we propose two practical applications of OakInk: intent-based interaction generation and handover generation. Our dataset and source code are publicly available at www.oakink.net.

Cite

Text

Yang et al. "OakInk: A Large-Scale Knowledge Repository for Understanding Hand-Object Interaction." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.02028

Markdown

[Yang et al. "OakInk: A Large-Scale Knowledge Repository for Understanding Hand-Object Interaction." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/yang2022cvpr-oakink/) doi:10.1109/CVPR52688.2022.02028

BibTeX

@inproceedings{yang2022cvpr-oakink,
  title     = {{OakInk: A Large-Scale Knowledge Repository for Understanding Hand-Object Interaction}},
  author    = {Yang, Lixin and Li, Kailin and Zhan, Xinyu and Wu, Fei and Xu, Anran and Liu, Liu and Lu, Cewu},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {20953-20962},
  doi       = {10.1109/CVPR52688.2022.02028},
  url       = {https://mlanthology.org/cvpr/2022/yang2022cvpr-oakink/}
}