Initial Analysis of Data-Driven Haptic Search for the Smart Suction Cup. Lee, J., Lee, S. D., Huh, T. M., & Stuart, H. S. October, 2023. arXiv:2401.06354 [cs]
Paper doi abstract bibtex Suction cups offer a useful gripping solution, particularly in industrial robotics and warehouse applications. Vision-based grasp algorithms, like Dex-Net, show promise but struggle to accurately perceive dark or reflective objects, sub-resolution features, and occlusions, resulting in suction cup grip failures. In our prior work, we designed the Smart Suction Cup, which estimates the flow state within the cup and provides a mechanically resilient end-effector that can inform arm feedback control through a sense of touch. We then demonstrated how this cup's signals enable haptically-driven search behaviors for better grasping points on adversarial objects. This prior work uses a model-based approach to predict the desired motion direction, which opens up the question: does a data-driven approach perform better? This technical report provides an initial analysis harnessing the data previously collected. Specifically, we compare the model-based method with a preliminary data-driven approach to accurately estimate lateral pose adjustment direction for improved grasp success.
@misc{lee_initial_2023,
title = {Initial {Analysis} of {Data}-{Driven} {Haptic} {Search} for the {Smart} {Suction} {Cup}},
url = {http://arxiv.org/abs/2401.06354},
doi = {10.48550/arXiv.2401.06354},
abstract = {Suction cups offer a useful gripping solution, particularly in industrial robotics and warehouse applications. Vision-based grasp algorithms, like Dex-Net, show promise but struggle to accurately perceive dark or reflective objects, sub-resolution features, and occlusions, resulting in suction cup grip failures. In our prior work, we designed the Smart Suction Cup, which estimates the flow state within the cup and provides a mechanically resilient end-effector that can inform arm feedback control through a sense of touch. We then demonstrated how this cup's signals enable haptically-driven search behaviors for better grasping points on adversarial objects. This prior work uses a model-based approach to predict the desired motion direction, which opens up the question: does a data-driven approach perform better? This technical report provides an initial analysis harnessing the data previously collected. Specifically, we compare the model-based method with a preliminary data-driven approach to accurately estimate lateral pose adjustment direction for improved grasp success.},
urldate = {2025-12-29},
publisher = {arXiv},
author = {Lee, Jungpyo and Lee, Sebastian D. and Huh, Tae Myung and Stuart, Hannah S.},
month = oct,
year = {2023},
note = {arXiv:2401.06354 [cs]},
keywords = {Computer Science - Robotics, ⭕/unread},
}
Downloads: 0
{"_id":"Gb9FFWhZd5nC8hGbD","bibbaseid":"lee-lee-huh-stuart-initialanalysisofdatadrivenhapticsearchforthesmartsuctioncup-2023","author_short":["Lee, J.","Lee, S. D.","Huh, T. M.","Stuart, H. S."],"bibdata":{"bibtype":"misc","type":"misc","title":"Initial Analysis of Data-Driven Haptic Search for the Smart Suction Cup","url":"http://arxiv.org/abs/2401.06354","doi":"10.48550/arXiv.2401.06354","abstract":"Suction cups offer a useful gripping solution, particularly in industrial robotics and warehouse applications. Vision-based grasp algorithms, like Dex-Net, show promise but struggle to accurately perceive dark or reflective objects, sub-resolution features, and occlusions, resulting in suction cup grip failures. In our prior work, we designed the Smart Suction Cup, which estimates the flow state within the cup and provides a mechanically resilient end-effector that can inform arm feedback control through a sense of touch. We then demonstrated how this cup's signals enable haptically-driven search behaviors for better grasping points on adversarial objects. This prior work uses a model-based approach to predict the desired motion direction, which opens up the question: does a data-driven approach perform better? This technical report provides an initial analysis harnessing the data previously collected. Specifically, we compare the model-based method with a preliminary data-driven approach to accurately estimate lateral pose adjustment direction for improved grasp success.","urldate":"2025-12-29","publisher":"arXiv","author":[{"propositions":[],"lastnames":["Lee"],"firstnames":["Jungpyo"],"suffixes":[]},{"propositions":[],"lastnames":["Lee"],"firstnames":["Sebastian","D."],"suffixes":[]},{"propositions":[],"lastnames":["Huh"],"firstnames":["Tae","Myung"],"suffixes":[]},{"propositions":[],"lastnames":["Stuart"],"firstnames":["Hannah","S."],"suffixes":[]}],"month":"October","year":"2023","note":"arXiv:2401.06354 [cs]","keywords":"Computer Science - Robotics, ⭕/unread","bibtex":"@misc{lee_initial_2023,\n\ttitle = {Initial {Analysis} of {Data}-{Driven} {Haptic} {Search} for the {Smart} {Suction} {Cup}},\n\turl = {http://arxiv.org/abs/2401.06354},\n\tdoi = {10.48550/arXiv.2401.06354},\n\tabstract = {Suction cups offer a useful gripping solution, particularly in industrial robotics and warehouse applications. Vision-based grasp algorithms, like Dex-Net, show promise but struggle to accurately perceive dark or reflective objects, sub-resolution features, and occlusions, resulting in suction cup grip failures. In our prior work, we designed the Smart Suction Cup, which estimates the flow state within the cup and provides a mechanically resilient end-effector that can inform arm feedback control through a sense of touch. We then demonstrated how this cup's signals enable haptically-driven search behaviors for better grasping points on adversarial objects. This prior work uses a model-based approach to predict the desired motion direction, which opens up the question: does a data-driven approach perform better? This technical report provides an initial analysis harnessing the data previously collected. Specifically, we compare the model-based method with a preliminary data-driven approach to accurately estimate lateral pose adjustment direction for improved grasp success.},\n\turldate = {2025-12-29},\n\tpublisher = {arXiv},\n\tauthor = {Lee, Jungpyo and Lee, Sebastian D. and Huh, Tae Myung and Stuart, Hannah S.},\n\tmonth = oct,\n\tyear = {2023},\n\tnote = {arXiv:2401.06354 [cs]},\n\tkeywords = {Computer Science - Robotics, ⭕/unread},\n}\n\n\n\n","author_short":["Lee, J.","Lee, S. D.","Huh, T. M.","Stuart, H. S."],"key":"lee_initial_2023","id":"lee_initial_2023","bibbaseid":"lee-lee-huh-stuart-initialanalysisofdatadrivenhapticsearchforthesmartsuctioncup-2023","role":"author","urls":{"Paper":"http://arxiv.org/abs/2401.06354"},"keyword":["Computer Science - Robotics","⭕/unread"],"metadata":{"authorlinks":{}},"downloads":0,"html":""},"bibtype":"misc","biburl":"https://bibbase.org/zotero/yliangbk","dataSources":["6LWsyvHEbAxHmChNb"],"keywords":["computer science - robotics","⭕/unread"],"search_terms":["initial","analysis","data","driven","haptic","search","smart","suction","cup","lee","lee","huh","stuart"],"title":"Initial Analysis of Data-Driven Haptic Search for the Smart Suction Cup","year":2023}