{"id":105519,"date":"2025-01-27T08:31:19","date_gmt":"2025-01-27T14:31:19","guid":{"rendered":"https:\/\/engineering.wisc.edu\/?post_type=tribe_events&p=105519"},"modified":"2025-01-30T08:36:57","modified_gmt":"2025-01-30T14:36:57","slug":"high-resolution-tactile-sensing-for-robotic-manipulation","status":"publish","type":"tribe_events","link":"https:\/\/engineering.wisc.edu\/event\/high-resolution-tactile-sensing-for-robotic-manipulation\/","title":{"rendered":"High-resolution Tactile Sensing for Robotic Manipulation\u00a0"},"content":{"rendered":"
<\/p>\n\n\n\n In the quest to enable robots to perform everyday tasks in unstructured environments, tactile sensing plays a crucial role. Tasks that benefit from refined tactile perception include industrial manufacturing, domestic chores, and the care of elderly individuals. However, high fidelity tactile sensing remains a challenge within the community. Our lab has made strides in the high-resolution tactile sensors and perception algorithms respond to this challenge. This seminar will showcase several prototypes designed to improve robotic manipulation through advanced tactile feedback. We will introduce the GelSight Wedge, a compact tactile sensor optimized for grasping in cluttered spaces, and VisTac, a sensor that integrates visual and tactile inputs for precise object localization, grasping, and insertion. Additionally, we will present VibTac, a sensor combining acoustic and vision-based tactile sensing for processing audio and vibration signals. Build upon the rich feedback, we explore diverse robotic manipulation tasks such as robot cable manipulation and dexterous in-hand manipulation. Our work aims to equip robots with the autonomy to intelligently perform a wide range of tasks, moving us closer to a future where robots seamlessly integrate into daily life. <\/p>\n\n\n<\/figure>\n\n\n\n