JOURNAL ARTICLE

Towards Dexterous In-Hand Manipulation through Tactile Sensing

Fernandes Veiga, Filipe

Year: 2018 Journal:   Technischen Universität Darmstadt   Publisher: Technical University of Darmstadt

Abstract

Currently, robots display manipulation capabilities that translate into actions such as picking and placing objects or poring liquid from containers. For actions that require finer in-hand manipulation to reposition objects or to use them as tools, robots are still not proficient enough. These shortcomings become even more apparent when considering the ease with which humans perform such manipulations on a daily basis, and while these limitations are not addressed, robots can not truly aid humans with their daily activities. The scope of possible interactions and the high dimensionality intrinsic to more dexterous robotic hands makes the manipulation problem hard to approach. Traditional control approaches to dexterous in-hand manipulation often work under the assumption that physical interactions, object properties, kinematics and dynamics of the robot can all be accurately modeled. Unfortunately, these modeling assumptions do not hold in most real environments, as uncertainty accumulates along the individual models. On the other hand, developing learning approaches that would generalize for all necessary manipulations proves difficult, as state spaces composed of the robots degrees of freedom and the necessary feedback channels often becomes too high dimensional and hence hard to explore. For dexterous in-hand manipulation, one of the most notable differences when considering human and robotic systems is the role of tactile information. The human system is composed of thousands of tactile afferents that provide detailed information of what is occurring at each interaction point during the manipulation action. On the other hand, traditional robotic manipulation approaches often rely on vision or on force feedback, either lacking information collected directly at the contact interaction or the various forms of information provided provided by tactile feedback. In this thesis, we explore tactile sensing as a means to bridge the gap in manipulation skills between humans and robots. We do so by assessing how to extract relevant feedback signals from the high dimensional tactile spaces, by exploring how to distribute the manipulation problem complexity onto modular components and to use these components to enable the use of powerful machine learning approaches, without loss of generalization capabilities. Chapter 2 will cover the recovery of relevant feedback signals from the tactile sensory information. Here, the desired signal is the state of the interaction between the robot and the object, particularly the knowledge of events such as slippage between the object and the finger surface. Through the use of machine learning, we predict such slip events in a manner that is generalizable to unknown objects. The ability to predict tactile slip allows analytically designed control solutions to stabilize objects when using a single finger. This is showed for cases where the opposing contact on the object is provided either by a static plane or by a human finger. In Chapter 3, we explore and extend a neurophysiological research hypothesis to the robotics domain. This hypothesis states that for stabilizing objects in-hand, digits can be controlled independently from each other, with no form of explicit coordination. Taking full advantage of the predictive slip feedback signals and ensuring smooth control responses by each of the finger controllers, we show that a stable grips on unknown objects emerge while controlling each digit independently. We show that coordination is achieved through the perturbations observed via the tactile feedback of each individual finger. Finally, in Chapter 4, use the modular nature of the grip stabilization control to enable the learning of manipulation policies in a hierarchical control setting. Reinforcement learning is used to learn a high-level control layer that exploits a lower-level composed of modular controllers that ensure the objects remains within the grip while being manipulated. In addition, we show that such a hierarchy facilitates the transfer of high-level policies learned in simulation onto real systems by using the low-level as an abstraction of the tactile information.

Keywords:
Robot Object (grammar) Tactile sensor Point (geometry) Robotics Degrees of freedom (physics and chemistry) Kinematics Scope (computer science)

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
0.52
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Geochemistry and Geologic Mapping
Physical Sciences →  Computer Science →  Artificial Intelligence
Geological Modeling and Analysis
Physical Sciences →  Earth and Planetary Sciences →  Geochemistry and Petrology
Electrical and Electromagnetic Research
Physical Sciences →  Physics and Astronomy →  Atomic and Molecular Physics, and Optics
© 2026 ScienceGate Book Chapters — All rights reserved.