Abstract

This paper presents a novel approach to robotic grasping by integrating embodied visual navigation with reinforcement learning. The primary objective is to determine the optimal location for a robot to stand for successful object grasping. The motivation for this research is to address the existing gap in the literature where navigation and grasping are often treated as separate problems, leading to suboptimal performance. Our approach leverages multimodal sensory data, including RGB images, depth images, and semantic information, to guide the robot's navigation. It also utilizes deep reinforcement learning to enable the robot to learn optimal navigation strategies from visual input. The effectiveness of this approach is demonstrated through a series of experiments conducted in simple and complex scenes with varying numbers of obstacles. The results show that our method achieves a high success rate and a fast grasping speed in different scenarios, outperforming other methods. This work contributes significantly to the field of robotic grasping by integrating embodied visual navigation and deep reinforcement learning, and by demonstrating its effectiveness through rigorous experiments.

Keywords:
Computer science Embodied cognition Artificial intelligence Reinforcement learning Robot Computer vision Mobile robot navigation Field (mathematics) Visualization Object (grammar) Human–computer interaction Mobile robot Robot control

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
10
Refs
0.29
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Robot Manipulation and Learning
Physical Sciences →  Engineering →  Control and Systems Engineering
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Robotic Locomotion and Control
Physical Sciences →  Engineering →  Biomedical Engineering

Related Documents

JOURNAL ARTICLE

Echo-Enhanced Embodied Visual Navigation

Yinfeng YuLele CaoFuchun SunChao YangHuicheng LaiWenbing Huang

Journal:   Neural Computation Year: 2023 Vol: 35 (5)Pages: 958-976
JOURNAL ARTICLE

Visual Navigation Subject to Embodied Mismatch

Xinzhu LiuDi GuoHuaping LiuXinyu ZhangFuchun Sun

Journal:   IEEE Transactions on Cognitive and Developmental Systems Year: 2023 Vol: 15 (4)Pages: 1959-1970
JOURNAL ARTICLE

FloNa: Floor Plan Guided Embodied Visual Navigation

Jiaxin LiWen‐Chih HuangZan WangWei LiangHuijun DiFeng Liu

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2025 Vol: 39 (14)Pages: 14610-14618
JOURNAL ARTICLE

Symmetry-aware Neural Architecture for Embodied Visual Navigation

Shuang LiuMasanori SuganumaTakayuki Okatani

Journal:   International Journal of Computer Vision Year: 2023 Vol: 132 (4)Pages: 1091-1107
© 2026 ScienceGate Book Chapters — All rights reserved.