JOURNAL ARTICLE

GPT-COPE: A Graph-Guided Point Transformer for Category-Level Object Pose Estimation

Lu ZouZhangjin HuangNaijie GuGuoping Wang

Year: 2023 Journal:   IEEE Transactions on Circuits and Systems for Video Technology Vol: 34 (4)Pages: 2385-2398   Publisher: Institute of Electrical and Electronics Engineers

Abstract

Category-level object pose estimation aims to predict the 6D pose and 3D metric size of objects from given categories. Due to significant intra-class shape variations among different instances, existing methods have mainly focused on estimating dense correspondences between observed point clouds and their canonical representations, i.e., normalized object coordinate space (NOCS). Subsequently, a similarity transformation is applied to recover the object pose and size. Despite these efforts, current approaches still cannot fully exploit the intrinsic geometric features to individual instances, thus limiting their ability to handle objects with complex structures (i.e., cameras). To overcome this issue, this paper introduces GPT-COPE, which leverages a graph-guided point transformer to explore distinctive geometric features from the observed point cloud. Specifically, our GPT-COPE employs a Graph-Guided Attention Encoder to extract multiscale geometric features in a local-to-global manner and utilizes an Iterative Non-Parametric Decoder to aggregate the multiscale geometric features from finer scales to coarser scales without learnable parameters. After obtaining the aggregated geometric features, the object NOCS coordinates and shape are regressed through the shape prior adaptation mechanism, and the object pose and size are obtained using the Umeyama algorithm. The multiscale network design enables perceiving the overall shape and structural information of the object, which is beneficial to handle objects with complex structures. Experimental results on the NOCS-REAL and NOCS-CAMERA datasets demonstrate that our GPT-COPE achieves state-of-the-art performance and significantly outperforms existing methods. Furthermore, our GPT-COPE shows superior generalization ability compared to existing methods on the large-scale in-the-wild dataset Wild6D and achieves better performance on the REDWOOD75 dataset, which involves objects with unconstrained orientations.

Keywords:
Pose Computer science Point cloud Artificial intelligence Encoder Computer vision Parametric statistics Geometric shape Graph Rigid transformation Pattern recognition (psychology) Mathematics Theoretical computer science

Metrics

8
Cited By
1.99
FWCI (Field Weighted Citation Impact)
44
Refs
0.83
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Robot Manipulation and Learning
Physical Sciences →  Engineering →  Control and Systems Engineering
3D Shape Modeling and Analysis
Physical Sciences →  Engineering →  Computational Mechanics
Anatomy and Medical Technology
Physical Sciences →  Engineering →  Biomedical Engineering

Related Documents

JOURNAL ARTICLE

GPV-Pose: Category-level Object Pose Estimation via Geometry-guided Point-wise Voting

Yan DiRuida ZhangZhiqiang LouFabian ManhardtXiangyang JiNassir NavabFederico Tombari

Journal:   2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Year: 2022 Pages: 6771-6781
JOURNAL ARTICLE

Diff-COPE: Diffusion-Based Category-Level 6D Object Pose Estimation

Yingbo TangZhiqiang CaoPeiyu GuanXurong GongMengyao WangJunzhi Yu

Journal:   IEEE Transactions on Circuits and Systems for Video Technology Year: 2025 Pages: 1-1
JOURNAL ARTICLE

CatFormer: Category-Level 6D Object Pose Estimation with Transformer

Sheng YuDi‐Hua ZhaiYuanqing Xia

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2024 Vol: 38 (7)Pages: 6808-6816
JOURNAL ARTICLE

UDA-COPE: Unsupervised Domain Adaptation for Category-level Object Pose Estimation

Taeyeop LeeByeong-Uk LeeInkyu ShinJaesung ChoeUkcheol ShinIn So KweonKuk‐Jin Yoon

Journal:   2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Year: 2022 Pages: 14871-14880
© 2026 ScienceGate Book Chapters — All rights reserved.