+ All Categories
Home > Documents > OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D...

OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D...

Date post: 31-May-2020
Category:
Upload: others
View: 6 times
Download: 0 times
Share this document with a friend
35
Yu Xiang, 7/12/2019 OBJECT PERCEPTION FOR ROBOT MANIPULATION
Transcript
Page 1: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

Yu Xiang, 7/12/2019

OBJECT PERCEPTION FOR ROBOT MANIPULATION

Page 2: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

2

MANIPULATION• The way of making physical changes to the world around us

Vs. question answering or autonomous driving

Page 3: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

3

MANIPULATION REQUIRES INTELLIGENCE• Understanding the 3D environment from sensing

• E.g., Vision, Tactile

• Grasp and motion planning / decision making

• E.g., Obstacle avoidance

• Dynamics / Control

• Learning from experience

Page 4: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

4

ROBOT MANIPULATION

IntelligentSystem

3D EnvironmentLearning

Sensing PlanningPerception Control

Page 5: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

5

Page 6: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

6

6D OBJECT POSE ESTIMATION

Page 7: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

7

USING 3D MODELS OF OBJECTS• The YCB Object and Model Set

B. Calli, A. Singh, A. Walsman, S. Srinivasa, P. Abbeel and A. M. Dollar, "The YCB object and Model set: Towards common benchmarks for manipulation research," International Conference on Advanced Robotics (ICAR), 2015.

Page 8: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

8

6D POSE ESTIMATION FOR GRASP PLANNING

From Graspit!

Page 9: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

9

POSECNN

Yu Xiang, Tanner Schmidt, Venkatraman Narayanan and Dieter Fox. PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes. In RSS, 2018.

PoseCNN Texture-less objects Symmetric objects Occlusions

Page 10: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

10

camera coordinate

POSECNN: DECOUPLE 3D TRANSLATION AND 3D ROTATION

object coordinate2D center

Distance

• 3D Translation• 3D Rotation

2D Center Localization 3D Rotation Regression

Page 11: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

11

POSECNN: SEMANTIC LABELING

LabelsSkip link

Input image

Fully convolutional networkEncoder Decoder

Page 12: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

12

POSECNN: 2D CENTER VOTING FOR HANDLING OCCLUSIONS

center

Page 13: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

13

POSECNN: 3D TRANSLATION ESTIMATION

Labels#classes

Center direction X

Center direction Y

Center distance

Hough voting layer

3 × #classes

Page 14: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

14

POSECNN: 3D ROTATION REGRESSION

Labels#classes

3 × #classes

Center direction X

Center direction Y

Center distance

Hough voting layer

RoIs

RoI pooling layers 6D Poses4 × #class

For each RoI

Page 15: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

15

POSECNN: HANDLE SYMMETRIC OBJECTS

Page 16: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

16

POSECNN: 3D ROTATION REGRESSION LOSS FUNCTIONS

Pose Loss (non-symmetric)

Shape-Match Loss for symmetric objects (symmetric)

3D model points

Ground truth rotation

Predicted rotation

Page 17: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

17

IMPLICIT ROTATION LEARNING

Encoder Decoder

Embedding

Input ReconstructionSundermeyer et al. Implicit 3D orientation learning for 6D object detection from RGB images. In ECCV, 2018.

Page 18: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

18

ROTATION ESTIMATION WITH CODEBOOK MATCHING

Encoder

Codebook

191,808 discrete rotations

Similarity scores

Input

Detection

Page 19: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

19

TRAINING DATA: DOMAIN RANDOMIZATION

Page 20: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

20

TRAINING DATA: DOMAIN RANDOMIZATION

Page 21: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

21

Page 22: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

22

DEEP ITERATIVE MATCHING FOR 6D OBJECT POSE ESTIMATION

Yi Li*, Gu Wang, Xiangyang Ji, Yu Xiang and Dieter Fox. DeepIM: Deep Iterative Matching for 6D Pose Estimation. In ECCV, 2018 (Oral) (*PhD student at UW).

Initial pose estimation

pose changeDeep Neural Network

Page 23: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

23

pose(0)

Δpose(0)

Network

Observed image

3D model

Renderer

Rendered image

pose(1)

Network

3D model

Renderer

×

Δpose(1)

×

DEEPIM PIPELINE

Rendered image

Page 24: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

24

FlowNetConvs [1]

Rotation

Translation

FC4

FC3

640x480

Observedimage 

Renderedimage

Zoomed input

FC256FC256

Feature map

NETWORK STRUCTURE

[1] Dosovitskiy, Alexey and Fischer, Philipp and Ilg, Eddy and Hausser, Philip and Hazirbas, Caner and Golkov, Vladimir and Van Der Smagt, Patrick and Cremers, Daniel and Brox, Thomas. Flownet: Learning optical flow with convolutional networks. In ICCV, 2015.

Page 25: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

25

TRAINING DATA: YCB OBJECTS

Page 26: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

26

Page 27: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

27

6D OBJECT POSE TRACKING

PoseRBPF

Input images Translation Orientationdistribution

Xinke Deng*, Arsalan Mousavian, Yu Xiang, Fei Xia*, Timothy Bretl and Dieter Fox. PoseRBPF: A Rao-Blackwellized Particle Filter for 6D Object Pose Tracking. In RSS, 2019 (*intern at NVIDIA).

Uncertainty in Pose Estimation

Page 28: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

28

PoseRBPF: Particle Representation

3D Translation𝑇

Orientation Distribution𝑷 𝑹𝒊 𝑻𝒊, 𝒁𝟏:𝒌

RoI

Encoder

Discretized Rotations

Codebook

ParticleCode

Rotation Likelihood

191,808 bins

Page 29: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

29

Results: YCB ObjectsExample: YCB mug (50 particles, ~20fps)

YCB-Video RGB PoseRBPF:

ADD: 62.1, ADD-S: 78.4 PoseCNN:

ADD: 53.7, ADD-S: 75.9

Page 30: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

30

Results: TLess ObjectsExample: TLess 01 (100 particles, ~11fps)

TLess RGBObject recall for Err_vsd < 0.3: PoseRBPF: 41.47% Sundermeyer et al: 18.35%

Page 31: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

3131

Yu Xiang and Dieter Fox. DA-RNN Semantic Mapping with Data Associated Recurrent Neural Network, RSS, 2017.

SEMANTIC MAPPING

Page 32: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

32

UNSEEN OBJECT INSTANCE SEGMENTATION

Christopher Xie*, Yu Xiang, Arsalan Mousavian and Dieter Fox. The Best of Both Modes: Separately Leveraging RGB and Depth for Unseen Object Instance Segmentation. Under Review, 2019 (*PhD student at UW).

Failu

res

Page 33: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

33

POSECNN FOR 20 YCB OBJECTS

Page 34: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

34

FUTURE WORK: SELF-SUPERVISED LEARNING

Simulation

Synthetic data model

Leaning

Real environment

Interacting

Apply

Updating

Training data in the real world• Reason about uncertainty/failure in

the real world to obtain annotations• Interact with the real world to

collect more data

New environment? Learning in that environment to

adapt the model!

Page 35: OBJECT PERCEPTION FOR ROBOT MANIPULATIONyuxng.github.io/Xiang_TRI_07122019.pdf ·  · 2020-04-076D OBJECT POSE TRACKING PoseRBPF Input images Translation Orientation distribution

Questions?


Recommended