Early Estimation of User’s Intention of Tele-Operation Using Object Affordance and Hand Motion in a Dual First-Person Vision

Early Estimation of User’s Intention of Tele-Operation
Using Object Affordance and Hand Motion in a Dual First-Person Vision

Motoki Kojima and Jun Miura
Department of Computer Science and Engineering
Toyohashi University of Technology
Abstract

This paper describes a method of estimating the intention of a user’s motion in a robot tele-operation scenario. One of the issues in tele-operation is latency, which occurs due to various reasons such as a slow robot motion and a narrow communication channel. An effective way of reducing the latency is to estimate human intention of motions and to move the robot proactively. To enable a reliable early intention estimation, we use both hand motion and object affordances in a dual first-person vision (robot and user) with an HMD. Experimental results in an object pickup scenario show the effectiveness of the method.

1 Introduction

Lifestyle support is one of the promising application domains of robotic technologies. A mobile manipulator is a useful tool for the user with a disability to perform various home tasks such as bringing a user-specified object from a remote room. Tele-operation through a robot’s view is a common approach to easily controlling such a robot. Human body motion-based control (e.g., [5]) is promising because it can provide an intuitive way of tele-operation. There are, however, two issues in body motion-based control. One is the body motion observation, and the other is the latency due to motion estimation or robot control.

Concerning body motion estimation, existing devices such as MoCap, RGB-D cameras, and wearable sensors are costly and/or not very reliable. Concerning the latency, even if the communication channel is fast enough, the latency is sometimes unavoidable due to, for example, the difficulty in controlling a robot precisely. A hybrid approach will then be a solution, which combines the user’s command and robot’s autonomous movement based on object recognition and motion planning capabilities, but keeping the intuitiveness of operation remains as a problem.

To address these issues, we propose an approach based on early intention estimation of human actions with a dual first-person view of the robot and the user. We suppose the situation where a user tries to pick up one of the objects on a remote table. Fig. 1 shows an intended scenario and an overview of the proposed method. The user obtains the robot’s current view through a head-mounted display (HMD). The user reaches out a hand to touch a target object in the view. Of course, the hand never touches the object at a remote site. The system uses a stream of hand images for intention estimation. For an early but reliable estimation of intention, we use a strong relationship between human hand shape (i.e., preshaping [2]) and object affordances. Unlike the previous work on action recognition in a first-person vision, we combine data from different views.

Figure 1: An intended scenario and the overview of the proposed action intention estimation method. The user with an HMD sees the image from the robot and trying to pick up one of the objects by moving his/her hand. The method analyzes both the first-person view of the hand and the robot view to determine which object he/she wants to pick up.

2 Related Work

2.1 Action recognition by first-person vision

Many methods have been proposed for End-to-End action recognition by first-person vision [11, 7]. These methods analyzes the relationships of objects and hand motions which both appear in first-person image sequences. In our tele-operation scenario, these methods cannot be directly adopted because hands and objects are at different locations and therefore observed in separate images.

2.2 Hand motion intention estimation using preshaping and affordances

Characteristics of an object which naturally induces a human action is called affordance [4]. Arbib et al. [1] show the relationship between the shape and the motion of a hand and the visible characteristics of the target object. This implies that preshaping provides useful information for identifying the target object [6]. In our problem setting, the user cannot get a visual feedback of her/his own hand position and motion. Some work show that even in such a case, as long as the location of the target of interaction is known, the reaching motion can be executed similarly [12, 10]; This supports the use of separately observed object locations/types and a hand motion stream.

3 Object Affordance Extraction

Do et al. [3] developed AffordanceNet. Affordance here is a part of an object which is closely related to some human action. AffordanceNet outputs object names and bounding boxes, and affordance-wise segmentation.

We use cups as target objects. We modified the original AffordanceNet such that the class output is only “cup” and the affordance output has the following four classes: contain, wrap-grasp, handle-grasp, and background. We also constructed a new dataset accordingly by ourselves. Fig. 2 shows examples in the new dataset and an affordance extraction result.

Figure 2: Example data with a simulated (left) and a real (center) background and an affordance extraction exmaple (right).

4 Intention Estimation Network

4.1 Network structure

Early intention estimation or action prediction usually necessitates integration and analysis of time series data. We design a network, called Intention Estimation Network (IEN), which is to calculate probable target location(s) of the current hand motion based on an integration of affordance information and hand information (see Fig. 3). For the former, we use the output of AffordanceNet trained with our dataset, which is composed of a set of four affordance masks and a bounding box mask. For the latter, we use the output of the ConvLSTM [9] with a sequence of hand images as an input. We concatenate both outputs and then supply it to a variant of U-Net [8], one of the popular networks for semantic segmentation. The final heatmap is normalized so that the values of all pixels sum to one.

Figure 3: Intention estimation network.

4.2 Dataset construction for intention estimation

The input to IEN includes the extracted object and affordance infomration as well as a hand image sequence. The latter is the heatmap of the target location. Since the output of AffordanceNet for a real setting is not free from errors, we first generated a variety of simulated scenes, which are with ground truth information of affordances and object bounding boxes, and then provided them to the user for taking a hand-reaching motion.

Our system uses a stereoscopic head-mounted display (HMD) (Occulus Rift CV1) for presenting images of the remote site to the user. We use an RGB-D camera (Intel RealSense D435) for acquiring hand motion sequences, which are recorded both in RGB and depth images. Fig. 1 shows a user wearing the system on the bottom left.

The steps for collecting a hand motion sequence for an affordance/stereo pair are as follows (see Fig. 4):

  1. Put two or three cups on the table. Textures, orientations, and locations of cups are set randomly.

  2. A stereo image pair is shown on the HMD.

  3. The user moves the right hand to try to pick up the designated object in the scene, and the hand image sequence is recorded with the RGB-D camera.

  4. The recorded sequence is saved with the output of AffordanceNet.

Figure 4: An example pair of stereo image and hand motion.

5 Performance Evaluation

5.1 Training and testing the network

We collected 156 pairs of affordance and hand motion information for training. One hand motion sequence consists of 60 frames (30fps, 2sec) and a set of 10-frame sequences are extracted with a sliding-window fashion. As a result, we have 50 sequences for each original sequence and 7,800 sequences in total. The ground truth of the output heatmap is generated by putting a Gaussian distribution at the location of the designated object. We used the KL-divergence as the loss function to evaluate the distance between the true and the predicted heatmap.

We generated an additional test dataset for evaluation by 50 trials. The input to the intention estimation network (IEN) in testing is not limited to a fixed-length (10 frames) sequence but a set of sequences with variable lengths from 1 to 15 to see how early the intention can be estimated.

We examined the following three points: (1) effectiveness of affordance information, (2) comparison of RGB and depth images, and (3) effectiveness of hand region extraction in RGB images.

5.2 Estimation examples

Fig. 5 shows the inputs and the corresponding results for depth images. There are two cups, and the user is asked to pick up the right one. With affordance information, the probability of the right cup gradually increased, while without affordance information, the probabilities for both objects remain almost the same.

Figure 5: Intention estimation results for depth images.

Fig. 6 examines the usefulness of hand extraction for RGB images. There are two cups, and the user is asked to pick up the right one. With hand region information, the probability of the right cup increased at the 6th frame, while both cups are equally probable without that information. Since we learned the use of hand extraction is always useful, we include hand extraction for RGB images in the subsequent quantitative comparison.

Figure 6: Intention estimation results for RGB images.

5.3 Quantitative evaluation

The purpose of the intention estimation is to determine the target object as early as possible. Since a heatmap can be interpreted as a probability distribution of the location of the target object, we use the sum of probabilities with the bounding box of an object as the confidence value. Such confidence values for all objects in the scene are normalized to be used as probabilities of being the target. The object whose probability exceeds a threhold is judged as the intended target object. We use the averaged f-value as a measure of estimation quality.

Table 1 compares actual f-values for four cases in order to summarize the above comparisons. Using depth images is more effective in the intention estimation because the hand region and shape information is more clearly obtained compared to RGB images. Using affordance information is useful for both depth and RGB images, and using RGB images with hand extraction and affordances is comparable to using depth images without affordances.

case/frame 2 4 6 8 10 12
Depth-AO 0.405 0.487 0.695 0.745 0.761 0.725
Depth-O 0.038 0.466 0.600 0.610 0.625 0.658
RGB-AO 0.242 0.286 0.350 0.432 0.568 0.667
RGB-O 0.039 0.145 0.317 0.286 0.556 0.691

case/frame 2 4 6 8 10 12
Depth-AO 0.175 0.323 0.507 0.615 0.675 0.707
Depth-O 0.039 0.113 0.323 0.355 0.500 0.556
RGB-AO 0.038 0.172 0.197 0.338 0.493 0.640
RGB-O 0.000 0.039 0.113 0.182 0.276 0.381

*-AO: with affordance, *-O: without affordance

Table 1: Summary of comparison in terms of f-values.

6 Summary

This paper proposes a method for early estimation of the user’s intention of tele-operation using a dual first-person images and object affordances. We deal with a tele-operated object pickup scenario where the user sees a view of a remote site through an HMD and moves his/her hand to try to pick up a target object. AffordanceNet extracts object and affordance information and ConvLSTM extracts hand motion and shape information. The U-Net combines these two kinds of information to generate a heatmap of the target location. We generated two new datasets, one is for training AffordanceNet and the other is for training the whole network. We compared several cases and show that affordance information is useful in early intention estimation both for depth and RGB images. Testing the developed method on a real robot and applying it to other objects are future work.

Acknowledgment

This work is in part supported by JSPS KAKENHI Grant Numbers 17H01799. The authors would like to thank Mr. Hiroaki Masuzawa for fruitful discussion.

References

  • [1] M. Arbib, T. Iberall, and D. Lyons. Coordinated control programs for movements of the hand. In Hand Function and the Neocortex, pages 135–170. Springer Verlag, 1985.
  • [2] C. Bard, J. Troccaz, and G. Vercelli. Shape analysis and hand preshaping for grasping. In Proc. IROS-91, pages 64–69, 1991.
  • [3] T.-T. Do, A. Nguyen, and I. Reid. AffordanceNet: An end-to-end deep learning approach for object affordance detection. In Proc. ICRA-2018, 2018.
  • [4] J. Gibson. The Ecological Approach to Visual Perception. Houghton Mifflin, 1979.
  • [5] C. Ha, S. Park, J. Her, I. Jang, Y. Lee, G. Cho, H. Son, and D. Lee. Whole-body multi-modal semi-autonomous teleoperation of mobile manipulator systems. In Proc. ICRA-2015, pages 164–170, 2015.
  • [6] N. Nakamaura, T. Miyasato, and F. Kishino. Preshaping analysis and target prediction in grasping hand motion. IEICE Trans. on Fundamentals, J80-A(6):1003–1013, 1997. (in Japanese).
  • [7] N. Rhinehart and K. Kitani. First-person activity forecasting from video with online inverse reinforcement learning. IEEE Trans. on Pattern Analysis and Machine Intelligence, 2018.
  • [8] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. arXiv preprint arXiv:1505.04597, 2015.
  • [9] X. Shi, Z. Chen, H. Wang, D.-Y. Yeung, W.-K. Wong, and W.-C. Woo. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In Advances in Neural Information Processing Systems, volume 25, pages 802–810, 2015.
  • [10] S. Shimawaki and N. Sakai. Effect of viewing conditions on preshaping movement in hand grasp motion. The Japanese J. of Ergonomics, 47(1):31–35, 2011.
  • [11] S. Singh, C. Arora, and C. Jawahar. First person action recognition using deep learned descriptors. In Proc. CVPR-2016, 2016.
  • [12] S. Winges, D. Weber, and M. Santello. The role of vision on hand preshaping during reach to grasp. Experimental Brain Research, 152:489–498, 2003.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
""
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
   
Add comment
Cancel
Loading ...
393296
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test
Test description