1 Introduction

Detecting individual persons and objects in isolation often does not provide sufficient information for understanding complex human activities. Moving beyond detecting/recognizing individual objects, we aim to detect persons, objects, and recognize their interaction relationships (if any) in the scene. This task, known as human-object interaction (HOI) detection, can produce rich semantic information with visual grounding.

Fig. 1.
figure 1

Human-object interaction (HOI) detection using dual relation graph. Predicting each HOI in isolation is ambiguous due to the lack of context. In this work, we propose to leverage a dual relation graph. For each human node h, we obtain a human-centric subgraph where all object nodes are connected to h. Similarly, we can obtain an object-centric subgraph for each object node o. The human subgraph helps to adjust single HOI’s prediction based on the same person’s other HOIs. For example, knowing a is wearing a and hitting a suggests that the may be holding a . Similarly, the object subgraph helps to refine the HOI’s prediction based on other HOIs associated with the same object. For example, knowing a is held by a lowers the chance that it is held by another . Our method exploits such cues for improving HOI detection.

State-of-the-art HOI detection methods often use appearance features from the detected human/object instances as well as their relative spatial layout for predicting the interaction relationships  [1, 3, 9, 12,13,14, 21, 24, 25, 32, 40, 41, 51]. These methods, however, often predict the interaction relationship between each human-object pair in isolation, thereby ignoring the contextual information in the scene. In light of this, several methods have been proposed to capture the contextual cues through iterative message passing  [34, 43] or attentional graph convolutional networks  [44]. However, existing approaches rely on complex appearance-based features to encode the human-object relation (e.g., deep features extracted from a union of two boxes) and do not exploit the informative spatial cues. In addition, the contexts are aggregated via a densely connected graph (where the nodes represent all the detected objects).

In this paper, we first propose to use spatial-semantic representation to describe each human-object pair. Specifically, our spatial-semantic representation encodes (1) the relative spatial layout between a person and an object and (2) the semantic word embedding of the object category. Using spatial-semantic representation for HOI prediction has two main advantages: First, it is invariant to complex appearance variations. Second, it enables knowledge transfer among object classes and helps with rare interaction during training and inference.

While such representations are informative, predicting HOI in isolation fails to leverage the contextual cues. In the example of Fig. 1, a model might struggle to recognize that the person (in the red box) is hitting the baseball, by using only the spatial-semantic features from this particular human-object pair. Such ambiguity, however, may be alleviated if given the relation among different HOIs from the same person, e.g., this person is wearing a baseball helmet and holding a baseball bat. Similarly, we can exploit the relations among different HOIs from the same object. For example, a model may recognize both persons are holding the same baseball bat when making prediction independently. Knowing that the person (red box) is more likely to hold the baseball bat reduces the probability of another person (blue box) holding the same baseball bat. Inspired by these observations, we construct a human-centric and an object-centric HOI subgraph and apply attentional graph convolution to encode and aggregate the contextual information. We refer to our method as Dual Relation Graph (DRG).

Our Contributions

  • We propose Dual Relation Graph, an effective method to capture and aggregate contextual cues for improving HOI predictions.

  • We demonstrate that using the proposed spatial-semantic representation alone (without using appearance features) can achieve competitive performance compared to the state-of-the-art.

  • We conduct extensive ablation study of our model, identifying contributions from individual components and exploring different model design choices.

  • We achieve competitive results compared with the state-of-the-art on the VCOCO and HICO-DET datasets.

Fig. 2.
figure 2

Leveraging contextual information. Given object detections in the scene (a), existing HOI detection algorithms only perform independent prediction for each Human-object pair (b), ignoring the rich contextual cues. Recent methods in visual relationship detection (or scene graph generation) perform joint inference on a densely connected graph (c). While being general, the large number of relations among the dense connections makes the learning and inference on such a graph challenging. In contrast, our work leverages the human/object-centric graph to focus only on relevant contexts for improved HOI detection (d).

2 Related Work

Human-Object Interaction Detection. The task of human-object interaction detection aims to localize persons, object instances, as well as recognize the interactions (if any) between each pair of a person and an object. State-of-the-art HOI detection algorithms generally rely on two types of visual cues: (1) appearance features of the detected persons and objects (e.g., using the ROI pooling features extracted from a ConvNet) and (2) the spatial relationship between each human-object pair (e.g., using the bounding box transformation between the agent and the object  [12,13,14], a two-channel interaction pattern  [3, 9], or modeling the mutual contexts of human pose and object  [14, 24, 46]). Recent advances focus on incorporating contexts to resolve potential ambiguity in interaction prediction based on independent human-object pairs, including pairwise body-parts  [7, 40] or object-parts  [51], instance-centric attention [9, 41], or message passing on a graph  [34]. Our work shares similar spirits with these recent efforts as we also aim to capture contextual cues. The key difference lies in that the above approaches learn to aggregate contextual information from the other objects, body parts, or the scene background, while our method exploits relations among different HOIs to refine the predictions.

Inspired by the design of two-stage object detectors  [35], recent works also show that filtering out candidate pairs with no relations using a relation proposal network  [44] or an interactiveness network  [24] improves the performance. Our method does not train an additional network for pruning unlikely relations. We believe that incorporating such a strategy may lead to further improvement.

Recent advances in HOI detection focus on tackling the long-tailed distributions of HOI classes. Examples include transferring knowledge from seen categories to unseen ones by an analogy transformation  [31], performing data augmentation of semantically similar objects  [1], or leveraging external knowledge graph  [20]. While we do not explicitly address rare HOI classes, our method shows a small performance gap between rare and non-rare classes. We attribute this to the use of our abstract spatial-semantic representation.

Visual Relationship Detection. Many recent efforts have been devoted to detecting visual relationships  [2, 5, 17, 22, 31, 50, 52]. Unlike object detection, the number of relationship classes can be prohibitively large due to the compositionality of object pairs, predicates, and limited data samples. To overcome this issue, some forms of language prior have been applied  [27, 33]. Our focus in this work is on one particular class of relationship: human-centric interactions. Compared with other object classes, the possible interactions (the predicate) between a person and objects are significantly more fine-grained.

Scene Graph. A scene graph is a graphical structure representation of an image where objects are represented as nodes, and the relationships between objects are represented as edges  [30, 43,44,45, 49]. As the scene graph captures richer information beyond categorizing scene types or localizing object instances, it has been successfully applied to image retrieval  [19], captioning  [23], and generation  [18]. Recent advances in scene graph generation leverage the idea of iterative message passing to capture contextual cues and produce a holistic interpretation of the scene  [34, 43, 44, 49]. Our work also exploits contextual information but has the following distinctions: (1) Unlike existing methods that apply message passing to update appearance features (e.g., the appearance feature extracted from the union of human-object pair) at each step, we use an abstract spatial-semantic representation with an explicit encoding of relative spatial layout. (2) In contrast to prior works that use a single densely connected graph structure where edges connecting all possible object pairs, we operate on human-centric and object-centric subgraphs to focus on relevant contextual information specifically for HOI. Figure 2 highlights the differences between methods that capture contextual cues.

The mechanisms for dynamically capturing contextual cues for resolving ambiguity in local predictions have also been successfully applied to sequence prediction  [38], object detection  [16], action recognition  [10, 37, 42], and HOI detection  [9]. Our dual relation graph shares a similar high-level idea with these approaches but with a focus on exploiting the contexts of spatial-semantic representations.

Visual Abstraction. The use of visual abstraction helps direct the focus to study the semantics of an image  [53]. Visual abstraction has also been applied to learn common sense  [39], forecasting object dynamics  [8], and visual question answering  [47]. Our work leverages the contexts of an abstract representation between human-object pairs for detecting HOIs.

Spatial-Semantic Representation. Spatial-semantic representation has also been applied in other problem domains such as image search  [28], multi-class object detection  [6], and image captioning  [48].

Fig. 3.
figure 3

Overview of the proposed model. Our network consists of three streams (human, object, and spatial-semantic). The human and object stream leverage the appearance feature \(\mathbf {f}_h\) and \(\mathbf {f}_o\). The spatial-semantic stream makes a prediction from the abstract spatial-semantic feature \(\mathbf {x}\). We apply our proposed dual relation graph (DRG) to this stream. The three streams predict the scores \(s_h^a\), \(s_o^a\), \(s_{spH}^a\) and \(s_{spO}^a\), which are fused to form final prediction.

3 Method

In this section, we present our network for HOI detection (Fig. 3). We start with an overview of our network (Sect. 3.1). We then introduce the spatial-semantic representation (Sect. 3.2) and describe how we can leverage the proposed Dual Relation Graph (DRG) to propagate contextual information (Sect. 3.3). Finally, we outline our inference (Sect. 3.4) and the training procedure (Sect. 3.5).

3.1 Algorithm Overview

Figure 3 provides a high-level overview of our HOI detection network. We decompose the HOI detection problem into two steps: (1) object detection and (2) HOI prediction. Following Gao et al.  [9], we first apply an off-the-shelf object detector Faster R-CNN  [35] to detect all the human/object instances in an image. We denote \(\mathbb {H}\) as the set of human detections, and \(\mathbb {O}\) as the set of object detections. Note that “person” is also an object category. We denote \(b_h\) as the detected bounding box for a person and \(b_o\) for an object instance. We use \(s_h\) and \(s_o\) to denote the confidence scores produced by the object detector for a detected person \(b_h\) and an object \(b_o\), respectively. Given the detected bounding boxes \(b_h\) and \(b_o\), we first extract the ROI pooled features and pass them into the human and object stream. We then pass the detected bounding boxes as well as the object category information to the spatial-semantic stream. We apply the proposed Dual Relation Graph (DRG) in the spatial-semantic stream. Lastly, we fuse the action scores from the three streams (human, object, and spatial-semantic) to produce our final predictions.

Human and Object Stream. Our human/object streams follow the standard object detection pipeline for feature extraction and classification. For each ROI pooled human/object feature, we pass it into a one-layer MLP followed by global average pooling and obtain the human appearance feature \(\mathbf {f}_h\) and the object appearance feature \(\mathbf {f}_o\). We then apply a standard classification layer to obtain the A-dim action scores \(s_h^a\) (from human stream) and \(s_o^a\) (from object stream).

Spatial-Semantic Stream. Our inputs to this stream are the spatial-semantic features (described in Sect. 3.2). In an image, we pair all the detected persons in \(\mathbb {H}\) with all the objects in \(\mathbb {O}\), and extract spatial-semantic features for each human-object pair. We then pass all the features into our proposed Dual Relation Graph (Sect. 3.3) to aggregate the contextual information and produce updated features for each human-object pair. Our dual relation graph consists of a human-centric subgraph and an object-centric subgraph. These two subgraphs produce the action scores, \(s_{spH}^a\) and \(s_{spO}^a\).

Fig. 4.
figure 4

HOI detection using Dual Relation Graph. (a) The input to our model are the detected objects in the given image. We denote \(\mathbb {H}\) as the set of human detections, and \(\mathbb {O}\) as the set of object detections. (b) We construct a relation graph from the detections where the two sets are \(\mathbb {H}\) and \(\mathbb {O}\). (c) For each human node h in \(\mathbb {H}\), we obtain a human-centric sub-graph where all nodes in \(\mathbb {O}\) are connected to h. Similarly, we can obtain an object-centric sub-graph for each object node o in \(\mathbb {O}\). Note that “person” is also an object category. For simplicity, we do not show it in the figure. (d) In order to predict HOIs, we need to construct the HOI graph explicitly. Taking human sub-graph for example, we insert an HOI node x between human node h and object node o. We then connect all the HOI nodes and obtain the human-centric HOI sub-graph and the object-centric HOI sub-graph. (e) We iteratively update the HOI node feature via a trainable attentional graph convolutional network. This helps to aggregate the contextual information. (f) We fuse the scores from both sub-graphs and make the final HOI prediction.

3.2 Spatial-Semantic Representation

We leverage the abstract visual representation of human-object pair for HOI prediction. The visual abstraction of a human-object pair allows us to construct representations that are invariant to intra-class appearance variations. In the context of human-object interaction, we consider the two important visual abstractions: (1) spatial relationship and (2) object category.Footnote 1

Capturing Pairwise Spatial Relationship. Following [3, 9], we use the two-channel binary image representation to model the spatial relationship between a person and an object. To do so, we take the union of the two bounding boxes as a reference and rescale it to a fixed size. A binary image with two channels can then be created by filling value ones within the human bounding box in the first channel and filling value ones within the object bounding box in the second channel. The remaining locations are then filled with value 0. We then feed these two-channel binary images into a simple two-layer ConvNet to extract the spatial relation feature.

Capturing Object Semantics. We find that using spatial features by itself leads to poor results in predicting the interaction. To address this issue, we augment the spatial feature with the word embedding of each object’s category, vector(o), using fastText  [29]. Let \(\mathbf {x}_{ij}\) denote the spatial-semantic feature between the i-th person and the j-th object. We construct \(\mathbf {x}_{ij} \in \mathbb {R}^{5708}\) by concatenating (1) the spatial feature and (2) the 300-dimensional word embedding vector.

3.3 Dual Relation Graph

Here, we introduce the Dual Relation Graph for aggregating the spatial-semantic features. Figure 4 illustrates the overall process.

Relation Graph. Given the object detection results, i.e., instances in \(\mathbb {H}\) and \(\mathbb {O}\), we construct a relation graph (Fig. 4b). There are two types of nodes in this graph, \(\mathbb {H}\) (human) and \(\mathbb {O}\) (object). For each node \(\textit{h}\) in \(\mathbb {H}\), we connect it to all the other nodes in \(\mathbb {O}\). Similarly, for each node \(\textit{o}\) in \(\mathbb {O}\), we connect it to all nodes in \(\mathbb {H}\).

Human-Centric Subgraph and Object-Centric Subgraph. Unlike previous methods [34, 44], we do not use the densely connected graphs. To exploit the relation among different HOIs performed by the same person, we construct a human-centric subgraph. Similarly, we construct an object-centric subgraph for the HOIs performed on the same object (Fig. 4c). So far, each node stands for an object instance detection. To explicitly represent the HOI, we insert an HOI node \( {x_{ij}}\) between each paired human node \( {h_i}\) and object node \( {o_j}\). We then connect all the HOI nodes and obtain human-centric HOI subgraph and object-centric HOI subgraph (Fig. 4d). We use the before mentioned spatial-semantic feature \(\mathbf {x}_{ij}\) to encode each HOI node between the i-th person and the j-th object.

Contextual Feature Aggregation. With these two HOI subgraphs, we follow a similar procedure for propagating and aggregating features as in relation network  [16], non-local neural network  [42], and attentional graph convolutional network  [44].

Human-Centric HOI Subgraph. To update node \( {x_{ij}}\), we aggregate all the spatial-semantic feature of the nodes involving the same i-th person \(\{x_{ij^\prime }| j^\prime \in \mathcal {N}(j)\}\). The feature aggregation can be written as:

$$\begin{aligned} \mathbf {x}_{ij}^{(l+1)} = \sigma \left( \mathbf {x}_{ij}^{(l)} + \sum _{j^\prime \in \mathcal {N}(j)} \alpha _{j j^\prime } W \mathbf {x}_{ij^\prime }^{(l)}\right) , \end{aligned}$$
(1)

where \(W \in \mathbb {R}^{5708 \times 5708}\) is a learned linear transformation that projects features into the embedding space, \(\alpha _{i j^\prime }\) is a learned attention weight.

We can rewrite this equation compactly with matrix operation:

$$\begin{aligned} \mathbf {x}_{ij}^{(l+1)}&= \sigma \left( W X^{(l)} \varvec{\alpha }_j\right) \end{aligned}$$
(2)
$$\begin{aligned} u_{jj^\prime }&= \left( W_q \mathbf {x}_{ij^\prime }^{(l)}\right) ^\top \left( W_k \mathbf {x}_{ij}^{(l)} \right) / \sqrt{d_k} \end{aligned}$$
(3)
$$\begin{aligned} \varvec{\alpha }_j&= \text {softmax} (\varvec{u}_{j}), \end{aligned}$$
(4)

where \(W_q, W_k \in \mathbb {R}^{1024 \times 5708}\) are linear projections that project feature into a query and a key embedding space. Following  [38], we calculate the attention weights using scaled dot-product, normalized by \(\sqrt{d_k}\) where \(d_k = 1024\) is the dimension of the key embedding space. We do not directly use the aggregated feature \(\sigma \left( W X^{(l)} \varvec{\alpha }_j\right) \) as our output updated feature. Instead, we add it back to the original spatial-semantic feature \(\mathbf {x}_{ij}^{(l)}\). We then pass the addition through a LayerNorm to get the final aggregated feature on the human-centric subgraph.

$$\begin{aligned} \mathbf {x}_{ij}^{(l+1)} = \text {LayerNorm}\left( \mathbf {x}_{ij}^{(l)} + \sigma \left( W X^{(l)} \varvec{\alpha }_j\right) \right) . \end{aligned}$$
(5)

The linear transformation W does not change the dimension of the input feature, thus the output \(\mathbf {x}_{ij}^{(l+1)}\) has the same size as input \(\mathbf {x}_{ij}^{(l)}\). As a result, we can perform several iterations of feature aggregation (Fig. 4e). We explore the effectiveness of more iteration in Table 3(a).

Object-Centric HOI Subgraph. Similarly, to update node \( {x_{ij}}\), we aggregate all the spatial-semantic feature of the nodes which involved the same object \(\{x_{i^\prime j}| i^\prime \in \mathcal {N}(i)\}\). The two subgraphs have independent weights and aggregate contextual information independently.

3.4 Inference

For each human-object bounding box pair (\(b_h\), \(b_o\)) in image I, we predict the score \(S^a_{h,o}\) for each action \(a \in \left\{ 1, \cdots , A \right\} \), where A denotes the total number of possible actions. The final score \(S^a_{h,o}\) depends on (1) the confidence for the individual object detections (\(s_h\) and \(s_o\)), (2) the prediction score from the appearance of the person \(s^a_h\) and the object \(s^a_o\), and (3) the prediction score based on the aggregated spatial-semantic feature, using human-centric and object-centric subgraph, \(s_{spH}^a\) and \(s_{spO}^a\). We compute the HOI score \(S^a_{h,o}\) for the human-object pair (\(b_h\), \(b_o\)) as

$$\begin{aligned} S^a_{h,o}&= s_h \cdot s_o \cdot s^a_{h} \cdot s^a_{o} \cdot s^a_{spH} \cdot s^a_{spO} \end{aligned}$$
(6)

Note that we are not able to obtain the action scores using object \(s^a_{o}\) or the spatial-semantic stream for some classes of actions as they do not involve any objects (e.g., walk, smile). For those cases, we use only the score \(s^a_h\) from the human stream. For those actions, our final scores are \(s_h \cdot s^a_h\).

3.5 Training

HOI detection is a multi-label classification problem because a person can simultaneously perform different actions on different objects, e.g., sitting on a chair and reading a book. Thus, we minimize the cross-entropy loss for each individual action class between the ground-truth action label and the score produced from each stream. The total loss is the summation of the loss at each stream.

Fig. 5.
figure 5

Sample HOI detections on V-COCO (first row) and HICO-DET (second row) test set.

4 Experimental Results

In this section, we first outline our experimental setup, including datasets, metrics, and implementation details. We then report the quantitative results on two large-scale HOI benchmark datasets and compare the performance with the state-of-the-art HOI detection algorithms. Next, we show sample visual results on HOI detection. We conduct a detailed ablation study to quantify the contributions from individual components and validate our design choices. More results can be found in the supplementary material. We will make the source code and pre-trained models publicly available to foster future research.

Table 1. Comparison with the state-of-the-art on the V-COCO test set. The best performance is in and the second best is . Character \(^\star \) indicates that the method uses both VCOCO and HICO-DET training data. “S-S only” shows the performance of our spatial-semantic stream.

4.1 Experimental Setup

Datasets. V-COCO dataset  [13] is constructed by augmenting the COCO dataset  [26] with additional human-object interaction annotations. Each person is annotated with a binary label vector for 29 different action categories (five of them do not involve associated objects). HICO-DET  [4] is a larger dataset containing 600 HOI categories over 80 object categories (same as [26]) with more 150 K annotated instances of human-object pairs. For applying our method on the HICO-DET dataset, we disentangle the 600 HOI categories into 117 object-agnostic action categories and train our network over these 117 action categories. At test time, we then combine the predicted action and the detected object and convert them back to the original 600 HOI classes. Note that the evaluation for the HICO-DET dataset remains the same.

Evaluation Metrics. To evaluate the performance of our model, we adopt the commonly used role mean average precision (role mAP)  [13] for both V-COCO and HICO datasets. The goal is to detect and correctly predict the \(\langle \) \(\texttt {human, verb, object}\) \(\rangle \) triplet. We consider a triplet as true positive if and only if it localizes the human and object accurately (i.e., with IoUs \(\ge 0.5\) w.r.t the ground truth annotations) and predicts the action correctly.

Table 2. Comparison with the state-of-the-art on HICO-DET test set. The best performance is in and the second best is . Character \(^\star \) indicates that the method uses both VCOCO and HICO-DET training data. For the object detector, “COCO” means that the detector is trained on COCO, while “HICO-DET” means that the detector is first pre-trained on COCO and then further fine-tuned on HICO-DET.

Implementation Details. We build our network with the publicly available PyTorch framework. Following Gao et al.  [9], we use the Detectron [11] with a feature backbone of ResNet-50 to generate human and object bounding boxes. For VCOCO, we conduct an ablation study on the validation split to determine the best threshold. We keep the detected human boxes with scores \(s_h\) higher than 0.8 and object boxes with scores \(s_o\) higher than 0.1. For HICO-DET, since there is no validation split available, we follow the setting in [32]. We use the score threshold 0.6 to filter out unreliable human boxes and threshold 0.4 to filter out unconfident object boxes. To augment the training data, we apply random spatial jitterring to the human and object bounding boxes and ensure that the IOU with the ground truth bounding box is greater than 0.7. We pair all the detected human and objects, and regard those who are not ground truth as negative training examples. We keep the negative to positive ratio to three.

We initialize our appearance feature backbone with the COCO pre-trained weight from Mask R-CNN [15]. We perform two iterations of feature aggregation on both human-centric and object-centric subgraphs. We train the three streams (human appearance, object appearance, and spatial-semantic) using the V-COCO train set. We use early stopping criteria by monitoring the validation loss. We train our network with a learning rate of 0.0025, a weight decay of 0.0001, and a momentum of 0.9 on both the V-COCO train set and HICO-DET train set. Training our network takes 14 h on a single NVIDIA P100 GPU on V-COCO and 24 h on HICO-DET. At test time, our model runs at 3.3 fps for VCOCO and 5 fps for HICO-DET.

Fig. 6.
figure 6

More iteration of feature aggregation leads to a more accurate prediction. The human-centric and object-centric subgraph in the spatial-semantic stream propagates contextual information to produce increasingly accurate HOI predictions.

4.2 Quantitative Evaluation

We report the main quantitative results in terms of \(AP_{role}\) on V-COCO in Table 1 and HICO-DET in Table 2. For the V-COCO dataset, our method compares favorably against state-of-the-art algorithms  [24, 41, 51] except PMFNet [40], which uses human pose as an additional feature. Since pose estimation required additional training data (with pose annotations), we expect to see performance gain using human pose. PMFNet [40] also reports the \(AP_{role}\) without human pose, which is 2.4 mAP lower to our method. We also note that the spatial-semantic stream alone without using any visual features achieves a competitive performance (47.1 mAP) when compared with the state-of-the-art. This highlights the effectiveness of the abstract spatial-semantic representation and contextual information. Compared with methods that perform joint inference on densely connected graph  [34], our approach produces significant performance gains.

For the HICO-DET dataset, our method also achieves competitive performance with state-of-the-art methods  [14, 24, 31, 40]. Our method achieves the best performance for the rare categories, showing that our method handles the long-tailed distributions of HOI classes well.

We note that the current best performing model  [1] uses an object detector which is fine-tuned on HICO-DET train set using the annotated object bounding boxes. For a fair comparison, we also fine-tune our object detector on HICO-DET and report our result. Note that we do not re-train our model, but only replace the object detector at the test time.

Here, the large performance gain from fine-tuning the object detector may not reflect the progress on the HOI detection task. This is because the objects (and the associated HOIs) in the HICO-DET dataset are not fully annotated. Using a fine-tuned object detector can thus improve the HOI detection performance by exploiting such annotation biases.

4.3 Qualitative Evaluation

HOI Detection Results. Here we show sample results on the V-COCO dataset and the HICO-DET dataset in Fig. 5. We highlight the detected person and the associated object with red and green bounding boxes, respectively.

Table 3. Ablation study on the V-COCO val set. We show the role mAP \(AP_{role}\).

Visualizing the Effect of the Dual Relation Graph. In Fig. 6, we show the effectiveness of the proposed DRG. In the first two rows, we show that by aggregating contextual information, using the human-centric subgraph produces more accurate HOI predictions. Another example in the top right image indicates that the human-centric subgraph can also suppress the scores for unrelated human-object pairs. In the bottom two rows, we show four examples of how the object-centric subgraph propagates contextual information in each step to produce increasingly more accurate HOI predictions. For instance, the bottom right images show that for a person and an object without interaction, our model learns to suppress the predicted score by learning from the relationship of other HOI pairs associated with this particular object (laptop). In this example, the model starts with predicting a high score for the woman working on a computer. By learning from the relationship between the man and the computer, our model suppresses the score in each iteration.

4.4 Ablation Study

We examine several design choices using the V-COCO val set.

More Iteration of Feature Aggregation. Table 3(a) shows the performance using different iterations of feature aggregation. For either human-centric or object-centric subgraph, using more iterations of feature aggregation improves the overall performance. This highlights the advantages of exploiting contextual information among different HOIs. Performing feature aggregation on both subgraphs further improves the final performance.

Effectiveness of Each Subgraph. To validate the effectiveness of the proposed subgraph, we show different variants of our model in Table 3(c). Adding only human-centric subgraph improves upon the baseline model (without using any subgraph) by 0.96 absolute mAP, while adding only object-centric subgraph gives a 0.64 absolute mAP. More importantly, our results show that the performance gain of each subgraph is complementary to each other. To further validate the effectiveness of the object-centric subgraph, we show in Table 3(d) the breakdown of Table 3(c) in terms of the number of persons in the scene. The object-centric subgraph is less effective for cases with few people. For example, if there is only one person, the object-centric subgraph has no effect. For images with a moderate amount of persons (4–6), however, our object-centric subgraph shows a clear 0.98 mAP gain. As the number of persons getting larger (7+), the object-centric subgraph shows a relatively smaller improvement due to clutter. Among the 2,867 testing images, 68% of them have only 1–3 persons. As a result, we do not see significant overall improvement.

Spatial-Semantic Representation. To demonstrate the advantage and effectiveness of the use of the abstract spatial-semantic representation, we show in Table 3(b) the comparison with alternative features, e.g., word2vec (as used in [27]) or appearance-based features. By using our spatial-semantic representation in the dual relation graph, we achieve 51.37 mAP. This shows a clear margin over the other alternative options, highlighting the contribution of spatial-semantic representation.

Fig. 7.
figure 7

Failure cases of our method.

4.5 Limitations

While we demonstrated improved performance, our model is far from perfect. Below, we discuss two main limitations of our approach, with examples in Fig. 7.

First, we leverage the off-the-shelf object detector to detect object instances in an image. The object detection does not benefit from the rich contextual cues captured by our method. We believe that a joint end-to-end training approach may help reduce this type of errors.

Second, our model may be confused by plausible spatial configuration and predicts incorrect action. In the third image, our model predicts that the person is sitting on a bench even though our model confidently predicts this person is standing and catching a Frisbee. Capturing the statistics of co-occurring actions may resolve such mistakes.

5 Conclusions

In this paper, we present a Dual Relation Graph network for HOI detection. Our core idea is to exploit the global object layout as contextual cues and use a human-centric as well as an object-centric subgraph to propagate and integrate rich relations among individual HOIs. We validate the efficacy of our approach on two large-scale HOI benchmark datasets and show our model achieves a sizable performance boost over the state-of-the-art algorithms. We also find that using the abstract spatial-semantic representation alone (i.e., without the appearance features extracted from a deep CNN) yields competitive accuracy, demonstrating a promising path of activity understanding through visual abstraction.