This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: The University of Sydney, Australia
22institutetext: JD Explore Academy, China
22email: [email protected], [email protected],[email protected]

Discovering Human-Object Interaction Concepts via Self-Compositional Learning

Zhi Hou 11    Baosheng Yu 11    Dacheng Tao 1122
Abstract

A comprehensive understanding of human-object interaction (HOI) requires detecting not only a small portion of predefined HOI concepts (or categories) but also other reasonable HOI concepts, while current approaches usually fail to explore a huge portion of unknown HOI concepts (i.e., unknown but reasonable combinations of verbs and objects). In this paper, 1) we introduce a novel and challenging task for a comprehensive HOI understanding, which is termed as HOI Concept Discovery; and 2) we devise a self-compositional learning framework (or SCL) for HOI concept discovery. Specifically, we maintain an online updated concept confidence matrix during training: 1) we assign pseudo labels for all composite HOI instances according to the concept confidence matrix for self-training; and 2) we update the concept confidence matrix using the predictions of all composite HOI instances. Therefore, the proposed method enables the learning on both known and unknown HOI concepts. We perform extensive experiments on several popular HOI datasets to demonstrate the effectiveness of the proposed method for HOI concept discovery, object affordance recognition and HOI detection. For example, the proposed self-compositional learning framework significantly improves the performance of 1) HOI concept discovery by over 10% on HICO-DET and over 3% on V-COCO, respectively; 2) object affordance recognition by over 9% mAP on MS-COCO and HICO-DET; and 3) rare-first and non-rare-first unknown HOI detection relatively over 30% and 20%, respectively. Code is publicly available at https://github.com/zhihou7/HOI-CL.

Keywords:
Human-Object Interaction, HOI Concept Discovery, Object Affordance Recognition

1 Introduction

Human-object interaction (HOI) plays a key role in analyzing the relationships between humans and their surrounding objects [23], which is of great importance for deep understanding on human activities/behaviors. Human-object interaction understanding has attracted extensive interests from the community, including image-based [8, 6, 19, 40, 55], video-based visual relationship analysis [13, 42], video generation [44], and scene reconstruction [66]. However, the distribution of HOI samples is naturally long-tailed: most interactions are rare and some interactions do not even occur in most scenarios, since we can not obtain an interaction between human and object until someone conducts such action in real-world scenarios. Therefore, recent HOI approaches mainly focus on the analysis of very limited predefined HOI concepts/categories, leaving the learning on a huge number of unknown HOI concepts [10, 3] poorly investigated, including HOI detection and object affordance recognition [53, 27, 28]. For example, there are only 600 HOI categories known in HICO-DET [7], while we can find 9,360 possible verb-object combinations from 117117 verbs and 8080 objects.

Refer to caption
Figure 1: An illustration of unknown HOI detection via concept discovery. Given some known HOI concepts (e.g., “drink_\_with cup”, “drink_\_with bottle”, and “hold bowl”), the task of concept discovery aims to identify novel HOI concepts (i.e., reasonable combinations between verbs and objects). For example, here we have some novel HOI concepts, “drink_\_with wine_\_glass”, “fill bowl”, and “fill bottle”. Specifically, the proposed self-compositional learning framework jointly optimizes HOI concept discovery and HOI detection on unknown concepts in an end-to-end manner.

Object affordance is closely related to HOI understanding from an object-centric perspective. Specifically, two objects with similar attributes usually share the same affordance, i.e. , humans usually interact with similar objects in a similar way [20]. For example, cup, bowl, and bottle share the same attributes (e.g., hollow), and all of these objects can be used to “drink with”. Therefore, object affordance [20, 28] indicates whether each action can be applied into an object, i.e. , if a verb-object combination is reasonable, we then find a novel HOI concept/category. An illustration of unknown HOI detection via concept discovery is shown in Fig. 1. Recently, it has turned out that an HOI model is not only capable of detecting interactions, but also able to recognize object affordances [28], especially novel object affordances using the composite HOI features. Particularly, novel object affordance recognition also indicates discovering novel reasonable verb-object combinations or HOI concepts. Inspired by this, we can introduce a simple baseline for HOI concept discovery by averaging the affordance predictions of training dataset into each object category [28].

Nevertheless, there are two main limitations when directly utilizing object affordance prediction [28] for concept discovery. First, the affordance prediction approach in [28] is time-consuming and unsuitable to be utilized during training phrase, since it requires to predict all possible combinations of verbs and objects using the whole training set. By contrast, we introduce an online HOI concept discovery method, which is able to collect concept confidence in a running mean manner with verb scores of all composite features in mini-batches during training. Second, also more importantly, the compositional learning approach [28] merely optimizes the composite samples with known concepts (e.g., 600 categories on HICO-DET), ignoring a large number of composite samples with unknown concepts (unlabeled composite samples). As a result, the model is inevitably biased to known object affordances (or HOI concepts), and leads to the similar inferior performance to the one in Positive-Unlabeled learning [14, 16, 49]. That is, without negative samples for training, the network will tend to predict high confidence on those impossible verb-object combinations or overfit verb patterns (please refer to Appendix A for more analysis). Considering that the online concept discovery branch is able to predict concept confidence during optimization, we can then construct pseudo labels [37] for all composite HOIs belonging to either known or unknown categories. Inspired by this, we introduce a self-compositional learning strategy (or SCL) to jointly optimize all composite representations and improve concept predictions in an iterative manner. Specifically, SCL combines the object representations with different verb representations to compose new samples for optimization, and thus implicitly pays attention to the object representations and improves the discrimination of composite representations. By doing this, we can improve the object affordance learning, and then facilitate the HOI concept discovery.

Our main contributions can be summarized as follows: 1) we introduce a new task for a better and comprehensive understanding on human-object interactions; 2) we devise a self-compositional learning framework for HOI concept discovery and object affordance recognition simultaneously; and 3) we evaluate the proposed approach on two extended benchmarks, and it significantly improves the performance of HOI concept discovery, facilitates object affordance recognition with HOI model, and also enables HOI detection with novel concepts.

2 Related Work

2.1 Human-Object Interaction

HOI understanding [23] is of great importance for visual relationship reasoning [61] and action understanding [5, 67]. Different approaches have been investigated for HOI understanding from various aspects, including HOI detection [7, 39, 40, 68, 34, 9, 71, 55, 65], HOI recognition [8, 33, 30], video HOI [13, 32], compositional action recognition [42], 3D scene reconstruction [66, 12], video generation [44], and object affordance reasoning [17, 28]. Recently, compositional approaches (e.g., VCL [27]) have been intensively proposed for HOI understanding using the structural characteristic [33, 27, 44, 38, 28]. Meanwhile, DETR-based methods (e.g., Qpic [55]) achieve superior performance on HOI detection. However, these approaches mainly consider the perception of known HOI concepts, and pay no attention to HOI concept discovery. To fulfill the gap between learning on known and unknown concepts, a novel task, i.e. , HOI concept discovery, is explored in this paper. Currently, zero-shot HOI detection also attracts massive interests from the community [53, 2, 48, 27, 29]. However, those approaches merely consider known concepts and are unable to discover HOI concepts. Some HOI approaches [48, 2, 59, 58] expand the known concepts via leveraging language priors. However, that is limited to existing knowledge and can not discover concepts that never appear in the language prior knowledge. HOI concept discovery is able to address the problem, and enable unknown HOI concept detection.

2.2 Object Affordance Learning

The notation of affordance is formally introduced in  [20], where object affordances are usually those action possibilities that are perceivable by an actor [45, 20, 21]. Noticeably, the action possibilities of an object also indicate the HOI concepts related to the object. Therefore, object affordance can also represent the existence of HOI concepts. Recent object affordance approaches mainly focus on the pixel-level affordance learning from human interaction demonstration [36, 18, 17, 25, 43, 15, 64]. Yao et al. [63] present a weakly supervised approach to discover object functionalities from HOI data in the musical instrument environment. Zhu et al. [70] introduce to reason affordances in knowledge-based representation. Recent approaches propose to generalize HOI detection to unseen HOIs via functionality generalization [2] or analogies [48]. However those approaches focus on HOI detection, ignoring object affordance recognition. Specifically, Hou et al. [28] introduce an affordance transfer learning (ATL) framework to enable HOI model to not only detect interactions but also recognize object affordances. Inspired by this, we further develop a self-compositional learning framework to facilitate the object affordance recognition with HOI model to discover novel HOI concepts for downstream HOI tasks.

2.3 Semi-Supervised Learning

Semi-supervised learning is a learning paradigm for constructing models that use both labeled and unlabeled data [62]. There are a wide variety of Deep Semi-Supervised Learning methods, such as Generative Networks [35, 54], Graph-Based methods [57, 22], Pseudo-Labeling methods [37, 60, 26]. HOI concept discovery shares a similar characteristic to semi-supervised learning approaches. HOI concept discovery has instances of labeled HOI concepts, but no instances of unknown concepts. We thus compose HOI representations for unknown concepts according to [50]. With composite HOIs, concept discovery and object affordance recognition can be treated as PU learning [14]. Moreover, HOI concept discovery requires to discriminate whether the combinations (possible HOI concepts) are reasonable and existing. Considering each value of the concept confidences also represents the possibility of the composite HOI, we construct pseudo labels [37, 50] for composite features from the concept confidence matrix, and optimize the composite HOIs in an end-to-end way.

3 Approach

In this section, we first formulate the problem of HOI concept discovery and introduce the compositional learning framework. We then describe a baseline for HOI concept discovery via affordance prediction. Lastly, we introduce the proposed self-compositional learning framework for online HOI concept discovery and object affordance recognition.

Refer to caption
Figure 2: Illustration of Self-Compositional Learning for HOI Concept Discovery. Specifically, following [27], verb and object features are extracted via RoI-Pooling from union box and object box respectively, which are then used to construct HOI features in HOI branch according to HOI annotation. Following [27], for SCL, verb and object features are further mutually combined to generate composite HOI features. Then, the feasible composite HOI features belonging to the known concepts are directly used to train the network in Compositional Branch. Here the classifier predicts verb classes directly. Meanwhile, we update the concept confidence 𝐌RNv×No\mathbf{M}\in R^{N_{v}\times N_{o}}, where NvN_{v} and NoN_{o} are the number of verb classes and object classes respectively, with the predictions of all composite HOI features. The concept discovery branch is optimized via a self-training approach to learn from composite HOI features with the concept confidence 𝐌\mathbf{M}.

3.1 Problem Definition

HOI concept discovery aims to discover novel HOI concepts/categories using HOI instances from existing known HOI categories. Given a set of verb categories 𝒱\mathcal{V} and a set of object categories 𝒪\mathcal{O}, let 𝒮=𝒱×𝒪\mathcal{S}=\mathcal{V}\times\mathcal{O} indicate the set of all possible verb-object combinations. Let 𝒮k\mathcal{S}^{k}, 𝒮u\mathcal{S}^{u}, and 𝒮o\mathcal{S}^{o} denote three disjoint sets, known HOI concepts, unknown HOI concepts, and invalid concepts (or impossible verb-object combinations), respectively. That is, we have 𝒮k𝒮u=\mathcal{S}^{k}\cap\mathcal{S}^{u}=\varnothing and 𝒮k𝒮u=𝒮\mathcal{S}^{k}\cup\mathcal{S}^{u}=\mathcal{S} if 𝒮o=\mathcal{S}^{o}=\varnothing. Let 𝒯={(hi,ci)}i=1L\mathcal{T}=\{(h_{i},c_{i})\}_{i=1}^{L} indicate the training dataset, where hih_{i} is a HOI instance (i.e. , verb-object visual representation pair), ci𝒮kc_{i}\in\mathcal{S}^{k} indicates the label of the ii-th HOI instance and LL is the total number of HOI instance.

We would also like to clarify the difference between the notations of “unknown HOI categories” and “unseen HOI categories” in current HOI approaches as follows. Let 𝒮z\mathcal{S}^{z} indicate the set of “unseen HOI categories” and we then have 𝒮z𝒮k\mathcal{S}^{z}\subseteq\mathcal{S}^{k}. Specifically, “unseen HOI category” indicates that the HOI concept is known but no corresponding HOI instances can be observed in the training data. Current HOI methods usually assume that unseen HOI categories 𝒮z\mathcal{S}^{z} are known HOI categories via the prior knowledge [53, 33, 48, 2, 27]. Therefore, existing HOI methods can not directly detect/recognize HOIs with unknown HOI concepts. HOI concept discovery aims to find 𝒮u\mathcal{S}^{u} from the existing HOI instances in 𝒯\mathcal{T} with only known HOI concepts in 𝒮k\mathcal{S}^{k}.

3.2 HOI Compositional Learning

Inspired by the compositional nature of HOI, i.e. , each HOI consists of a verb and an object, visual compositional learning has been intensively explored for HOI detection by combining visual verb and object representations [33, 27, 29, 28]. Let 𝐡i=𝐱vi,𝐱oi\mathbf{h}_{i}=\langle\mathbf{x}_{v_{i}},\mathbf{x}_{o_{i}}\rangle indicate a HOI instance, where 𝐱vi\mathbf{x}_{v_{i}} and 𝐱oi\mathbf{x}_{o_{i}} denote the verb and object representations, respectively. The HOI compositional learning then aims to achieve the following objective,

gh(𝐱~vi,𝐱~oi)gh(𝐱vi,𝐱oi),g_{h}(\langle\widetilde{\mathbf{x}}_{v_{i}},\widetilde{\mathbf{x}}_{o_{i}}\rangle)\approx g_{h}(\langle\mathbf{x}_{v_{i}},\mathbf{x}_{o_{i}}\rangle), (1)

where ghg_{h} indicates the HOI classifier, 𝐱vi\mathbf{x}_{v_{i}} and 𝐱oi\mathbf{x}_{o_{i}} indicate the real verb-object representation pair (i.e. , annotated HOI pair in dataset), 𝐱~vi,𝐱~oi\langle\widetilde{\mathbf{x}}_{v_{i}},\widetilde{\mathbf{x}}_{o_{i}}\rangle indicates the composite verb-object pair. Specifically, 𝐱~oi\widetilde{\mathbf{x}}_{o_{i}} can be obtained from either real HOIs [27], fabricated objects or language embedding [29, 2, 48], or external object datasets [28], while 𝐱~vi\widetilde{\mathbf{x}}_{v_{i}} can be from real HOIs (annotated verb-object pair) and language embeddings [33, 48]. As a result, when composite HOIs are similar to real HOIs, we are then able to augment HOI training samples in a compositional manner. However, current compositional approaches for HOI detection [27, 28] simply remove the composite HOI instances out of the label space, which may also remove a large number of feasible HOIs (e.g., “ride zebra” as shown Figure 2). Furthermore, the compositional approach can not only augment the training data for HOI recognition, but also provide a method to determinate whether 𝐱~vi\widetilde{\mathbf{x}}_{v_{i}} and 𝐱~oi\widetilde{\mathbf{x}}_{o_{i}} are combinable to form a new HOI or not [28], i.e. , discovering the HOI concepts.

3.3 Self-Compositional Learning

In this subsection, we introduce the proposed self-compositional learning framework for HOI concept discovery as follows. As shown in Figure 2, the main HOI concept discovery framework falls into the popular two-stage HOI detection framework [27]. Specifically, we compose novel HOI samples from pair-wise images to optimize the typical HOI branch (annotated HOIs), compositional branch (the composite HOIs out of the label space are removed [27, 28]) and the new concept discovery branch (all composite HOIs are used). The main challenge of HOI concept discovery is the lack of instances for unknown HOI concepts, but we can infer to discover new concepts according to the shared verbs and objects. Specifically, we find that the affordance transfer learning [28] can be used for not only the object affordance recognition but also the HOI concept discovery, and we thus first introduce the affordance-based method as a baseline as follows.

3.3.1 Affordance Prediction

The affordance transfer learning [28] or ATL is introduced for affordance recognition using the HOI detection model. However, it has been ignored that the affordance prediction can also enable HOI concept discovery, i.e. , predicting a new affordance for an object although the affordance is not labeled during training. We describe a vanilla approach for HOI concept discovery using affordance prediction [28]. Specifically, we predict the affordances for all objects in the training set according to  [28]. Then, we average the affordance predictions according to each object category to obtain the HOI concept confidence matrix 𝐌RNv×No\mathbf{M}\in R^{N_{v}\times N_{o}}, where each value represents the concept confidence of the corresponding combination between a verb and an object. NvN_{v} and NoN_{o} are the numbers of verb and object categories, respectively. For simplicity, we may use both vector and matrix forms of the confidence matrix 𝐌RNvNo\mathbf{M}\in R^{N_{v}N_{o}} and 𝐌RNv×No\mathbf{M}\in R^{N_{v}\times N_{o}} in this paper. Though affordance prediction can be used for HOI concept discovery, it is time-consuming since it requires to predict affordances of all objects in training set. Specifically, we need an extra offline affordance prediction process to infer concepts with the computational complexity O(N2)O(N^{2}) in [28], where NN is the number of total training HOIs, e.g., it takes 8 hours with one GPU to infer the concept matrix 𝐌\mathbf{M} on HICO-DET. However, we can treat the verb representation as affordance representation [28], and obtain the affordance predictions for all objects in each mini-batch during training stage. Inspired by the running mean manner in [31], we devise an online HOI concept discovery framework via averaging the predictions in each mini-batch.

3.3.2 Online Concept Discovery

As shown in Figure 2, we keep a HOI concept confidence vector during training, 𝐌RNvNo\mathbf{M}\in R^{N_{v}N_{o}}, where each value represents the concept confidence of the corresponding combination between a verb and an object. To achieve this, we first extract all verb and object representations among pair-wise images in each batch as 𝐱v\mathbf{x}_{v} and 𝐱o\mathbf{x}_{o}. We then combine each verb representation and all object representations to generate the composite HOI representations 𝐱h\mathbf{x}_{h}. After that, we use the composite HOI representations as the input to the verb classifier and obtain the corresponding verb predictions 𝐘^vRNN×Nv\mathbf{\hat{Y}}_{v}\in R^{NN\times N_{v}}, where NN indicates the number of real HOI instances (i.e. , verb-object pair) in each mini-batch and NNNN is then the number of all composite verb-object pairs (including unknown HOI concepts). Let 𝐘vRN×Nv\mathbf{Y}_{v}\in R^{N\times N_{v}} and 𝐘oRN×No\mathbf{Y}_{o}\in R^{N\times N_{o}} denote the label of verb representations 𝐱v\mathbf{x}_{v} and object representations 𝐱o\mathbf{x}_{o}, respectively. We then have all composite HOI labels 𝐘h=𝐘v𝐘o\mathbf{Y}_{h}=\mathbf{Y}_{v}\otimes\mathbf{Y}_{o}, where 𝐘hRNN×NvNo\mathbf{Y}_{h}\in R^{NN\times N_{v}N_{o}}, and the superscripts hh, vv, and oo indicate HOI, verb, and object, respectively. Similar to affordance prediction, we repeat 𝐘^v\mathbf{\hat{Y}}_{v} by NoN_{o} times to obtain concept predictions 𝐘^hRNN×NvNo\mathbf{\hat{Y}}_{h}\in R^{NN\times N_{v}N_{o}}. Finally, we update 𝐌\mathbf{M} in a running mean manner [31] as follows,

𝐌\displaystyle\mathbf{M} 𝐌𝐂+iNN𝐘^h(i,:)𝐘h(i,:)𝐂+iNN𝐘h(i,:),\displaystyle\leftarrow\frac{\mathbf{M}\odot\mathbf{C}+\sum_{i}^{NN}{\mathbf{\hat{Y}}_{h}(i,:)\odot\mathbf{Y}_{h}(i,:)}}{\mathbf{C}+\sum_{i}^{NN}{\mathbf{Y}_{h}(i,:)}}, (2)
𝐂\displaystyle\mathbf{C} 𝐂+iNN𝐘h(i,:),\displaystyle\leftarrow\mathbf{C}+\sum_{i}^{NN}{\mathbf{Y}_{h}(i,:)}, (3)

where \odot indicates the element-wise multiplication, 𝐘^h(i,:)𝐘h(i,:)\mathbf{\hat{Y}}_{h}(i,:)\odot\mathbf{Y}_{h}(i,:) aims to filter out predictions whose labels are not 𝐘h(i,:)\mathbf{Y}_{h}(i,:), each value of 𝐂RNvNo\mathbf{C}\in R^{N_{v}N_{o}} indicates the total number of composite HOI instances in each verb-object pair (including unknown HOI categories). Actually, 𝐘^h(i,:)𝐘h(i,:)\mathbf{\hat{Y}}_{h}(i,:)\odot\mathbf{Y}_{h}(i,:) follows the affordance prediction process [28]. The normalization with 𝐂\mathbf{C} is to avoid the model bias to frequent categories. Specifically, both 𝐌\mathbf{M} and 𝐂\mathbf{C} are zero-initialized. With the optimization of HOI detection, we can obtain the vector 𝐌\mathbf{M} to indicate the HOI concept confidence of each combination between verbs and objects.

3.3.3 Self-Training

Existing HOI compositional learning approaches [27, 29, 28] usually only consider the known HOI concepts and simply discard the composite HOIs out of label space during optimization. Therefore, there are only positive data for object affordance learning, leaving a large number of unlabeled composite HOIs ignored. Considering that the concept confidence on HOI concept discovery also demonstrates the confidence of affordances (verbs) that can be applied to an object category, we thus try to explore the potential of all composite HOIs, i.e., both labeled and unlabeled composite HOIs, in a semi-supervised way. Inspired by the way used in PU learning [14] and pseudo-label learning [37], we devise a self-training strategy by assigning the pseudo labels to each verb-object combination instance using the concept confidence matrix 𝐌\mathbf{M}, and optimize the network with the pseudo labels in an end-to-end way. With the self-training, the online concept discovery can gradually improve the concept confidence 𝐌\mathbf{M}, and in turn optimize the HOI model for object affordance learning with the concept confidence. Specifically, we construct the pseudo labels 𝐘~vRNN×Nv\mathbf{\tilde{Y}}_{v}\in R^{NN\times N_{v}} from the concept confidence matrix 𝐌RNv×No\mathbf{M}\in R^{N_{v}\times N_{o}} for composite HOIs 𝐱h\mathbf{x}_{h} as follows,

𝐘~v(i,:)=jNo𝐌(:,j)max(𝐌)𝐘h(i,:,j),\mathbf{\tilde{Y}}_{v}(i,:)=\sum_{j}^{N_{o}}\frac{\mathbf{M}(:,j)}{\max(\mathbf{M})}\odot\mathbf{Y}_{h}(i,:,j), (4)

where 0j<No0\leq j<N_{o} indicates the index of object category, 0i<NN0\leq i<NN is the index of HOI representations. Here, NN is the number of HOIs in each mini-batch, and is usually very small on HICO-DET and V-COCO. Thus the time complexity of Equation 4 is small. The labels of composite HOIs are reshaped as 𝐘hRNN×Nv×No\mathbf{Y}_{h}\in R^{NN\times N_{v}\times N_{o}}. Noticeably, in each label 𝐘h(i,:,:)\mathbf{Y}_{h}(i,:,:), there is only one vector 𝐘h(i,:,j)\mathbf{Y}_{h}(i,:,j) larger than 0 because each HOI has only one object. As a result, we obtain pseudo verb label 𝐘~v(i,:)\mathbf{\tilde{Y}}_{v}(i,:) for HOI 𝐱hi\mathbf{x}_{h_{i}}. Finally, we use composite HOIs with pseudo labels to train the models, and the loss function is defined as follows,

d=1NNiNN(1NvkNvBCE(𝐙(i,k)T,𝐘~v(i,k))),\mathcal{L}_{d}=\frac{1}{NN}\sum_{i}^{NN}(\frac{1}{N_{v}}\sum_{k}^{N_{v}}\mathcal{L}_{\text{BCE}}(\frac{\mathbf{Z}(i,k)}{T},\mathbf{\tilde{Y}}_{v}(i,k))), (5)

where 𝐙(i,:)\mathbf{Z}(i,:) is the prediction of the ii-th composite HOI, 0k<Nv0\leq k<N_{v} means the index of predictions, TT is the temperature hyper-parameter to smooth the predictions (the default value is 1 in experiment), BCE\mathcal{L}_{\text{BCE}} indicates the binary cross entropy loss. Finally, we optimize the network using d\mathcal{L}_{d}, h\mathcal{L}_{h} and c\mathcal{L}_{c} in an end-to-end way, where h\mathcal{L}_{h} indicate the typical classification loss for known HOIs and c\mathcal{L}_{c} is the compositional learning loss [27].

4 Experiments

In this section, we first introduce the datasets and evaluation metrics. We then compare the baseline and the proposed method for HOI concept discovery and object affordance recognition. We also demonstrate the effectiveness of the proposed method for HOI detection with unknown concepts and zero-shot HOI detection. Lastly, we provide some visualizations results of self-compositional learning. Moreover, ablation studies and the full results of HOI detection with self-compositional learning are provided in Appendix D, F, respectively.

4.1 Datasets and Evaluation Metrics

Datasets. We extend two popular HOI detection datasets, HICO-DET [7] and V-COCO [24], to evaluate the performance of different methods for HOI concept discovery. Specifically, we first manually annotate all the possible verb-object combinations on HICO-DET (117 verbs and 80 objects) and V-COCO (24 verbs and 80 objects). As a result, we obtain 1,681 concepts on HICO-DET and 401 concepts on V-COCO, i.e. , 1,681 of 9,360 verb-object combinations on HICO-DET and 401 of 1,920 verb-object combinations on V-COCO are reasonable. Besides, 600 of 1,681 HOI concepts on HICO-DET and 222 of 401 HOI concepts on V-COCO are known according to existing annotations. Thus, the HOI concept discovery task requires to discover the other 1,081 concepts on HICO-DET and 179 concepts on V-COCO. See more details about the annotation process, the statistics of annotations, and the novel HOI concepts in Appendix B.

Evaluation Metrics. HOI concept discovery aims to discover all reasonable combinations between verbs and objects according to existing HOI training samples. We report the performance by using the average precision (AP) for concept discovery and mean AP (or mAP) for object affordance recognition. For HOI detection, we also report the performance using mAP. We follow [28] to evaluate object affordance recognition with HOI model on COCO validation 2017 [41], Object 365 validation [52], HICO-DET test set [7] and Novel Objects from Object 365 [52].

4.2 Implementation Details

We implement the proposed method with TensorFlow [1]. During training, we have two HOI images (randomly selected) in each mini-batch and we follow  [19] to augment ground truth boxes via random crop and random shift. We use a modified HOI compositional learning framework, i.e. , we directly predict the verb classes and optimize the composite HOIs using SCL. Following  [27, 29], the overall loss function is defined as =λ1h+λ2c+λ3d\mathcal{L}=\lambda_{1}\mathcal{L}_{h}+\lambda_{2}\mathcal{L}_{c}+\lambda_{3}\mathcal{L}_{d}, where λ1=2\lambda_{1}=2, λ2=0.5\lambda_{2}=0.5, λ3=0.5\lambda_{3}=0.5 on HICO-DET, and λ1=0.5\lambda_{1}=0.5, λ2=0.5\lambda_{2}=0.5, λ3=0.5\lambda_{3}=0.5 on V-COCO, respectively. Following [29], we also include a sigmoid loss for verb representation and the loss weight is 0.30.3 on HICO-DET. For self-training, we remove the composite HOIs when its corresponding concept confidence is 0, i.e. , the concept confidence has not been updated. If not stated, the backbone is ResNet-101. The Classifier is a two-layer MLP. We train the model for 3.0M iterations on HICO-DET and 300K iterations on HOI-COCO with an initial learning rate of 0.01. For zero-shot HOI detection, we keep human and objects with the score larger than 0.3 and 0.1 on HICO-DET, respectively. See more ablation studies (e.g., hyper-parameters, modules) in Appendix D. Experiments are conducted using a single Tesla V100 GPU (16GB), except for experiments on Qpic [55], which uses four V100 GPUs with PyTorch [46].

4.3 HOI Concept Discovery

Baseline and Methods. We perform experiments to evaluate the effectiveness of our proposed method for HOI concept discovery. For a fair comparison, we build several baselines and methods as follows,

  • Random: we randomly generate the concept confidence to evaluate the performance.

  • Affordance: discover concepts via affordance prediction [28] as described in Sec 3.3.1.

  • GAT [56]: build a graph attention network to mine the relationship among verbs during HOI detection, and discover concepts via affordance prediction.

  • Qpic[55]: convert verb and object predictions of  [55] to concept confidence similar as online discovery.

  • Qpic* [55] +SCL: utilize concept confidence to update verb labels, and optimize the network (Self-Training). Here, we have no composite HOIs.

Please refer to the Appendix for more details, comparisons (e.g., re-training, language embedding), and qualitative discovered concepts with analysis.

Table 1: The performance of the proposed method for HOI concept discovery. We report all performance using the average precision (AP) (%). SCL means self-compositional learning. SCL- means online concept discovery without self-training.
Method HICO-DET V-COCO
Unknown (%) Known (%) Unknown (%) Known (%)
Random 12.52 6.56 12.53 13.54
Affordance [28] 24.38 57.92 20.91 95.71
GAT [56] 26.35 76.05 18.35 98.09
Qpic* [55] 27.53 87.68 15.03 13.21
SCL- 22.25 83.04 24.89 96.70
Qpic* [55] + SCL 28.44 88.91 15.48 13.34
SCL 33.58 92.65 28.77 98.95

Results Comparison. Table 1 shows affordance prediction is capable of HOI concept discovery since affordance transfer learning [28] also transfers affordances to novel objects. Affordance prediction achieves 24.38% mAP on HICO-DET and 21.36% mAP on V-COCO, respectively, significantly better than the random baseline. With graph attention network, the performance is further improved a bit. Noticeably, [28] completely ignores the possibility of HOI concept discovery via affordance prediction. Due to the strong ability of verb and object prediction, Qpic achieves 27.42% on HICO-DET, better than affordance prediction. However, Qpic has poor performance on V-COCO. The inference process of affordance prediction for concept discovery is time-consuming (over 8 hours with one GPU). Thus we devise an efficient online concept discovery method which directly predicts all concept confidences. Specifically, the online concept discovery method (SCL-) achieves 22.25% mAP on HICO-DET, which is slightly worse than the result of affordance prediction. On V-COCO, the online concept discovery method improves the performance of concept discovery by 3.98% compared to the affordance prediction. The main reason for the above observation might be due to that V-COCO is a small dataset and the HOI model can easily overfit known concepts on V-COCO. Particularly, SCL significantly improves the performance of HOI concept discovery from 22.36% to 33.58% on HICO-DET and from 24.89% to 28.77% on V-COCO, respectively. We find we can also utilize self-training to improve concept discovery on Qpic [55] (ResNet-50) though the improvement is limited, which might be because verbs and objects are entangled with Qpic. Lastly, we meanwhile find SCL largely improves concept discovery of known concepts on both HICO-DET and V-COCO.

Table 2: Comparison of object affordance recognition with HOI network (trained on HICO-DET) among different datasets. Val2017 is the validation 2017 of COCO [41]. Obj365 is the validation of Object365 [52] with only COCO labels. Novel classes are selected from Object365 with non-COCO labels. ATL\ast means ATL optimized with COCO data. Numbers are copied from the appendix in  [28]. Unknown affordances indicate we evaluate with our annotated affordances. Previous approaches [27, 28] are usually trained by less 0.8M iterations (Please refer to the released checkpoint in [27, 28]). We thus also illustrate SCL under 0.8M iterations by default. SCL- means SCL without self-training. Results are reported by Mean Average Precision (%).
Method Known Affordances Unknown Affordances
Val2017 Obj365 HICO Novel Val2017 Obj365 HICO Novel
FCL [29] 25.11 25.21 37.32 6.80 - - - -
VCL [27] 36.74 35.73 43.15 12.05 28.71 27.58 32.76 12.05
ATL [28] 52.01 50.94 59.44 15.64 36.80 34.38 42.00 15.64
ATL\ast [28] 56.05 40.83 57.41 8.52 37.01 30.21 43.29 8.52
SCL- 50.51 43.52 57.29 14.46 44.21 41.37 48.68 14.46
SCL 59.64 52.70 67.05 14.90 47.68 42.05 52.95 14.90
SCL (3M iters) 72.08 57.53 82.47 18.55 56.19 46.32 64.50 18.55

4.4 Object Affordance Recognition

Following [28] that has discussed average precision (AP) is more robust for evaluating object affordance, we evaluate object affordance recognition with AP on HICO-DET. Table 2 illustrates SCL largely improves SCL- (without self-training) by over 9% on Val2017, Object365, HICO-DET under the same training iterations. SCL requires more iterations to converge, and SCL greatly improves previous methods on all datasets with 3M iterations (Please refer to Appendix D.2 for convergence analysis). Noticeably, SCL directly predicts verb rather than HOI categories, and removes the spatial branch. Thus, SCL without self-training (SCL-) is a bit worse than ATL. Previous approaches ignore the unknown affordance recognition. We use the released models of  [28] to evaluate the results on novel affordance recognition. Here, affordances of novel classes (annotated by hand [28]) are the same in the two settings. We find SCL improves the performance considerably by over 10% on Val2017 and HICO-DET.

4.5 HOI Detection with Unknown Concepts

HOI concept discovery enables zero-shot HOI detection with unknown concepts by first discovering unknown concepts and then performing HOI detection. The experimental results of HOI detection with unknown concepts are shown in Table 3. We follow [27] to evaluate HOI detection with 120 unknown concepts in two settings: rare first selection and non-rare first selection, i.e. , we select 120 unknown concepts from head and tail classes respectively. Different from [27, 29] where the existence of unseen categories is known and the HOI samples for unseen categories are composed during optimization, HOI detection with unknown concepts does not know the existence of unseen categories. Therefore, we select top-KK concepts according to the confidence score during inference to evaluate the performance of HOI detection with unknown concepts (that is also zero-shot) in the default mode [7].

As shown in Table 3, with more selected unknown concepts according to concept confidence, the proposed approach further improves the performance on unseen categories on both rare first and non-rare first settings. Specifically, it demonstrates a large difference between rare first unknown concepts HOI detection and non-rare first unknown concepts HOI detection in Table 3. Considering that the factors (verbs and objects) of rare-first unknown concepts are rare in the training set [29], the recall is very low and thus degrades the performance on unknown categories. However, with concept discovery, the results with top 120 concepts on unknown categories are improved by relatively 34.52% (absolutely 0.58%) on rare first unknown concepts setting and by relatively 20.31% (absolutely 1.19%) on non-rare first setting, respectively. with more concepts, the performance on unknown categories is also increasingly improved.

Table 3: Illustration of HOI detection with unknown concepts and zero-shot HOI detection with SCL. KK is the number of selected unknown concepts. HOI detection results are reported by mean average precision (mAP)(%). We also report the recall rate of the unseen categories in the top-KK novel concepts. “KK = all” indicates the results of selecting all concepts, i.e. , common zero-shot. \ast means we train Qpic [55](ResNet-50) with the released code in zero-shot setting and use the discovered concepts of SCL to evaluate HOI detection with unknown concepts. Un indicates Unknown/Unseen, Kn indicates Known/Seen, while Rec indicates Recall.
Method KK Rare First Non-rare First
Un Kn Full Rec (%) Un Kn Full Rec (%)
SCL 0 1.68 22.72 18.52 0.00 5.86 16.70 14.53 0.00
SCL 120 2.26 22.72 18.71 10.83 7.05 16.70 14.77 21.67
SCL 240 3.66 22.72 18.91 15.00 7.17 16.70 14.80 25.00
SCL 360 4.09 22.72 19.00 15.83 7.91 16.70 14.94 30.83
SCL all 9.64 22.72 19.78 100.00 13.30 16.70 16.02 100.00
Qpic\ast [55] 0 0.0 30.47 24.37 0.00 0.0 23.73 18.98 0.0
Qpic\ast [55] 120 2.32 30.47 24.84 10.83 14.90 22.19 20.58 21.67
Qpic\ast [55] 240 3.35 30.47 25.04 15.00 14.90 22.79 21.22 25.00
Qpic\ast [55] 360 3.72 30.47 25.12 15.83 14.91 23.13 21.48 30.83
Qpic\ast [55] all 15.24 30.44 27.40 100.00 21.03 23.73 23.19 100.00
ATL [28] all 9.18 24.67 21.57 100.00 18.25 18.78 18.67 100.00
FCL [29] all 13.16 24.23 22.01 100.00 18.66 19.55 19.37 100.00
Qpic + SCL all 19.07 30.39 28.08 100.00 21.73 25.00 24.34 100.00

We also utilize the discovered concept confidences with SCL to evaluate HOI detection with unknown concepts on Qpic [55]. For a fair comparison, we use the same concept confidences to SCL. Without concept discovery, the performance of Qpic [55] degrades to 0 on Unseen categories though Qpic significantly improves zero-shot HOI detection. Lastly, we show zero-shot HOI detection (the unseen categories are known) in Table 3 (Those rows where KK is all). We find that SCL significantly improves Qpic, and forms a new state-of-the-art on zero-shot setting though we merely use ResNet-50 as backbone in Qpic. We consider SCL improves the detection of rare classes (include unseen categories in rare first and seen categories in non-rare first) via stating the distribution of verb and object. See Appendix F for more analysis, e.g., SCL improves Qpic particularly for rare categories on Full HICO-DET.

4.6 Visualization

Figure 3 illustrates the Grad-CAM under different methods. We find the proposed SCL focus on the details of objects and small objects, while the baseline and VCL mainly highlight the region of human and the interaction region, e.g., SCL highlights the details of the motorbike, particularly the front-wheel (last row). Besides, SCL also helps the model via emphasizing the learning of small objects (e.g., frisbee and bottle in the last two columns), while previous works ignore the small objects. This demonstrates SCL facilitates affordance recognition and HOI concept discovery via exploring more details of objects. A similar trend can be observed in Appedix G (Qpic+SCL).

Refer to caption
Figure 3: A visual comparison of recent methods using the Grad-CAM [51] tool. The first row is input image, the second row is baseline without compositional approach, the third row is VCL [27] and the last row is the proposed SCL. We do not compare with ATL [28], since that ATL uses extra training datasets. Here, we compare all models using the same dataset.

5 Conclusion

We propose a novel task, Human-Object Interaction Concept Discovery, which aims to discover all reasonable combinations (i.e. , HOI concepts) between verbs and objects according to a few training samples of known HOI concepts/categories. Furthermore, we introduce a self-compositional learning or SCL framework for HOI concept discovery. SCL maintains an online updated concept confidence matrix, and assigns pseudo labels according to the matrix for all composite HOI features, and thus optimize both known and unknown composite HOI features via self-training. SCL facilitates affordance recognition of HOI model and HOI concept discovery via enabling the learning on both known and unknown HOI concepts. Extensive experiments demonstrate SCL improves HOI concept discovery on HICO-DET and V-COCO and object affordance recognition with HOI model, enables HOI detection with unknown concepts, and improves zero-shot HOI detection.
Acknowledgments Mr. Zhi Hou and Dr. Baosheng Yu are supported by ARC FL-170100117, DP-180103424, IC-190100031, and LE-200100049.

References

  • [1] Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., Isard, M., et al.: Tensorflow: A system for large-scale machine learning. In: 12th symposium on operating systems design and implementation (OSDI). pp. 265–283 (2016)
  • [2] Bansal, A., Rambhatla, S.S., Shrivastava, A., Chellappa, R.: Detecting human-object interactions via functional generalization. In: AAAI (2020)
  • [3] Best, J.B.: Cognitive psychology. West Publishing Co (1986)
  • [4] Byrd, J., Lipton, Z.: What is the effect of importance weighting in deep learning? In: ICML. pp. 872–881. PMLR (2019)
  • [5] Carreira, J., Zisserman, A.: Quo vadis, action recognition? a new model and the kinetics dataset. In: CVPR. pp. 6299–6308 (2017)
  • [6] Chao, Y.W., Liu, Y., Liu, X., Zeng, H., Deng, J.: Learning to detect human-object interactions. In: 2018 ieee winter conference on applications of computer vision (wacv). pp. 381–389. IEEE (2018)
  • [7] Chao, Y.W., Liu, Y., Liu, X., Zeng, H., Deng, J.: Learning to detect human-object interactions. In: WACV. pp. 381–389. IEEE (2018)
  • [8] Chao, Y.W., Wang, Z., He, Y., Wang, J., Deng, J.: Hico: A benchmark for recognizing human-object interactions in images. In: ICCV. pp. 1017–1025 (2015)
  • [9] Chen, M., Liao, Y., Liu, S., Chen, Z., Wang, F., Qian, C.: Reformulating hoi detection as adaptive set prediction. In: CVPR. pp. 9004–9013 (2021)
  • [10] Coren, S.: Sensation and perception. Handbook of psychology pp. 85–108 (2003)
  • [11] Cui, Y., Jia, M., Lin, T.Y., Song, Y., Belongie, S.: Class-balanced loss based on effective number of samples. In: CVPR. pp. 9268–9277 (2019)
  • [12] Dabral, R., Shimada, S., Jain, A., Theobalt, C., Golyanik, V.: Gravity-aware monocular 3d human-object reconstruction. In: ICCV. pp. 12365–12374 (2021)
  • [13] Damen, D., Doughty, H., Maria Farinella, G., Fidler, S., Furnari, A., Kazakos, E., Moltisanti, D., Munro, J., Perrett, T., Price, W., et al.: Scaling egocentric vision: The epic-kitchens dataset. In: ECCV. pp. 720–736 (2018)
  • [14] De Comité, F., Denis, F., Gilleron, R., Letouzey, F.: Positive and unlabeled examples help learning. In: International Conference on Algorithmic Learning Theory. pp. 219–230. Springer (1999)
  • [15] Deng, S., Xu, X., Wu, C., Chen, K., Jia, K.: 3d affordancenet: A benchmark for visual object affordance understanding. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1778–1787 (2021)
  • [16] Elkan, C., Noto, K.: Learning classifiers from only positive and unlabeled data. In: Proceedings of the 14th ACM SIGKDD international conference on Knowledge Discovery and Data Mining. pp. 213–220 (2008)
  • [17] Fang, K., Wu, T.L., Yang, D., Savarese, S., Lim, J.J.: Demo2vec: Reasoning object affordances from online videos. In: CVPR (2018)
  • [18] Fouhey, D.F., Delaitre, V., Gupta, A., Efros, A.A., Laptev, I., Sivic, J.: People watching: Human actions as a cue for single view geometry. IJCV 110, 259–274 (2014)
  • [19] Gao, C., Zou, Y., Huang, J.B.: ican: Instance-centric attention network for human-object interaction detection. BMVC (2018)
  • [20] Gibson, J.J.: The ecological approach to visual perception (1979)
  • [21] Gibson, J.J.: The ecological approach to visual perception: classic edition. Psychology Press (2014)
  • [22] Gilmer, J., Schoenholz, S.S., Riley, P.F., Vinyals, O., Dahl, G.E.: Neural message passing for quantum chemistry. In: ICML. pp. 1263–1272. PMLR (2017)
  • [23] Gupta, A., Kembhavi, A., Davis, L.S.: Observing human-object interactions: Using spatial and functional compatibility for recognition. IEEE PAMI 31(10), 1775–1789 (2009)
  • [24] Gupta, S., Malik, J.: Visual semantic role labeling. arXiv preprint arXiv:1505.04474 (2015)
  • [25] Hassan, M., Dharmaratne, A.: Attribute based affordance detection from human-object interaction images. In: Image and Video Technology. pp. 220–232. Springer (2015)
  • [26] Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)
  • [27] Hou, Z., Peng, X., Qiao, Y., Tao, D.: Visual compositional learning for human-object interaction detection. In: ECCV (2020)
  • [28] Hou, Z., Yu, B., Qiao, Y., Peng, X., Tao, D.: Affordance transfer learning for human-object interaction detection. In: CVPR (2021)
  • [29] Hou, Z., Yu, B., Qiao, Y., Peng, X., Tao, D.: Detecting human-object interaction via fabricated compositional learning. In: CVPR (2021)
  • [30] Huynh, D., Elhamifar, E.: Interaction compass: Multi-label zero-shot learning of human-object interactions via spatial relations. In: ICCV. pp. 8472–8483 (2021)
  • [31] Ioffe, S., Szegedy, C.: Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: ICML. pp. 448–456. PMLR (2015)
  • [32] Ji, J., Desai, R., Niebles, J.C.: Detecting human-object relationships in videos. In: ICCV. pp. 8106–8116 (2021)
  • [33] Kato, K., Li, Y., Gupta, A.: Compositional learning for human object interaction. In: ECCV. pp. 234–251 (2018)
  • [34] Kim, B., Lee, J., Kang, J., Kim, E.S., Kim, H.J.: Hotr: End-to-end human-object interaction detection with transformers. In: CVPR. pp. 74–83 (2021)
  • [35] Kingma, D.P., Mohamed, S., Rezende, D.J., Welling, M.: Semi-supervised learning with deep generative models. In: NIPS (2014)
  • [36] Kjellström, H., Romero, J., Kragić, D.: Visual object-action recognition: Inferring object affordances from human demonstration. Computer Vision and Image Understanding 115(1), 81–90 (2011)
  • [37] Lee, D.H., et al.: Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In: Workshop on challenges in representation learning, ICML (2013)
  • [38] Li, Y.L., Liu, X., Wu, X., Li, Y., Lu, C.: Hoi analysis: Integrating and decomposing human-object interaction. NeuIPS 33 (2020)
  • [39] Li, Y.L., Zhou, S., Huang, X., Xu, L., Ma, Z., Fang, H.S., Wang, Y.F., Lu, C.: Transferable interactiveness prior for human-object interaction detection. In: CVPR (2019)
  • [40] Liao, Y., Liu, S., Wang, F., Chen, Y., Feng, J.: Ppdm: Parallel point detection and matching for real-time human-object interaction detection. In: CVPR (2020)
  • [41] Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: ECCV. pp. 740–755. Springer (2014)
  • [42] Materzynska, J., Xiao, T., Herzig, R., Xu, H., Wang, X., Darrell, T.: Something-else: Compositional action recognition with spatial-temporal interaction networks. In: CVPR. pp. 1049–1059 (2020)
  • [43] Nagarajan, T., Grauman, K.: Learning affordance landscapes for interaction exploration in 3d environments. Advances in Neural Information Processing Systems 33, 2005–2015 (2020)
  • [44] Nawhal, M., Zhai, M., Lehrmann, A., Sigal, L., Mori, G.: Generating videos of zero-shot compositions of actions and objects. In: ECCV (2020)
  • [45] Norman, D.A.: The Design of Everyday Things. Basic Books, Inc., USA (2002)
  • [46] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., Chintala, S.: Pytorch: An imperative style, high-performance deep learning library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., Garnett, R. (eds.) NeurIPS, pp. 8024–8035. Curran Associates, Inc. (2019), http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
  • [47] Pennington, J., Socher, R., Manning, C.D.: Glove: Global vectors for word representation. In: EMNLP. pp. 1532–1543 (2014)
  • [48] Peyre, J., Laptev, I., Schmid, C., Sivic, J.: Detecting unseen visual relations using analogies. In: ICCV (October 2019)
  • [49] Scott, C., Blanchard, G.: Novelty detection: Unlabeled data definitely help. In: Artificial intelligence and statistics. pp. 464–471. PMLR (2009)
  • [50] Scudder, H.: Probability of error of some adaptive pattern-recognition machines. IEEE Transactions on Information Theory 11(3), 363–371 (1965)
  • [51] Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: Visual explanations from deep networks via gradient-based localization. In: ICCV (2017)
  • [52] Shao, S., Li, Z., Zhang, T., Peng, C., Sun, J.: Objects365: A large-scale, high-quality dataset for object detection. In: ICCV (2019)
  • [53] Shen, L., Yeung, S., Hoffman, J., Mori, G., Fei-Fei, L.: Scaling human-object interaction recognition through zero-shot learning. In: WACV. pp. 1568–1576. IEEE (2018)
  • [54] Springenberg, J.T.: Unsupervised and semi-supervised learning with categorical generative adversarial networks. arXiv preprint arXiv:1511.06390 (2015)
  • [55] Tamura, M., Ohashi, H., Yoshinaga, T.: QPIC: Query-based pairwise human-object interaction detection with image-wide contextual information. In: CVPR (2021)
  • [56] Veličković, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. In: ICLR (2017)
  • [57] Wang, D., Cui, P., Zhu, W.: Structural deep network embedding. In: Proceedings of the 22nd ACM SIGKDD international conference on Knowledge Discovery and Data Mining. pp. 1225–1234 (2016)
  • [58] Wang, S., Yap, K.H., Ding, H., Wu, J., Yuan, J., Tan, Y.P.: Discovering human interactions with large-vocabulary objects via query and multi-scale detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 13475–13484 (2021)
  • [59] Wang, S., Yap, K.H., Yuan, J., Tan, Y.P.: Discovering human interactions with novel objects via zero-shot learning. In: CVPR. pp. 11652–11661 (2020)
  • [60] Xie, Q., Luong, M.T., Hovy, E., Le, Q.V.: Self-training with noisy student improves imagenet classification. In: CVPR. pp. 10687–10698 (2020)
  • [61] Xu, D., Zhu, Y., Choy, C.B., Fei-Fei, L.: Scene graph generation by iterative message passing. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 5410–5419 (2017)
  • [62] Yang, X., Song, Z., King, I., Xu, Z.: A survey on deep semi-supervised learning. arXiv preprint arXiv:2103.00550 (2021)
  • [63] Yao, B., Ma, J., Li, F.F.: Discovering object functionality. In: ICCV (2013)
  • [64] Zhai, W., Luo, H., Zhang, J., Cao, Y., Tao, D.: One-shot object affordance detection in the wild. arXiv preprint arXiv:2108.03658 (2021)
  • [65] Zhang, A., Liao, Y., Liu, S., Lu, M., Wang, Y., Gao, C., Li, X.: Mining the benefits of two-stage and one-stage hoi detection. In: Advances in Neural Information Processing Systems. vol. 34 (2021)
  • [66] Zhang, J.Y., Pepose, S., Joo, H., Ramanan, D., Malik, J., Kanazawa, A.: Perceiving 3d human-object spatial arrangements from a single image in the wild. In: ECCV (2020)
  • [67] Zheng, S., Chen, S., Jin, Q.: Skeleton-based interactive graph network for human object interaction detection. In: 2020 IEEE International Conference on Multimedia and Expo (ICME). pp. 1–6. IEEE (2020)
  • [68] Zhong, X., Ding, C., Qu, X., Tao, D.: Polysemy deciphering network for human-object interaction detection. In: European Conference on Computer Vision. pp. 69–85. Springer (2020)
  • [69] Zhong, X., Qu, X., Ding, C., Tao, D.: Glance and gaze: Inferring action-aware points for one-stage human-object interaction detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 13234–13243 (2021)
  • [70] Zhu, Y., Fathi, A., Fei-Fei, L.: Reasoning about object affordances in a knowledge base representation. In: ECCV. pp. 408–424. Springer (2014)
  • [71] Zou, C., Wang, B., Hu, Y., Liu, J., Wu, Q., Zhao, Y., Li, B., Zhang, C., Zhang, C., Wei, Y., et al.: End-to-end human object interaction detection with hoi transformer. In: CVPR. pp. 11825–11834 (2021)

Appendix 0.A Detailed Analysis for the Motivation

Actually, after we generate the composite HOI features, we have features for both known and unknown concepts. We merely know the HOI features of the known concepts are existing, while we do not know whether the HOI features of unknown concepts are reasonable or not. This actually fall into a typical semi-supervised learning, in which part of samples are labeled (known). Therefore, inspired by the popular semi-supervised learning method, we propose to design a self-training strategy with pseudo labels.

SCL largely improves concept discovery. At first, during training, SCL involves both HOI instances from known or unknown concepts (via pseudo-labeling). Another important thing is that SCL uses both positive and negative unknown concepts, which prevents the model from only fitting the verb patterns. For example, the classifier may predict a reasonable concept for the verb “eat” regardless of the object representation, if there are no negative unknown concepts, e.g., “eat TV”. Lastly, as shown in Figure 4, SCL also reduces the risk of overfitting known concepts compared with ATL. e.g., we observe high confidence for the novel concept ”squeeze banana”(sort in 2027) in SCL, while the confidence of ”squeeze banana” is merely 0.0017 (sort in 7554) in ATL.

Appendix 0.B Annotation

In order to evaluate the proposed method, we manually annotate the novel concepts for both HICO and V-COCO dataset. Specifically, we annotate the concepts that people can infer from existing concepts. The final set of concepts are provided in the supplemental material.

Statistically, there are about 1.3% and 1.9% mislabeled pairs on HICO-DET and V-COCO, respectively. Meanwhile, there are about 1.7% and 1.1% unlabeled pairs (including ambiguous verbs) on the remaining categories of HICO-DET and V-COCO.

To evaluate the effect of annotation quality of concept annotation on HOI concept discovery, we illustrate the result of different models with different annotations. We compare two versions of annotations, both of which are provided in supplemental materials. Specifically, the file “label_hoi_concept.csv” is the worse version, while “label_hoi_concept_new.csv” is the refined version. Table 4 shows SCL even achieves better performance when evaluate SCL with better annotation, while the performance of baseline is not improved. This experiments together with Table 1 in the main paper show the quality of current annotation is enough for the evaluation of the proposed method.

Table 4: The performance of the proposed method for HOI concept discovery under different annotations. Better Annotation indicates we remove some wrongly labeled concepts in annotation. We report all performance using the average precision (AP) (%). UC means unknown concepts and KC means known concepts. SCL means self-compositional learning. SCL- means online concept discovery without self-training.
Method Better Annotation UC KC
SCL- 22.36 83.04
SCL 33.26 93.06
SCL- 22.25 83.04
SCL 33.58 92.65

Appendix 0.C Qualitative illustration

We also illustrate the discover concepts in this Section. Here, we choose the concepts after removing the known concepts from the prediction list because the confidence of known concepts in the prediction of SCL is usually very higher. We choose 5 concepts with high confidence and 5 concepts with low confidence to illustrate. Table 5 shows the discovered concepts in SCL are usually more reasonable. We provide the full prediction list with confidence in “result_conf_SCL.txt” in supplementary materials.

Table 5: The illustration of discovered concepts.
Method Concepts with high confidence Concepts with low confidence
SCL- type_on sink,inspect refrigerator,feed suitcase, inspect chair,carry stop_sign zip zebra, sign dog, chase broccoli, set parking_meter, tag teddy_bear
SCL ride bear, board truck, carry bowl, wash fire_hydrant, hop_on motorcycle zip zebra, flush parking_meter, stop_at hair_drier, stop_at microwave

Appendix 0.D Ablation Studies

0.D.1 Modules

We conduct ablation studies on three modules: verb auxiliary loss [29], union verb [27], and spatial branch [19]. Union verb indicates that we extract verb representation from the union box of human and object. When we remove the union verb representation, we directly extract verb representation from the human bounding box; In our experiment, we remove the spatial branch. Here, we demonstrate we achieve better performance without the spatial branch.

Spatial branch. We remove the spatial branch in [19], which is very effective for HOI detection. We find that the spatial branch degrades the performance of HOI concept discovery: the performance of HOI concept discovery increases from 32.56% to 33.26% without spatial branch, as shown in Table 6. We thus remove spatial branch.

Verb auxiliary loss. We follow [29] to utilize a verb auxiliary loss to regularize verb representations. As shown in Table 6, the model without using a verb auxiliary loss drops by nearly 3% on unseen concepts, which demonstrates the importance of verb auxiliary loss for HOI concept discovery.

Union verb. Table 6 demonstrates that extracting verb representation from union box is of great importance for HOI concept discovery. When we extract verb representation from human bounding box, the result of HOI concept discovery apparently drops from 32.56% to 28.30%.

Though verb auxiliary loss and union verb representation are very helpful for concept discovery, the performance without the two strategies still outperform our baseline, i.e. , online concept discovery without self-training.

Table 6: Ablation studies of different modules on HICO-DET. UC means unknown concepts and KC means known concepts. Verb aux loss means Verb auxiliary loss (i.e. , binary cross entropy loss). Results are reported by average precision (%).
Spatial branch Verb aux loss Union Verb UC KC
32.56 94.39
- 33.26 93.06
- 29.56 93.36
- 28.30 94.27

0.D.2 Convergence Analysis

Refer to caption
Figure 4: Illustration of the convergence with self-training strategy.

To some extent, the self-training approach makes use of all composite HOIs, and thus significantly enriches the training data. As a result, the self-training strategy usually requires more iterations to converge to a better result. Figure 4 illustrates the comparison of convergence between online concept discovery and self-training. For online concept discovery, we observe that the model begins to overfit the known concepts after 2,000,000 iterations, and we thus have an early stop during the optimization. We notice that the result on unknown concepts of self-training increases to 32.%, while the baseline (i.e. , online concept discovery) begins to overfit after 800,000 iterations. This might be because the self-training utilizes all composite HOIs including many impossible combinations (i.e. , negative samples for HOI concept discovery).

Table 7: Ablation studies of hyper-parameters on V-COCO. UC means unknown concepts and KC means known concepts. Results are reported by average precision (%).
λ3\lambda_{3} 0.5 0.5 0.5 0.25 1. 2. 4.
TT 1 2 0.5 1. 1. 1. 1.
UC (%) 29.52 28.60 29.69 28.06 29.94 31.33 29.78
KC (%) 97.57 96.76 97.57 95.32 97.87 97.81 97.94

0.D.3 Hyper-parameters

In the main paper, we have several hyper-parameters (i.e. λ1\lambda_{1}, λ2\lambda_{2}, λ3\lambda_{3}, TT, where λ1=2.\lambda_{1}=2., λ2=0.5\lambda_{2}=0.5, λ3=0.5\lambda_{3}=0.5 and T=1.T=1.). For λ1\lambda_{1} and λ2\lambda_{2}, we follow the settings in  [27]. For λ3\lambda_{3} and TT, we perform ablation studies on V-COCO as shown in Table 7. We notice that both TT and λ3\lambda_{3} have an important effect on the HOI concept discovery. As shown in Table 7, the performance increases from 29.52% to 31.33% on unseen concepts when we set λ3=2.\lambda_{3}=2., which is much better than the results reported in the main paper. This also illustrates that d\mathcal{L}_{d} is more important than CL\mathcal{L}_{CL} for HOI concept discovery.

In our experiment, we apply the temperature TT to predictions. As shown in Table 7, we find that when TT decreases to 0.5, the performance also slightly increases from 29.52% to 29.69%. Thus, we further conduct ablation experiments on TT in Table 8. Specifically, to quickly evaluate the effect of TT, we remove spatial branch and run all experiments with 1,000,000 iterations. Noticeably, when we set T=0.25T=0.25, the performance on concept discovery further increases from 30.36% to 33.66%, which indicates a smaller temperature helps HOI concept discovery. In our experiments, we also find this result further increases to over 35.% when T=0.5T=0.5 after convergence, which is much better than the result (33.26%) of T=1T=1. This might be because smaller temperature is less sensitive to noise data, since composite HOIs can be regard as noise data.

Table 8: Ablation studies of hyper-parameter TT on HICO-DET. Here, we run all experiments with only 1,000,000 iterations and remove the spatial branch to evaluate TT. UC means unknown concepts and KC means known concepts. Results are reported by average precision (%).
TT 2 1 0.5 0.25 0.125
UC (%) 27.15 30.36 33.54 33.66 33.25
KC (%) 85.53 88.72 91.71 93.62 94.32
Table 9: Illustration of normalized pseudo labels on HICO-DET and V-COCO. Experiments results are reported by average precision (%). Here, the SCL model uses spatial branch.
Method HICO-DET V-COCO
UC (%) KC (%) UC (%) KC (%)
SCL 32.56 94.39 29.52 97.57
w/o normalization 32.30 94.2 29.32 97.93

0.D.4 Normalization for Pseudo-labels

In our experiment, we normalize the confidence matrix for pseudo-labels. Table 9 illustrates the normalization approach has a slight effect on the concept discovery performance.

Appendix 0.E HOI Detection with Unknown Concepts

0.E.1 Additional Comparisons

Table 10 demonstrates SCL consistent improves the baseline (i.e. , SCL without Self-Training). Here, we use the same concepts for a fair comparison. Thus, the recall is the same. Meanwhile, Table 10 also shows Self-Training effectively improves the HOI detection. when we select all concepts to evaluate HOI detection, it is common zero-shot HOI detection, i.e. , all unseen classes are known. Particularly, for application, one can directly detect unknown concepts with concept discovery from the model itself, e.g., Qpic [55]. Here, we mainly demonstrate different methods with the same concept confidence for a fair comparison.

Table 10: Illustration of HOI detection with unknown concepts and zero-shot HOI detection with SCL. KK is the number of selected unknown concepts. HOI detection results are reported by mean average precision (mAP)(%). We also report the recall of the unseen categories in the top-KK novel concepts. KK = all indicates the results of selecting all concepts, i.e. , common zero-shot. \ast means we train Qpic [55](ResNet-50) with the released code in zero-shot setting and use the discovered concepts of SCL to evaluate HOI detection with unknown concepts.
Method KK Rare First Non-rare First
Unknown Known Full Recall (%) Unknown Known Full Recall (%)
Baseline 0 1.68 22.10 18.52 0.00 5.86 16.30 14.21 0.00
Baseline 120 3.06 22.10 18.29 10.83 6.16 16.30 14.27 21.67
Baseline 240 3.28 22.10 18.34 13.33 6.90 16.30 14.42 25.00
Baseline 360 3.86 22.10 18.45 15.83 7.29 16.30 14.50 30.83
Baseline all 9.62 22.10 19.61 100.00 12.82 16.30 15.60 100.00
SCL 0 1.68 22.72 18.52 0.00 5.86 16.70 14.53 0.00
SCL 120 2.26 22.72 18.71 10.83 7.05 16.70 14.77 21.67
SCL 240 3.66 22.72 18.91 15.00 7.17 16.70 14.80 25.00
SCL 360 4.09 22.72 19.00 15.83 7.91 16.70 14.94 30.83
SCL all 9.64 22.72 19.78 100.00 13.30 16.70 16.02 100.00

0.E.2 Novel Objects

In the main paper, we illustrate the result on two compositional zero-shot settings. Here, we further illustrate the effectiveness of HOI concept discovery for novel object HOI detection. Novel object HOI detection requires to detect HOI with novel objects, i.e. , the object of an unseen HOI is never seen in the HOI training set. We follow [28] to select 100 categories as unknown concepts. The remaining categories do not include the objects of unseen categories. Here we use a unique object detector to detect objects. To enable the novel object HOI detection and novel object HOI concept discovery, we follow [28] to incorporate external objects (e.g.COCO [41]) to compose novel object HOI samples. Specifically, we only choose the novel types of objects from COCO [41] as objects images in the framework [28] for novel object HOI detection with unknown concepts.

Table 11 demonstrates concept discovery largely improves the performance on unseen category from 3.92% to 11.41% (relatively by 191%) with top 100 unknown concepts. We meanwhile find the recall increases to 41.00% with only the top 100 unknown concepts. Nevertheless, when we select all unknown concepts, the performance on unseen category is 17.19%. This shows we should improve the performance of concept discovery.

Table 11: Illustration of the effectiveness of HOI concept discovery for HOI detection with unknown concepts (novel objects). KK is the number of selected unknown concepts. HOI detection results are reported by mean average precision (mAP)(%). Recall is evaluated for the unseen categories under the top-kk novel concepts. The last row indicates the results of selecting all concepts.
KK Unseen Seen Full Recall (%)
0 3.92 19.45 16.86 0.00
100 11.41 19.45 18.11 41.00
200 12.40 19.45 18.28 48.00
300 13.52 19.45 18.46 52.00
400 13.52 19.45 18.46 52.00
500 13.91 19.45 18.53 56.00
600 13.91 19.45 18.53 56.00
all 17.19 19.45 19.07 100.00
Table 12: Additional Comparison on HOI concept discovery. We report all performance using the average precision (AP) (%). UC means unknown concepts and KC means known concepts. SCL means self-compositional learning. SCL- means online concept discovery without self-training. SCL (COCO) means we train the network via composing between verbs from HICO and objects from COCO 2014 training set.
Method HICO-DET V-COCO
UC (%) KC (%) UC (%) KC (%)
Random 12.52 6.56 12.53 13.54
language embedding 16.08 29.64 - -
Re-Training 26.09 50.32 - -
SCL- (COCO) 17.01 55.50 26.04 81.47
SCL (COCO) 31.92 86.43 27.90 90.04
SCL- 22.36 83.04 26.64 95.59
SCL 33.26 93.06 29.52 97.57
Table 13: Illustration of the effectiveness of self-training on HOI detection based on ground truth box. Results are reported by mean average precision (%).
Method Full Rare NonRare
SCL 42.92 36.60 44.81
w/o Self-Training 42.66 35.81 44.70
Table 14: Illustration of the effectiveness of self-training for Qpic (ResNet-50). Results are reported by mean average precision (%). * means we use the released code to reproduce the results for a fair comparison. S1 means Scenario 1, while S2 means Scenario 2.
Method HICO-DET V-COCO
Full Rare NonRare S1 S2
GGNet [69] 23.47 16.48 25.60 - 54.7
ATL [28] 23.81 17.43 25.72 - -
HOTR [34] 25.10 17.34 27.42 55.2 64.4
AS-Net[9] 28.87 24.25 30.25 - 53.9
Qpic [55] 29.07 21.85 31.23 58.8 61.0
Qpic* [55] 29.19 23.01 31.04 61.29 62.10
Qpic + SCL 29.75 24.78 31.23 61.55 62.38

Appendix 0.F HOI Detection

One-Stage Method. We also evaluate SCL on Qpic [55], i.e. , the state-of-the-art HOI detection method based on Transformer, for HOI detection. Code is provided in https://github.com/zhihou7/SCL. We first obtain concept confidence similar as Section 3.3.2 in the main paper. Denote 𝐘^vRN×Nv\mathbf{\hat{Y}}_{v}\in R^{N\times N_{v}} as verb predictions, 𝐘^oRN×No\mathbf{\hat{Y}}_{o}\in R^{N\times N_{o}} as verb predictions, we obtain concept predictions 𝐘^h\mathbf{\hat{Y}}_{h} as follows,

𝐘^h=𝐘^v𝐘^o.\mathbf{\hat{Y}}_{h}=\mathbf{\hat{Y}}_{v}\otimes\mathbf{\hat{Y}}_{o}. (6)

Then, we update MM according to Equation 2 and Equation 3 in the main paper. After training, we evaluate HOI concept discovery with MM.

For self-training on Qpic [55], we use MM to update the verb label 𝐘vRN×Nv\mathbf{Y}_{v}\in R^{N\times N_{v}} for annotated HOIs. Here, we do not have composite HOIs because Qpic has entangled verb and object predictions, and we update verb labels with 𝐌\mathbf{M}. Specifically, given an HOI with a verb labeled as yvRvNy_{v}\in R^{N}_{v} and an object labeled as yoRoNy_{o}\in R^{N}_{o}, where 0yo<No0\leq y_{o}<N_{o} denotes the index of object category, we update yvy_{v} as follows,

y~v=max(yv+𝐌(:,yo),1)\widetilde{y}_{v}=max(y_{v}+\mathbf{M}(:,y_{o}),1) (7)

where maxmax means we clip the value to 1 if the value is larger than 1. Then, we obtain pseudo verb label y~v\widetilde{y}_{v} to optimize the samples of the HOI similar as Equation 7 (here, we only have annotated HOI samples). We think the running concept confidence 𝐌\mathbf{M} have implicitly counted the distribution of verb and object in the dataset. Meanwhile, the denominator in Equation 2 can also normalize the confidence according to the frequency, and thus ease the long-tailed issue. Thus, with the pseudo labels constructed from 𝐌\mathbf{M}, we can re-balance the distribution of the dataset, which is a bit similar to re-weighting strategy [4, 11]. However, SCL does not require to set the weights for each class manually.

Table 15 demonstrates SCL greatly improves Qpic on Unseen category on rare first zero-shot detection, while SCL significantly facilitates rare category on non-rare first zero-shot detection. In Full HOI detection on HICO-DET, Table 14 shows SCL largely facilitates HOI detection on rare category. Particularly, the seen category in rare first setting includes 120 rare classes, while the seen category in non-rare first setting only includes 18 classes (all rare classes are in unseen category in non-rare first setting). Thus, SCL actually improves HOI detection for rare category. We think the concept confidence matrix internally learns the distribution of verb and objects and in the dataset. e.g., given an object, 𝐌\mathbf{M} illustrates the corresponding verb distribution.

Table 15: Zero-Shot HOI detection based on Qpic. Results are reported by mean average precision (%). Here, we split the classes of HOI into four categories in zero-shot setting, i.e. , Seen are categorized into rare and non-rare.
Method Unseen Rare NonRare Full
Qpic [55] (non-rare first) 21.03 19.12 25.59 23.19
Qpic+SCL (non-rare first) 21.73 22.43 26.03 24.34
Qpic [55] (rare first) 15.24 16.72 30.98 27.40
Qpic+SCL (rare first) 19.07 16.19 30.89 28.08

Two-Stage method. Considering the HOI concept discovery is mainly based on two-stage HOI detection approaches [27], it is direct and simple to evaluate the performance of self-training on HOI detection. Table 13 demonstrates the HOI detection results on ground truth boxes. Noticeably, we directly predict the verb category, rather than HOI category. Thus, the baseline of HOI detection (i.e. visual compositional learning [27]) is a bit worse. We can find self-training also slightly improves the performance, especially on rare category.

Appendix 0.G Visualization

In this section, we provide more visualized illustrations.

More Grad-CAM Visualizations Figure 5 demonstrates the visualization of Qpic and Qpic+SCL: the second row is Qpic and the third row is Qpic+SCL, where we observe a similar trend to the Gram-CAM illustration in main paper.

Refer to caption
Figure 5: Visualized Illustration of SCL+Qpic and Qpic [55].

Concept Visualization. We illustrate the visualized comparisons of concept discovery in Figure 6. According to the ground truth and known concepts, we find some verb (affordance) classes can be applied to most of objects (the row is highlighted in the ground truth figure). This observation is reasonable because some kinds of actions can be applied to most of objects in visual world, e.g., hold. As shown in Figure 6, there are many false positive predictions in the results of affordance prediction, and affordance prediction tends to overfit the known concepts, especially those with frequently appeared verbs. Methods of online HOI concept discovery on V-COCO have fewer false positive predictions compared to affordance prediction. However, the two methods tend to predict concepts composed of frequent verbs in known concepts due to the verb and object imbalance issues in HOI dataset [29]. Particularly, the false positive predictions are largely eased with self-training (e.g., the top right region). In addition, the blank columns in Figure 6 are because there are only 69 objects in V-COCO training set, and we can ease it via training network with additional object images [28] as illustrated in the last figure of Figure 6. See more visualized results on HICO-DET and V-COCO in the supplemental material. Particularly, we further notice there are dependencies between verb classes (See verb dependency analysis).

Refer to caption
Figure 6: Visualized Comparison of different methods on V-COCO dataset. The column is the object classes and the row represents the verb classes. Known Concepts are the concepts that we have known. SCL- means online concept discovery without self-training. For better illustration, we filter out known concepts in proposed methods. “+ Novel Objects” means self-training with novel object images.

Appendix 0.H Additional Concept Discovery Approaches

We provide More comparisons in this Section. For a fair comparison with ATL [28] (i.e. , affordance prediction), we use the same number of verbs (21 verbs) on V-COCO. The code includes how to convert V-COCO to 21 verbs, i.e. merge “_instr” and “_obj” and remove actions without object (e.g., stand, smile, run).

Language embedding baseline. In the main paper, we illustrate a random baseline. Here we further illustrate the results with language embedding [47]. Different from extracting verb/object features from real HOI images, we use the corresponding language embedding representations of verb/object as input, i.e. discovering concepts from language embedding. Table 12 shows the performance is just a bit better than random result, and is much worse than online concept discovery. Similar to the main paper, when we evaluate the unknown concepts, we mask out the known concepts to avoid the disturbance from known concepts.

Re-Training. We first train the HOI model via visual compositional learning [27], and then predict the concept confidence. Next, we use the predicted concept confidence to provide pseudo labels for the composite HOIs. Table 12 shows the performance of Re-Training is worse than SCL.

With COCO dataset. Table 12 also demonstrates the baseline (SCL-) with COCO datasets has poor performance on concept discovery. We think it is because the domain shift between COCO dataset and HICO-DET dataset. However, SCL still achieves significant improvement on concept discovery.

Qpic+SCL. The details are provided in Section D.