Object-to-Scene: Learning to Transfer Object Knowledge to Indoor Scene Recognition
Abstract
Accurate perception of the surrounding scene is helpful for robots to make reasonable judgments and behaviours. Therefore, developing effective scene representation and recognition methods are of significant importance in robotics. Currently, a large body of research focuses on developing novel auxiliary features and networks to improve indoor scene recognition ability. However, few of them focus on directly constructing object features and relations for indoor scene recognition. In this paper, we analyze the weaknesses of current methods and propose an Object-to-Scene (OTS) method, which extracts object features and learns object relations to recognize indoor scenes. The proposed OTS first extracts object features based on the segmentation network and the proposed object feature aggregation module (OFAM). Afterwards, the object relations are calculated and the scene representation is constructed based on the proposed object attention module (OAM) and global relation aggregation module (GRAM). The final results in this work show that OTS successfully extracts object features and learns object relations from the segmentation network. Moreover, OTS outperforms the state-of-the-art methods by more than 2% on indoor scene recognition without using any additional streams. Code is publicly available at: https://github.com/FreeformRobotics/OTS.
I Introduction
Intelligent robots are human companions and assistants with the goal to improve the quality of life. For instance, robots can be designed to guide blind people to navigate by telling them where they are and which direction to go. For this application, there is an urgent need to equip robots with semantic scene understanding ability, e.g. the ability to perceive the surrounding objects and the scene that all objects constitute. Therefore, developing an effective method for indoor scene representation and recognition bears importance in improving the level of robotic perception and intelligence.
Indoor scene representation based recognition has been proposed for more than one decade, and it is still a challenging task in robotics and computer vision due to several issues: (1) conventional networks cannot focus on every object in scenes because they do not extract features at the object granularity [1]; (2) handcrafted object features are sub-optimal to represent scenes [2]; (3) an effective paradigm is lacking to represent coexisting object relations [3].

This work primarily aims to mitigate the above-mentioned issues and improve the accuracy of indoor scene representation and recognition. To achieve these goals, we propose a novel one-stream method, namely Object-to-Scene (OTS), that constructs object features and calculates object relations to realize indoor scene representation and recognition. In this work, instead of simply using handcrafted features such as [3] to represent the existence of each object, object features with higher semantic level are extracted using the proposed object feature aggregation module (OFAM) and a segmentation network. Object attention module (OAM) is then used to learn object relations implicitly, the attention mechanism of the module can obtain long-range dependencies, and thus coexisting objects in all situations can be learned. After that, Global relation aggregation module (GRAM) based on a strip depthwise convolution and a pointwise convolution is used to aggregate the object features and convert object features into a scene representation vector. Finally, a fully-connected layer is used for indoor scene recognition, and recognition results are compared with the existing state-of-the-art methods. Fig. 1 illustrates the superiority of our method, where the charts on the right show the top 5 recognition scores of our method and ResNet50 on the image. Compared to ResNet50 that focuses on major objects such as stool and table in scenes, the segmentation network in our method can detect tiny but important objects, such as stove and microwave in Fig. 1, and our method can represent object features and relations implicitly. Therefore, ResNet50 hesitates between wet bar and kitchen but OTS, based on object features, can recognize kitchen confidently.
In summary, the major contributions of this paper are as follows:
-
•
We propose a novel framework OTS that enables using object features and relations for indoor scene representation and recognition, and OTS outperforms existing methods by more than 2%.
-
•
We propose OFAM that can extract object features from the segmentation network for indoor scene representation and recognition.
- •
-
•
We propose GRAM to fuse the object features and relations into scene representation at a higher semantic level.
The remainder of the paper is organized as follows. Section II describes a detailed list of recent works in scene representation and recognition. The details of the proposed OTS are described in Section III. Section IV presents the experimental settings and analyzes the results of OTS. Finally, the conclusion and future work are presented in Section V.
II Related Works
Understanding surrounding scenes is helpful for robots to make reasonable judgments and behaviours, and a proper scene representation is an important prerequisite [6, 7, 8]. Many early approaches used statistics-based and hand-crafted features for scene representation [9, 10]. Li et al. proposed codebooks to represent local features and scenes [11]. Quattoni et al. used prototypes that contain object information for different indoor scene representations [12]. Liu et al. proposed fast adaptive color tags to describe each indoor scene, and used tag matching methods in U-V color space and geometric space for inference [13]. However, these statistics-based and hand-crafted features have limited semantic information and are sub-optimal to represent scenes as there are many combinations of objects in the same and different scenes. In that case, an effective scene representation method is urgently needed.
As computing power increases, many deep learning methods have been proposed for various vision tasks [14, 15, 16, 17, 18]. ResNet is one of the most prominent backbone feature extractors since it cleverly avoids the vanishing gradient problem in neural networks [1]. However, ResNet behaves poorly in indoor scene representation and recognition due to the lack of effective expression of coexisting small objects and object relations. To solve this problem, some methods try to combine the backbone features with the object and semantic features from detection networks and segmentation networks for scene representation and recognition. Chen et al. merged backbone features, detection features and segmentation features into an embedding for indoor scene representation [19]. Sun et al. used the method of spatial fisher vectors to extract object features from detection network, and combined the object features with contextual and global appearance features for scene recognition [20]. Alejandro et al. combined the semantic features from segmentation network with backbone features for scene recognition using an attention module [21]. Pal et al. used detection network to compute a binary feature vector that represents whether each object existed in the scene, the binary object vector is then combined with backbone features for indoor scene recognition [3]. Zhou et al. used a Bayesian method to represent the co-occurrence of object pairs for better scene representation and recognition [22]. Zeng et al. added scene attributes into image features and patch features at multi-scale for scene recognition [23]. Although the above mentioned methods improved the scene representation ability and recognition accuracy. These methods are sub-optimal in terms of constructing object features and relations. As shown in Fig. 2, the co-occurrence probability of some coexisting objects, such as bed and lamp, varies significantly from the bedroom to other scenes. Therefore, learning object relations and object features well could further improve the representation ability and recognition accuracy of models. Inspired by the above observations, we propose a novel one-stream method called OTS that enables object features and relations for scene representation and recognition, and our method outperforms existing methods.

III Methodology
We propose OTS for indoor scene representation and recognition. In OTS, we propose OFAM which enables small but discriminative objects to determine the final recognition results. To ensure effective utilization of object features for scene representation, we propose OAM to learn object relations based on its attention mechanism and a GRAM to fuse object features and relations. Fig. 3 illustrates the framework of the proposed OTS. Given an input image , where is the image data and is the corresponding category, OFAM is used to calculate the object features at first. OAM and GRAM are then used to calculate the scene representation . Finally, a fully connected layer and the softmax function are used to calculate the final probability of each scene .

III-A Object Feature Aggregation Module (OFAM)
The proposed OFAM can extract object features based on the segmentation network. In this paper, PSPNet that is pretrained on ADE20k dataset is used as the segmentation network [24, 25]. Fig. 4 illustrates the proposed OFAM. OFAM uses Conv4_6 feature map and segmentation score map that obtained from PSPNet to calculate object features , where is the number of channels, is the number of objects and is the number of feature spatial positions. In this paper, is set to 150 because of the object number of ADE20K, and is set to 1024 because Conv4_6 is used. For convenience, we call each spatial position in as a unit, the channel values of each unit form the feature vector of the unit. The binary mask of objects is calculated as:
where is the unit index and is the object index. Then, the object feature vector is calculated as:
where is the unit index and is the object index, and represents the unit number. Finally, the feature vectors stacked together to form the object features .
III-B Object Attention Module (OAM)
In addition to object features, relations between objects are also required to improve scene representation ability and recognition accuracy. For example, a model will be more confident about the kitchen label if stove and refrigerator have been detected together. In that case, an attention mechanism is a good choice since it can capture long-range dependencies between all objects [26]. However, only global attention mechanism is suitable for our case because object features do not include spatial relations.

Self-attention [4] and non-local [5] are one of the most effective global attention mechanisms where three nodes are calculated in parallel with 11 convolutions and the input feature map. We call the three nodes Query (), Key (), and Value () for convenience. and are then multiplied together to calculate the correlation matrix of each unit in the feature map, and softmax is used to calculate the activation map. After that, and the activation map are multiplied together, and an additional 11 convolution is used to form the final attention. Finally, the attention multiplies a learnable scale parameter and adds back to the input feature map to form the output. Fig. 5(b) illustrates the structure of self-attention and non-local. Although self-attention and non-local are effective, they can be further improved. For example, the attention will be more efficient if and are calculated based on , because has fewer channels compared with input features. Leveraging this insight, we propose an OAM that consists of cascaded object attention blocks as shown in Fig. 5(a). Given an input feature map , is first obtained using 11 convolutions, where is the channel number, is the unit number, is a factor to control the compression rate and the output channel number, and . and are then calculated based on , where and . The object attention is then calculated as follow:
Parm. (M) | FLOPs (M) | ||||||
---|---|---|---|---|---|---|---|
Self-attention | 1024 | 1024 | 128 | 128 | 512 | 1.3 | 211.0 |
Non-local | 1024 | 1024 | 512 | 512 | 512 | 2.1 | 337.6 |
Object Attention Block | 1024 | 1024 | 512 | 512 | 512 | 1.1 | 180.3 |
Where is a learnable scale parameter that can be updated gradually. The intuition for why we propose object attention block is straightforward. Firstly, is used to refine input information with fewer channels (512/) in object attention, calculating and based on enables them to perceive refined input information with half the computational cost. Secondly, is used to compress the block and control the output channel number, the connectivity of , and allows them to be automatically compressed proportionally with the change of . Thirdly, concatenation mechanism in object attention avoids using extra convolution and ensures the later block perceive the refined input information and attention features separately. Table I compares the computational cost of three methods. It can be seen that our object attention block is the most efficient of the three methods under the same input and output channel number configuration. Our object attention block is also more flexible because we can adjust to reduce the computational cost arbitrarily, and the channel number of , , , and output will be changed automatically.

III-C Global Relation Aggregation Module (GRAM)
Aggregation of object features is essential for accurate scene recognition. To solve this problem, we propose a GRAM that consists of a strip depthwise convolution and a pointwise convolution as shown in Fig. 6. Given an input feature map , the strip depthwise convolution first aggregates object features and relations at each channel, and converts the feature map into , pointwise convolution is then used to generate scene representation vector that has higher semantic level. Depthwise convolution combined with pointwise convolution is much more efficient compared with conventional convolution kernel [27]. The conventional depthwise convolution has a 3x3 kernel to learn local information in each channel. However, we aim to aggregate the object features and relations into a scene representation vector, and learn the global relation at the same time. Therefore, the conventional depthwise convolution is not suitable for our case. To solve this problem, we proposed a strip depthwise convolution, which operates on a list of object features rather than a patch of spatial features. The intuition for why we use this is clear. Firstly, the strip depthwise convolution can convert object feature map into a representation vector, that is more suitable for the final recognition layer. Secondly, the convolution is also able to aggregate the relations between all objects in each channel. Finally, the convolution is more efficient not only because of the depthwise characteristic, but also the avoiding of the use of large amount of fully connected layers.

IV Experiments and Results
IV-A Implementation Details
Our network is implemented in the Pytorch library [28], and a single RTX 2080Ti GPU is used in the experiments. We use the most common settings to implement the experiments [1]. The batch size for all experiments is set to 256 and cross-entropy loss is used in our method. We adopt Stochastic Gradient Descent (SGD) optimizer with the base learning rate of 0.1 while momentum and weight decay are set to 0.9 and 0.0001, respectively. The learning rate is divided by 10 every 10 epochs. Training is performed for 40 epochs. Meanwhile, object features are calculated off-line to increase the training speed.
IV-B Places365 Dataset
The reduced Places365 dataset is used in this paper since it is the largest scene recognition dataset with various indoor environment categories [29]. To verify the effectiveness of our method, we used two different class settings of indoor scenes for a fair comparison with other state-of-the-art methods. The first one contains 7 classes: Bathroom, Bedroom, Corridor, Dining room, Kitchen, Living room, and Office. We extract both training data and test data from the target 7 classes in Places365 and form the Places365-7classes, the class setting is totally the same as [3]. The second one includes 14 indoor scenes in home environment: Balcony, Bedroom, Dining room, Home office, Kitchen, Living room, Staircase, Bathroom, Closet, Garage, Home theater, Laundromat, Playroom, and Wet bar. We also extract both training data and test data from the target 14 classes in Places365 and form the Places365-14classes, the class setting is totally the same as [19].
IV-C SUN-RGBD Dataset
SUN-RGBD is one of the most challenging datasets for scene understanding [30], which includes images from various sources: 3784 images captured by Kinect v2; 1159 images captured by Intel RealSense cameras; 1449 images from NYU Depth V2 [31] and 554 manually selected realistic scene images from Berkeley B3DO [32] captured by Kinect v1; 3389 selected frames by filtering out significantly blurred frames from the SUN3D videos [33] captured by Asus Xtion. The diversity of categories and sources makes SUN-RGBD more suitable for verifying the generalization ability of methods. We only consider the RGB images in this work, and the 7 classes that chosen in Places365 are used to test our model’s generalization ability. Notably, we only used the official test images in SUN-RGBD, and these images are evaluated by the model trained using Places365-7 classes.
[1] | [3] | Our | ||||
Category | ResNet18 | ResNet50 | Scene | Obj. | Scene+Obj. | OTS |
Bathroom | 87 | 94 | 92 | 65 | 91 | 92 |
Bedroom | 82 | 83 | 90 | 74 | 90 | 97 |
Corridor | 96 | 93 | 94 | 90 | 96 | 95 |
Dining room | 81 | 71 | 79 | 94 | 79 | 88 |
Kitchen | 83 | 84 | 87 | 62 | 87 | 92 |
Living room | 55 | 66 | 84 | 25 | 80 | 79 |
Office | 79 | 88 | 85 | 29 | 94 | 88 |
Avg. Acc. (%) | 80.4 | 82.7 | 87.3 | 62.6 | 88.1 | 90.1 |
IV-D Main Results
To evaluate the effectiveness of the proposed OTS, we compare it with other benchmark methods on the Places365-7classes, Places365-14classes and reduced SUN-RGBD datasets. As shown in Table II, OTS significantly outperforms the ResNet50 baseline with 7.4% on Places365-7classes which shows the effectiveness of OTS. Moreover, OTS has 2% higher accuracy than [3] with only one stream. [3] used a detection network to calculate one-hot object existence features, and added an additional stream to calculate auxiliary image features.
To verify the generalization ability of OTS, we further test OTS on SUN-RGBD. It is noteworthy that SUN-RGBD is just used as a test set to verify the generalization ability of models, and all models are trained using Plces365-7classes. As shown in Table III, OTS achieves the highest accuracy compared with other methods on SUN-RGBD. The results demonstrate the generalization ability of OTS.
We further compare OTS with [19] on the larger Places365-14classes. [19] used a segmentation network to calculate Word2Vec features, and added an additional stream to calculate image features. As shown in Table IV, our OTS is 2.2% higher than [19] without any additional streams. The results show the effectiveness of our method. Better segmentation models have the potential to improve the performance of OTS, but they are not the focus of this work. In addition to evaluating the benchmark datasets, we also inference OTS in a real-world office environment, which can be found in our video supplement files.
ResNet50 | OFAM | OAM | GRAM | Acc. (%) |
---|---|---|---|---|
✓ | 80.0 | |||
✓ | ✓ | 77.2 | ||
✓ | ✓ | ✓ | 82.0 | |
✓ | ✓ | ✓ | ✓ | 85.9 |

IV-E Ablation Experiments
We run ablation experiments to analyze the results of our method. Unless specified otherwise, the dataset used in ablation experiments is Places365-14classes since it is larger and contains more categories. We firstly run an ablation experiment to show the effectiveness of each module. As shown in Table V, both OAM and GRAM significantly improve the scene recognition accuracy because of the ability to capture the long-range dependencies between all objects. Meanwhile, only adding OFAM degrades the performance of the model because ResNet50 uses the features of Conv5 layer but OFAM extracts the object features from Conv4 layer as shown in Fig. 3, which means the performance gains of our method rely on the object relation construction instead of the backbone features of the pre-trained segmentation model.
Num. | SUM. | CAT. | Acc. (%) | Parm. (M) | FLOPs (M) | |
---|---|---|---|---|---|---|
OAB | 1 | ✓ | 85.0 | 0.4 | 70.5 | |
OAB | 2 | ✓ | 85.2 | 1.8 | 321.3 | |
OAB | 1 | ✓ | 85.4 | 0.4 | 70.5 | |
OAB | 2 | ✓ | 85.9 | 1.2 | 211.5 |
IV-E1 Number of object attention blocks
As stated in Section III-B, OAM is built based on one or several cascaded object attention blocks. Therefore, we run ablation experiments to find how many object attention blocks is the optimal choice to balance efficiency and effectiveness. Unless specified otherwise, the of the first object attention block is 2 and the of the second object attention block is 0.5 to control the output channel number. Before that, we first verify the difference between using summation (SUM.) and using concatenation (CAT.) in the last step of object attention blocks. The experiment results are shown in Table VI. It can be seen that using concatenation in object attention blocks is both more efficient and effective compared with using summation. This is because simple summation cannot perfectly fuse object features and object relation features together, but concatenation can store both types of features separately. Then, we use concatenation in object attention blocks to find the optimal number of blocks. As shown in Table VII, we list the results as well as FLOPs obtained with different number of object attention blocks. It can be seen that when two object attention blocks are cascaded, OTS has the best accuracy 85.9%. However, adding more object attention blocks exert an adverse impact on the results because these blocks make the model too complicated.
Num. | Acc. (%) | Parm. (M) | FLOPs (M) | |
---|---|---|---|---|
OAB | 1 | 85.4 | 0.4 | 70.5 |
OAB | 2 | 85.9 | 1.2 | 211.5 |
OAB | 3 | 85.0 | 1.4 | 262.3 |
OAB | 4 | 85.0 | 1.7 | 313.2 |
IV-E2 Object attention block vs Self-attention and Non-local
We have shown the effectiveness of OAM in Table V. In addition, We run an ablation experiment to show the effectiveness and efficiency of the proposed object attention blocks in OAM. As shown in Table VIII, the accuracy of the proposed object attention block has an improvement of 1.2% compared with self-attention, and an improvement of 3.6% compared with non-local. What is more, the proposed object attention block has the least FLOPs compared with others under a similar condition. The object attention block (S) means only one object attention block is used in OAM.
IV-E3 Global relation aggregation module (GRAM)
We set two kinds of control groups to show the effectiveness of the proposed GRAM. In the ablation experiment, fully connected layer and Pooling layer are used to replace GRAM, separately. As shown in Table IX, GRAM obtained the best accuracy with only 2.3M parameters. It is noteworthy that FC in Table IX is equal to a large conventional convolution kernel with the size of input feature map. FC not only has more computational cost, but also has inferior performance compared with the proposed GRAM. The results indicate that GRAM can balance efficiency and effectiveness well.
Acc. (%) | Parm. (M) | FLOPs (M) | |
---|---|---|---|
Non-Local [5] | 82.3 | 4.2 | 675.2 |
Self-Attention [4] | 84.7 | 2.6 | 422.0 |
Object Attention Block | 85.9 | 1.2 | 211.5 |
Object Attention Block (S) | 85.4 | 0.4 | 70.5 |
Acc. (%) | Parm. (M) | FLOPs (M) | |
---|---|---|---|
FC | 85.0 | 314.6 | 314.6 |
Max & Avg. Pooling | 82.0 | 0 | 0.3 |
GRAM | 85.9 | 2.3 | 2.3 |
IV-E4 Failure cases
During experiments, we find that ResNet50 misclassified many common images. However, OTS successfully recognized the common images in scenes because it can detect the objects and learn their relations in each scene. Then, we analyze the failure cases of OTS to demonstrate the pros and cons. As shown in Fig. 7, OTS misclassified some images in each scene mainly caused by the missed or wrong detection. For example, kitchen is misclassified as bathroom because the pot is mistaking for a toilet, and office is misclassified as corridor because only wall, painting, ceiling, floor, and light are detected, which are the common coexisting objects in corridor. Therefore, the performance of OTS could be further improved with the help of more accurate segmentation results.
V Conclusions and Future Work
In this paper, we analyzed the weakness of existing scene representation and recognition methods, and proposed OTS to solve these issues. We further demonstrated that OTS can effectively use object features and relations for scene representation and recognition by comparing OTS with other existing state-of-the-art methods. Based on numerous ablation experiments, we also showed that OAM and GRAM perform well in learning object relations for scene representation. Moreover, the results of our work reflect several interesting conclusions: 1) object features can perform well as long as an appropriate object feature and relation learning method is used; 2) the backbone features in segmentation network can also be used for scene recognition instead of adding an additional stream to calculate; 3) attention mechanism is very suitable for computing object relations. We hope these results could guild future works for scene understanding. What is more, our work can also be promoted in the future. First of all, data augmentation methods can be used to extract more diverse object features during training. Secondly, enriching the number of object categories can offer a better scene representation, and thus improve scene recognition ability of the model. In the future, we plan to extend our methods to other mobile robots and establish more accurate semantic maps. Therefore, they can be better used to improve human’s life quality.
References
- [1] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- [2] P. Quelhas, F. Monay, J.-M. Odobez, D. Gatica-Perez, T. Tuytelaars, and L. Van Gool, “Modeling scenes with local descriptors and latent aspects,” in Tenth IEEE International Conference on Computer Vision (ICCV’05) Volume 1, vol. 1. IEEE, 2005, pp. 883–890.
- [3] A. Pal, C. Nieto-Granda, and H. I. Christensen, “Deduce: Diverse scene detection methods in unseen challenging environments,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2019, pp. 4198–4204.
- [4] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena, “Self-attention generative adversarial networks,” in International Conference on Machine Learning, 2019, pp. 7354–7363.
- [5] X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 7794–7803.
- [6] Y. Liao, S. Kodagoda, Y. Wang, L. Shi, and Y. Liu, “Understand scene categories by objects: A semantic regularized scene classifier using convolutional neural networks,” in 2016 IEEE international conference on robotics and automation (ICRA). IEEE, 2016, pp. 2318–2325.
- [7] C. Ye, Y. Yang, R. Mao, C. Fermüller, and Y. Aloimonos, “What can i do around here? deep functional scene understanding for cognitive robots,” in 2017 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2017, pp. 4604–4611.
- [8] F. Yan, S. Nannapaneni, and H. He, “Robotic scene understanding by using a dictionary,” in 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO). IEEE, 2019, pp. 895–900.
- [9] S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories,” in 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), vol. 2. IEEE, 2006, pp. 2169–2178.
- [10] S. H. Khan, M. Bennamoun, F. Sohel, and R. Togneri, “Geometry driven semantic labeling of indoor scenes,” in European Conference on Computer Vision. Springer, 2014, pp. 679–694.
- [11] L. Fei-Fei and P. Perona, “A bayesian hierarchical model for learning natural scene categories,” in 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), vol. 2. IEEE, 2005, pp. 524–531.
- [12] A. Quattoni and A. Torralba, “Recognizing indoor scenes,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 2009, pp. 413–420.
- [13] M. Liu, D. Scaramuzza, C. Pradalier, R. Siegwart, and Q. Chen, “Scene recognition with omnidirectional vision for topological map using lightweight adaptive descriptors,” in 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2009, pp. 116–121.
- [14] H. Lei, N. Akhtar, and A. Mian, “Spherical kernel for efficient graph convolution on 3d point clouds,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
- [15] K. He, G. Gkioxari, P. Dollár, and R. Girshick, “Mask r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961–2969.
- [16] Y. Hristov, D. Angelov, M. Burke, A. Lascarides, and S. Ramamoorthy, “Disentangled relational representations for explaining and learning from demonstration,” in Conference on Robot Learning, 2020, pp. 870–884.
- [17] L. Zhang, G. Zhu, L. Mei, P. Shen, S. A. A. Shah, and M. Bennamoun, “Attention in convolutional lstm for gesture recognition,” in Advances in Neural Information Processing Systems, 2018, pp. 1953–1962.
- [18] L. Herranz, S. Jiang, and X. Li, “Scene recognition with cnns: objects, scales and dataset bias,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 571–579.
- [19] B. X. Chen, R. Sahdev, D. Wu, X. Zhao, M. Papagelis, and J. K. Tsotsos, “Scene classification in indoor environments for robots using context based word embeddings,” in 2018 IEEE International Conference on Robotics and Automation (ICRA) Workshop, 2018.
- [20] N. Sun, W. Li, J. Liu, G. Han, and C. Wu, “Fusing object semantics and deep appearance features for scene recognition,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 29, no. 6, pp. 1715–1728, 2018.
- [21] A. López-Cifuentes, M. Escudero-Viñolo, J. Bescós, and Á. García-Martín, “Semantic-aware scene recognition,” Pattern Recognition, vol. 102, p. 107256, 2020.
- [22] L. Zhou, C. Jun, X. Wang, Z. Sun, T. L. Lam, and Y. Xu, “Borm: Bayesian object relation model for indoor scene recognition,” in IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2021.
- [23] H. Zeng, X. Song, G. Chen, and S. Jiang, “Learning scene attribute for scene recognition,” IEEE Transactions on Multimedia, 2019.
- [24] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 2881–2890.
- [25] B. Zhou, H. Zhao, X. Puig, S. Fidler, A. Barriuso, and A. Torralba, “Scene parsing through ade20k dataset,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 633–641.
- [26] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in neural information processing systems, 2017, pp. 5998–6008.
- [27] C. François, “Xception: Deep learning with depthwise separable convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1800–1807.
- [28] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “Pytorch: An imperative style, high-performance deep learning library,” in Advances in neural information processing systems, 2019, pp. 8026–8037.
- [29] B. Zhou, A. Lapedriza, A. Khosla, A. Oliva, and A. Torralba, “Places: A 10 million image database for scene recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017.
- [30] S. Song, S. P. Lichtenberg, and J. Xiao, “Sun rgb-d: A rgb-d scene understanding benchmark suite,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 567–576.
- [31] N. Silberman, D. Hoiem, P. Kohli, and R. Fergus, “Indoor segmentation and support inference from rgbd images,” in European conference on computer vision. Springer, 2012, pp. 746–760.
- [32] A. Janoch, S. Karayev, Y. Jia, J. T. Barron, M. Fritz, K. Saenko, and T. Darrell, “A category-level 3d object dataset: Putting the kinect to work,” in Consumer depth cameras for computer vision. Springer, 2013, pp. 141–165.
- [33] J. Xiao, A. Owens, and A. Torralba, “Sun3d: A database of big spaces reconstructed using sfm and object labels,” in Proceedings of the IEEE International Conference on Computer Vision, 2013, pp. 1625–1632.