ABSTRACT
Spatial scene classification has long been a prominent area of research in the field of geographic information science. In the past, traditional approaches heavily relied on retrieval methods based on image features. However, given the rapid advancements in deep learning and artificial intelligence, the efficient classification of complex spatial scenes has become increasingly crucial. This paper presents a novel framework named WYA (Where You At) that combines surrounding object detection with knowledge graph to automate the process of spatial scene classification. Initially, the input images undergo processing using object detection techniques to identify key entities within the scenes. Subsequently, a knowledge graph, which encompasses various spatial scenes, entities, and their relationships, is utilized to identity spatial scene catogories. To validate the effectiveness of the framework, experiments were conducted using eight spatial scene categories as an example. The results demonstrated a high level of consistency with actual spatial types, thus affirming the efficacy of the framework and highlighting its potential application value in the domain of spatial scene classification.
- Alexey Bochkovskiy, Chien-Yao Wang, and Hong-Yuan Mark Liao. 2020. YOLOv4: Optimal Speed and Accuracy of Object Detection. ArXiv abs/2004.10934 (2020). https://api.semanticscholar.org/CorpusID:216080778Google Scholar
- Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. 2016. The Cityscapes Dataset for Semantic Urban Scene Understanding. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 3213--3223. https://doi.org/10.1109/CVPR.2016.350Google ScholarCross Ref
- N. Dalal and B. Triggs. 2005. Histograms of oriented gradients for human detection. In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), Vol. 1. 886--893 vol. 1. https://doi.org/10.1109/CVPR.2005.177Google ScholarDigital Library
- Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. ImageNet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition. 248--255. https://doi.org/10.1109/CVPR.2009.5206848Google ScholarCross Ref
- Pedro F. Felzenszwalb, Ross B. Girshick, and David McAllester. 2010. Cascade object detection with deformable part models. In 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2241--2248. https://doi.org/10.1109/CVPR.2010.5539906Google ScholarCross Ref
- Ross Girshick. 2015. Fast R-CNN. In 2015 IEEE International Conference on Computer Vision (ICCV). 1440--1448. https://doi.org/10.1109/ICCV.2015.169Google ScholarDigital Library
- Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. 2017. Mask R-CNN. In 2017 IEEE International Conference on Computer Vision (ICCV). 2980--2988. https://doi.org/10.1109/ICCV.2017.322Google ScholarCross Ref
- Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 770--778.Google Scholar
- Luis Herranz, Shuqiang Jiang, and Xiangyang Li. 2016. Scene Recognition with CNNs: Objects, Scales and Dataset Bias. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016), 571--579. https://api.semanticscholar.org/CorpusID:15429030Google ScholarCross Ref
- Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. ImageNet classification with deep convolutional neural networks. Commun. ACM 60 (2012), 84--90.Google ScholarDigital Library
- Wenbin Li, Sajad Saeedi, John McCormac, Ronald Clark, Dimos Tzoumanikas, Qing Ye, Yuzhong Huang, Rui Tang, and Stefan Leutenegger. 2018. InteriorNet: Mega-scale Multi-sensor Photo-realistic Indoor Scenes Dataset. In British Machine Vision Conference (BMVC).Google Scholar
- Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. 2017. Feature Pyramid Networks for Object Detection. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 936--944. https://doi.org/10.1109/CVPR.2017.106Google ScholarCross Ref
- Tsung Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollár. 2017. Focal Loss for Dense Object Detection. IEEE Transactions on Pattern Analysis Machine Intelligence PP, 99 (2017), 2999--3007.Google Scholar
- Tsung Yi Lin, Michael Maire, Serge Belongie, James Hays, and C. Lawrence Zitnick. 2014. Microsoft COCO: Common Objects in Context. In European Conference on Computer Vision.Google Scholar
- Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng Yang Fu, and Alexander C. Berg. 2016. SSD:Single Shot MultiBox Detector. In European Conference on Computer Vision, Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling (Eds.). 21--37.Google Scholar
- David G. Lowe. 2004. Distinctive Image Features from Scale-Invariant Keypoints. International Journal of Computer Vision 60, 2 (2004), 91--110.Google ScholarDigital Library
- Ariadna Quattoni and Antonio Torralba. 2009. Recognizing indoor scenes. In 2009 IEEE Conference on Computer Vision and Pattern Recognition. 413--420. https://doi.org/10.1109/CVPR.2009.5206537Google ScholarCross Ref
- Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You Only Look Once: Unified, Real-Time Object Detection. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 779--788. https://doi.org/10.1109/CVPR.2016.91Google ScholarCross Ref
- Joseph Redmon and Ali Farhadi. 2017. YOLO9000: Better, Faster, Stronger. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 6517--6525. https://doi.org/10.1109/CVPR.2017.690Google ScholarCross Ref
- Joseph Redmon and Ali Farhadi. 2018. YOLOv3: An Incremental Improvement. ArXiv abs/1804.02767 (2018). https://api.semanticscholar.org/CorpusID:4714433Google Scholar
- Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. 2017. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence 39, 6 (2017), 1137--1149. https://doi.org/10.1109/TPAMI.2016.2577031Google ScholarDigital Library
- K. Simonyan and A. Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In International Conference on Learning Representations.Google Scholar
- Mingxing Tan, Ruoming Pang, and Quoc V. Le. 2020. EfficientDet: Scalable and Efficient Object Detection. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 10778--10787. https://doi.org/10.1109/CVPR42600.2020.01079Google ScholarCross Ref
- P. Viola and M. Jones. 2001. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001, Vol. 1. I--I. https://doi.org/10.1109/CVPR.2001.990517Google ScholarCross Ref
- Yajun Wang, Qizi Mu, Moyin Yu, Honghai Wang, and Liyuan Zhu. 2023. Out-door Scene Recognition Based on Convolutional Neural Network. Automation Application 64, 201--207 (2023).Google Scholar
- Degang Xu, Lu Wand, and Fan Li. 2021. Review of Typical Object Detection Algorithms for Deep Learning. Computer Engineering and Applications 10--25 (2021).Google Scholar
- Shun Zhang, Yihong Gong, and Jinjun Wang. 2019. The Development of Deep Convolution Neural Network and Its Applications on Computer Vision. Chinese Journal of Computers 42, 453--482 (2019).Google Scholar
- Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. 2018. Places: A 10 Million Image Database for Scene Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 6 (2018), 1452--1464. https://doi.org/10.1109/TPAMI.2017.2723009Google ScholarCross Ref
- Zhengxia Zou, Keyan Chen, Zhenwei Shi, Yuhong Guo, and Jieping Ye. 2023. Object Detection in 20 Years: A Survey. Proc. IEEE 111, 3 (2023), 257--276.Google ScholarCross Ref
Index Terms
- WYA: A Novel Spatial Scene Classification Framework Based on Surrounding Object Detection
Recommendations
NDMFCS: An automatic fruit counting system in modern apple orchard using abatement of abnormal fruit detection
Highlights- Abatement of abnormal fruit detection led to more accurate fruit detection.
- Tracked trunk offered a reliable reference displacement in consecutive video frames.
- Identity document (ID) assignment efficiently counted fruits in modern ...
AbstractAutomatic fruit counting is an important task for growers to estimate yield and manage orchards. Although many deep-learning-based fruit detection algorithms have been developed to improve performance of automatic fruit counting systems, abnormal ...
Robust object tracking via multi-cue fusion
A long-term object tracking method based on calibrated binocular cameras by fusing information of the two channels and binocular geometry constraints is proposed.The stereo filter which is built based on the epipolar geometry of the binocular cameras is ...
Complete and accurate holly fruits counting using YOLOX object detection
Highlights- Complete and accurate counting of fruit on the tree.
- High performance single ...
AbstractFruits counting is important in management of orchard and plantation since better decision for labor and logistic can be made based on complete and accurate counting of fruits. Computer vision-based fruits counting has been research ...
Comments