Abstract
The game of Tibetan Go faces the scarcity of expert knowledge and research literature. Therefore, we study the zero learning model of Tibetan Go under limited computing power resources and propose a novel scale-invariant U-Net style two-headed output lightweight network TibetanGoTinyNet. The lightweight convolutional neural networks and capsule structure are applied to the encoder and decoder of TibetanGoTinyNet to reduce computational burden and achieve better feature extraction results. Several autonomous self-attention mechanisms are integrated into TibetanGoTinyNet to capture the Tibetan Go board’s spatial and global information and select important channels. The training data are generated entirely from self-play games. TibetanGoTinyNet achieves 62%–78% winning rate against other four U-Net style models including Res-UNet, Res-UNet Attention, Ghost-UNet, and Ghost Capsule-UNet. It also achieves 75% winning rate in the ablation experiments on the attention mechanism with embedded positional information. The model saves about 33% of the training time with 45%–50% winning rate for different Monte-Carlo tree search (MCTS) simulation counts when migrated from 9 × 9 to 11 × 11 boards. Code for our model is available at https://github.com/paulzyy/TibetanGoTinyNet.
摘要
藏式围棋面临专家知识和研究文献匮乏的问题。因此, 我们研究了有限计算能力资源下藏式围棋的零学习模型, 并提出一种新颖的尺度不变U型网络(U-Net)风格的双头输出轻量级网络TibetanGoTinyNet。该网络的编码和解码器应用了轻量级卷积神经网络(CNN)和胶囊网络, 以减少计算负担并提升特征提取效果。网络中集成了数种自注意力机制, 以捕获藏式围棋棋盘的空间和全局信息, 并选择有价值通道。训练数据完全由自我对弈生成。TibetanGoTinyNet在与Res-UNet, Res-UNet Attention, Ghost-UNet和Ghost Capsule-UNet4个U-Net风格模型的对弈中获得了62%–78%的胜率。在捕获棋盘位置信息的轻量级自注意机制消融实验中, 它也实现了75%的胜率。当模型从99棋盘直接迁移到1111棋盘时, 该模型在不同的蒙特卡洛树搜索(MCTS)次数下节省了约33%的训练时间, 并获得了45%–50%的胜率。本文模型代码可在https://github.com/paulzyy/TibetanGoTinyNet上获取。
Similar content being viewed by others
Data availability
The data that support the findings of this study are available from the corresponding authors upon reasonable request. The code for the model is available at https://github.com/paulzyy/TibetanGoTinyNet.
References
Azad R, Bozorgpour IA, Asadi-Aghbolaghi M, et al., 2021. Deep frequency re-calibration U-Net for medical image segmentation. IEEE/CVF Int Conf on Computer Vision Workshops, p.3267–3276. https://doi.org/10.1109/ICCVW54120.2021.00366
Azad R, Aghdam EK, Rauland A, et al., 2022a. Medical image segmentation review: the success of U-Net. https://doi.org/10.48550/arXiv.2211.14830
Azad R, Khosravi N, Merhof D, 2022b. SMU-Net: style matching U-Net for brain tumor segmentation with missing modalities. https://arxiv.org/abs/2204.02961v1
Bougourzi F, Distante C, Dornaika F, et al., 2023. PDAtt-Unet: pyramid dual-decoder attention Unet for Covid-19 infection segmentation from CT-scans. Med Image Anal, 86:102797. https://doi.org/10.1016/j.media.2023.102797
Ding XW, Wang SS, 2021. Efficient Unet with depth-aware gated fusion for automatic skin lesion segmentation. J Intell Fuzzy Syst, 40(5):9963–9975. https://doi.org/10.3233/JIFS-202566
Gao YF, Wu LZ, Li HY, 2021. GomokuNet: a novel UNet-style network for Gomoku zero learning via exploiting positional information and multiscale features. IEEE Conf on Games, p.1–4. https://doi.org/10.1109/CoG52621.2021.9619111
Guo CL, Szemenyei M, Yi YG, et al., 2021. SA-UNet: spatial attention U-Net for retinal vessel segmentation. 25th Int Conf on Pattern Recognition, p.1236–1242. https://doi.org/10.1109/ICPR48806.2021.9413346
Guo YH, Cai B, Liang PP, et al., 2022. Efficient network with ghost tied block for heart segmentation. Proc SPIE 12032, Medical Imaging 2022: Image Processing, Article 120320A. https://doi.org/10.1117/12.2605538
Hai JJ, Qiao K, Chen J, et al., 2019. Fully convolutional DenseNet with multiscale context for automated breast tumor segmentation. J Healthc Eng, 2019:8415485. https://doi.org/10.1155/2019/8415485
Han K, Wang YH, Tian Q, et al., 2020. GhostNet: more features from cheap operations. Proc IEEE/CVF Conf on Computer Vision and Pattern Recognition, p.1577–1586. https://doi.org/10.1109/CVPR42600.2020.00165
He KM, Zhang XY, Ren SQ, et al., 2016. Deep residual learning for image recognition. Proc IEEE Conf on Computer Vision and Pattern Recognition, p.770–778. https://doi.org/10.1109/CVPR.2016.90
Heidler K, Mou LC, Baumhoer C, et al., 2022. HED-UNet: combined segmentation and edge detection for monitoring the Antarctic coastline. IEEE Trans Geosci Remote Sens, 60:4300514. https://doi.org/10.1109/TGRS.2021.3064606
Hou QB, Zhou DQ, Feng JS, 2021. Coordinate attention for efficient mobile network design. IEEE/CVF Conf on Computer Vision and Pattern Recognition, p.13708–13717. https://doi.org/10.1109/CVPR46437.2021.01350
Howard AG, Zhu ML, Chen B, et al., 2017. MobileNets: efficient convolutional neural networks for mobile vision applications. https://doi.org/10.48550/arXiv.1704.04861
Hu J, Shen L, Sun G, 2018. Squeeze-and-excitation networks. Proc IEEE/CVF Conf on Computer Vision and Pattern Recognition, p.7132–7141. https://doi.org/10.1109/CVPR.2018.00745
Huang G, Liu Z, Van Der Maaten L, et al., 2017. Densely connected convolutional networks. Proc IEEE Conf on Computer Vision and Pattern Recognition, p.2261–2269. https://doi.org/10.1109/CVPR.2017.243
Huang Z, Zhao YW, Liu YH, et al., 2021. GCAUNet: a group cross-channel attention residual UNet for slice based brain tumor segmentation. Biomed Signal Process Contr, 70:102958. https://doi.org/10.1016/j.bspc.2021.102958
Ibtehaz N, Rahman MS, 2020. MultiResUNet: rethinking the U-Net architecture for multimodal biomedical image segmentation. Neur Netw, 121:74–87. https://doi.org/10.1016/j.neunet.2019.08.025
Jing JF, Wang Z, Rätsch M, et al., 2022. Mobile-Unet: an efficient convolutional neural network for fabric defect detection. Text Res J, 92(1–2):30–42. https://doi.org/10.1177/0040517520928604
Kazerouni IA, Dooly G, Toal D, 2021. Ghost-UNet: an asymmetric encoder-decoder architecture for semantic segmentation from scratch. IEEE Access, 9:97457–97465. https://doi.org/10.1109/ACCESS.2021.3094925
Kocsis L, Szepesvári C, 2006. Bandit based Monte-Carlo planning. 17th European Conf on Machine Learning, p.282–293. https://doi.org/10.1007/11871842_29
Mamoon S, Manzoor MA, Zhang FE, et al., 2020. SPSSNet: a real-time network for image semantic segmentation. Front Inform Technol Electron Eng, 21(12):1770–1782. https://doi.org/10.1631/FITEE.1900697
Ronneberger O, Fischer P, Brox T, 2015. U-Net: convolutional networks for biomedical image segmentation. 18th Int Conf on Medical Image Computing and Computer-Assisted Intervention, p.234–241. https://doi.org/10.1007/978-3-319-24574-4_28
Sabour S, Frosst N, Hinton GE, 2017. Dynamic routing between capsules. Proc 31st Int Conf on Neural Information Processing Systems, p.3859–3869.
Saeed MU, Ali G, Bin W, et al., 2021. RMU-Net: a novel residual mobile U-Net model for brain tumor segmentation from MR images. Electronics, 10(16):1962. https://doi.org/10.3390/electronics10161962
Silver D, Huang A, Maddison CJ, et al., 2016. Mastering the game of Go with deep neural networks and tree search. Nature, 529(7587):484–489. https://doi.org/10.1038/nature16961
Silver D, Hubert T, Schrittwieser J, et al., 2017a. Mastering chess and shogi by self-play with a general reinforcement learning algorithm. https://doi.org/10.48550/arXiv.1712.01815
Silver D, Schrittwieser J, Simonyan K, et al., 2017b. Mastering the game of Go without human knowledge. Nature, 550(7676):354–359. https://doi.org/10.1038/nature24270
Soemers DJNJ, Piette É, Stephenson M, et al., 2022. The Ludii game description language is universal. https://doi.org/10.48550/arXiv.2205.00451
Tan MX, Le Q, 2019. EfficientNet: rethinking model scaling for convolutional neural networks. Proc 36th Int Conf on Machine Learning, p.6105–6114.
Tang YH, Han K, Guo JY, et al., 2022. GhostNetV2: enhance cheap operation with long-range attention. Proc 36th Int Conf on Neural Information Processing Systems.
Tian MJ, Li XL, Kong SH, et al., 2022. A modified YOLOv4 detection method for a vision-based underwater garbage cleaning robot. Front Inform Technol Electron Eng, 23(8):1217–1228. https://doi.org/10.1631/FITEE.2100473
Tran M, Vo-Ho VK, Le NTH, 2022. 3DConvCaps: 3DUnet with convolutional capsule encoder for medical image segmentation. 26th Int Conf on Pattern Recognition, p.4392–4398. https://doi.org/10.1109/ICPR56361.2022.9956588
Trebing K, Staùczyk T, Mehrkanoon S, 2021. SmaAt-UNet: precipitation nowcasting using a small attention-UNet architecture. Patt Recogn Lett, 145:178–186. https://doi.org/10.1016/j.patrec.2021.01.036
Woo S, Park J, Lee JY, et al., 2018. CBAM: convolutional block attention module. Proc 15th European Conf on Computer Vision, p.3–19. https://doi.org/10.1007/978-3-030-01234-2_1
Wu YH, Gao SH, Mei J, et al., 2021. JCS: an explainable COVID-19 diagnosis system by joint classification and segmentation. IEEE Trans Image Process, 30:3113–3126. https://doi.org/10.1109/TIP.2021.3058783
Xu YH, Li Q, He SY, et al., 2022. Ghost-Unet: an efficient convolutional neural network for spine MR image segmentation: lightweight segmentation method for spine MRI. Proc 4th Int Conf on Robotics, Intelligent Control and Artificial Intelligence, p.1159–1163. https://doi.org/10.1145/3584376.3584581
Xue LY, Lin JW, Cao XR, et al., 2019. A saliency and Gaussian net model for retinal vessel segmentation. Front Inform Technol Electron Eng, 20(8):1075–1086. https://doi.org/10.1631/FITEE.1700404
Author information
Authors and Affiliations
Contributions
Xiali LI designed the research. Yanyin ZHANG processed the data. Xiali LI and Yanyin ZHANG drafted the paper. Yandong CHEN helped process the data. Junzhi YU helped organize the paper. Licheng WU and Junzhi YU revised and finalized the paper.
Corresponding authors
Ethics declarations
All the authors declare that they have no conflict of interest.
Additional information
Project supported by the National Natural Science Foundation of China (Nos. 62276285 and 62236011) and the Major Projects of Social Science Fundation of China (No. 20&ZD279)
List of supplementary materials
1 Introduction
2 AlphaGo family and its improvements
3 Preliminary
4 Network structure
5 Experiments and discussion
Fig. S1 Hierarchical module of the proposed Tibetan-GoTinyNet
Fig. S2 The beginning and end modules of the network Fig. S3 TibetanGoTinyNet trained at a learning rate of 0.001 compared to the model trained at other learning rates
Fig. S4 The results of TibetanGoTinyNet against other models under several rollout set training conditions
Table S1 Winning rates of TibetanGoTinyNet against other models under different learning rate conditions
Electronic supplementary material
Rights and permissions
About this article
Cite this article
Li, X., Zhang, Y., Wu, L. et al. TibetanGoTinyNet: a lightweight U-Net style network for zero learning of Tibetan Go. Front Inform Technol Electron Eng 25, 924–937 (2024). https://doi.org/10.1631/FITEE.2300493
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1631/FITEE.2300493