Skip to main content

Advertisement

Log in

TibetanGoTinyNet: a lightweight U-Net style network for zero learning of Tibetan Go

TibetanGoTinyNet:一种应用于藏式围棋的U型网络风格的轻量级零学习模型

  • Research Article
  • Published:
Frontiers of Information Technology & Electronic Engineering Aims and scope Submit manuscript

Abstract

The game of Tibetan Go faces the scarcity of expert knowledge and research literature. Therefore, we study the zero learning model of Tibetan Go under limited computing power resources and propose a novel scale-invariant U-Net style two-headed output lightweight network TibetanGoTinyNet. The lightweight convolutional neural networks and capsule structure are applied to the encoder and decoder of TibetanGoTinyNet to reduce computational burden and achieve better feature extraction results. Several autonomous self-attention mechanisms are integrated into TibetanGoTinyNet to capture the Tibetan Go board’s spatial and global information and select important channels. The training data are generated entirely from self-play games. TibetanGoTinyNet achieves 62%–78% winning rate against other four U-Net style models including Res-UNet, Res-UNet Attention, Ghost-UNet, and Ghost Capsule-UNet. It also achieves 75% winning rate in the ablation experiments on the attention mechanism with embedded positional information. The model saves about 33% of the training time with 45%–50% winning rate for different Monte-Carlo tree search (MCTS) simulation counts when migrated from 9 × 9 to 11 × 11 boards. Code for our model is available at https://github.com/paulzyy/TibetanGoTinyNet.

摘要

藏式围棋面临专家知识和研究文献匮乏的问题。因此, 我们研究了有限计算能力资源下藏式围棋的零学习模型, 并提出一种新颖的尺度不变U型网络(U-Net)风格的双头输出轻量级网络TibetanGoTinyNet。该网络的编码和解码器应用了轻量级卷积神经网络(CNN)和胶囊网络, 以减少计算负担并提升特征提取效果。网络中集成了数种自注意力机制, 以捕获藏式围棋棋盘的空间和全局信息, 并选择有价值通道。训练数据完全由自我对弈生成。TibetanGoTinyNet在与Res-UNet, Res-UNet Attention, Ghost-UNet和Ghost Capsule-UNet4个U-Net风格模型的对弈中获得了62%–78%的胜率。在捕获棋盘位置信息的轻量级自注意机制消融实验中, 它也实现了75%的胜率。当模型从99棋盘直接迁移到1111棋盘时, 该模型在不同的蒙特卡洛树搜索(MCTS)次数下节省了约33%的训练时间, 并获得了45%–50%的胜率。本文模型代码可在https://github.com/paulzyy/TibetanGoTinyNet上获取。

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Data availability

The data that support the findings of this study are available from the corresponding authors upon reasonable request. The code for the model is available at https://github.com/paulzyy/TibetanGoTinyNet.

References

Download references

Author information

Authors and Affiliations

Authors

Contributions

Xiali LI designed the research. Yanyin ZHANG processed the data. Xiali LI and Yanyin ZHANG drafted the paper. Yandong CHEN helped process the data. Junzhi YU helped organize the paper. Licheng WU and Junzhi YU revised and finalized the paper.

Corresponding authors

Correspondence to Xiali Li  (李霞丽) or Junzhi Yu  (喻俊志).

Ethics declarations

All the authors declare that they have no conflict of interest.

Additional information

Project supported by the National Natural Science Foundation of China (Nos. 62276285 and 62236011) and the Major Projects of Social Science Fundation of China (No. 20&ZD279)

List of supplementary materials

1 Introduction

2 AlphaGo family and its improvements

3 Preliminary

4 Network structure

5 Experiments and discussion

Fig. S1 Hierarchical module of the proposed Tibetan-GoTinyNet

Fig. S2 The beginning and end modules of the network Fig. S3 TibetanGoTinyNet trained at a learning rate of 0.001 compared to the model trained at other learning rates

Fig. S4 The results of TibetanGoTinyNet against other models under several rollout set training conditions

Table S1 Winning rates of TibetanGoTinyNet against other models under different learning rate conditions

Electronic supplementary material

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, X., Zhang, Y., Wu, L. et al. TibetanGoTinyNet: a lightweight U-Net style network for zero learning of Tibetan Go. Front Inform Technol Electron Eng 25, 924–937 (2024). https://doi.org/10.1631/FITEE.2300493

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1631/FITEE.2300493

Key words

关键词

CLC number