skip to main content
10.1145/3640771.3640774acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiscaiConference Proceedingsconference-collections
research-article

Optimization of Chinese Character Font Recognition Model Based on Incremental Learning

Published:29 March 2024Publication History

ABSTRACT

Chinese character font recognition is a vital research area with applications in text recognition, artistic font design, and handwriting recognition. Traditional methods for this task are time-consuming, labor-intensive, and error-prone. The rise of deep learning has opened new possibilities in this field. However, neural network models often struggle with generalization and robustness, especially with new data, leading to catastrophic forgetting of old data. To address this, we propose an optimization algorithm for the Chinese character font recognition model SwordNet, combining parameter isolation and regularization. We designed an incremental module for SwordNet, enhancing it for better performance and scalability. The loss function incorporates a regularization term based on knowledge distillation principles, limiting significant parameter changes for old data and ensuring the model retains knowledge while learning new data. To tackle the diversity of Chinese character fonts and aid model recognition, we introduced a data augmentation method using noise, rotation, and occlusion. Experiments demonstrate that our regularized incremental learning optimization algorithm achieves a 98.81% accuracy rate with only new data, comparable to models trained on both new and old datasets. Notably, it also reduces training time by approximately 30%, marking a significant improvement in efficiency and accuracy for Chinese character font recognition.

References

  1. Abhishek Aich. 2021. Elastic Weight Consolidation (EWC): Nuts and Bolts. arxiv:2105.04093Google ScholarGoogle Scholar
  2. Rahaf Aljundi, Lucas Caccia, Eugene Belilovsky, Massimo Caccia, Min Lin, Laurent Charlin, and Tinne Tuytelaars. 2019. Online Continual Learning with Maximally Interfered Retrieval. arxiv:1908.04742Google ScholarGoogle Scholar
  3. Xiang Bai, Baoguang Shi, Chengquan Zhang, Xuan Cai, and Li Qi. 2017. Text/Non-Text Image Classification in the Wild with Convolutional Neural Networks. Pattern Recogn. 66, C (jun 2017), 437–446. https://doi.org/10.1016/j.patcog.2016.12.005Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Francisco M. Castro, Manuel J. Marín-Jiménez, Nicolás Guil, Cordelia Schmid, and Karteek Alahari. 2018. End-to-End Incremental Learning. arxiv:1807.09536 [cs.CV]Google ScholarGoogle Scholar
  5. Arslan Chaudhry, Marc’Aurelio Ranzato, Marcus Rohrbach, and Mohamed Elhoseiny. 2019. Efficient Lifelong Learning with A-GEM. arxiv:1812.00420Google ScholarGoogle Scholar
  6. Jingchao Chen, Shiyi Mu, Shugong Xu, and Youdong Ding. 2021. HENet: Forcing a Network to Think More for Font Recognition. arxiv:2110.10872 [cs.CV]Google ScholarGoogle Scholar
  7. Matthias Delange, Rahaf Aljundi, Marc Masana, Sarah Parisot, Xu Jia, Ales Leonardis, Greg Slabaugh, and Tinne Tuytelaars. 2021. A continual learning survey: Defying forgetting in classification tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence (2021), 1–1. https://doi.org/10.1109/tpami.2021.3057446Google ScholarGoogle ScholarCross RefCross Ref
  8. Arthur Douillard, Matthieu Cord, Charles Ollion, Thomas Robert, and Eduardo Valle. 2020. PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning. arxiv:2004.13513 [cs.CV]Google ScholarGoogle Scholar
  9. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the Knowledge in a Neural Network. arxiv:1503.02531 [stat.ML]Google ScholarGoogle Scholar
  10. Xinting Hu, Kaihua Tang, Chunyan Miao, Xian-Sheng Hua, and Hanwang Zhang. 2021. Distilling Causal Effect of Data in Class-Incremental Learning. arxiv:2103.01737 [cs.AI]Google ScholarGoogle Scholar
  11. James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, Demis Hassabis, Claudia Clopath, Dharshan Kumaran, and Raia Hadsell. 2017. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences 114, 13 (mar 2017), 3521–3526. https://doi.org/10.1073/pnas.1611835114Google ScholarGoogle ScholarCross RefCross Ref
  12. Xudong Li, Jingyi Wang, Haiyang Zhang, Yongke Huang, and Huihui Huang. 2022. SwordNet: Chinese Character Font Style Recognition Network. IEEE Access 10 (2022), 8388–8398. https://doi.org/10.1109/ACCESS.2022.3143795Google ScholarGoogle ScholarCross RefCross Ref
  13. Zhizhong Li and Derek Hoiem. 2017. Learning without Forgetting. arxiv:1606.09282Google ScholarGoogle Scholar
  14. Charles X. Ling and Tanner Bohn. 2021. A Deep Learning Framework for Lifelong Machine Learning. arxiv:2105.00157Google ScholarGoogle Scholar
  15. Zheda Mai, Ruiwen Li, Jihwan Jeong, David Quispe, Hyunwoo Kim, and Scott Sanner. 2021. Online Continual Learning in Image Classification: An Empirical Survey. arxiv:2101.10423Google ScholarGoogle Scholar
  16. Arun Mallya and Svetlana Lazebnik. 2018. PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning. arxiv:1711.05769 [cs.CV]Google ScholarGoogle Scholar
  17. Sylvestre-Alvise Rebuffi, Alexander Kolesnikov, Georg Sperl, and Christoph H. Lampert. 2017. iCaRL: Incremental Classifier and Representation Learning. arxiv:1611.07725Google ScholarGoogle Scholar
  18. Joan Serrà, Dídac Surís, Marius Miron, and Alexandros Karatzoglou. 2018. Overcoming catastrophic forgetting with hard attention to the task. arxiv:1801.01423Google ScholarGoogle Scholar
  19. Zhiqing Sun, Hongkun Yu, Xiaodan Song, Renjie Liu, Yiming Yang, and Denny Zhou. 2020. MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices. arxiv:2004.02984Google ScholarGoogle Scholar
  20. Shipeng Yan, Jiangwei Xie, and Xuming He. 2021. DER: Dynamically Expandable Representation for Class Incremental Learning. arxiv:2103.16788 [cs.CV]Google ScholarGoogle Scholar

Index Terms

  1. Optimization of Chinese Character Font Recognition Model Based on Incremental Learning

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          ISCAI '23: Proceedings of the 2023 2nd International Symposium on Computing and Artificial Intelligence
          October 2023
          120 pages
          ISBN:9798400708954
          DOI:10.1145/3640771

          Copyright © 2023 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 29 March 2024

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited
        • Article Metrics

          • Downloads (Last 12 months)15
          • Downloads (Last 6 weeks)15

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format