ABSTRACT
To mine high-dimensional rules in Learning Classifier Systems (LCSs) through a reduction of the dimensionality of input data, this paper proposes a novel approach that indirectly learns the rules in the "latent space" based on the rewards of the reconstructed rules in the "observation space". We call this approach Learning Strategy by exploring rules in Observation space via Latent space (LS-OvL), which is based on two rule representations in the observation and latent space. Concretely, LS-OvL explores the rules by searching the latent space as the reduced dimensional input space by an autoencoder and evaluates them in the observation space by reconstructing them from the latent space. Such a design is significant because it prevents the generation of inaccurate rules during the reconstitution process from the latent space to the observation space. Through a comparison LS-OvL with the conventional learning strategy, which explores and evaluates the rules in the only latent space and reconstructs them in the observation space, the experimental results show that (1) LS-OvL outperforms the conventional learning strategy in terms of the acquired reward and the population size, and (2) LS-OvL can generate explainable and classifiable high-dimensional rules.
- Isidro M. Alvarez, Will N. Browne, and Mengjie Zhang. 2014. Reusing Learned Functionality in XCS: Code Fragments with Constructed Functionality and Constructed Features. In Proceedings of the Companion Publication of the 2014 Annual Conference on Genetic and Evolutionary Computation (Vancouver, BC, Canada) (GECCO Comp '14). Association for Computing Machinery, New York, NY, USA, 969--976. Google ScholarDigital Library
- Alejandro Barredo Arrieta, Natalia Diaz Rodriguez, Javier Del Ser, Adrien Bennetot, Siham Tabik, Alberto Barbado González, Salvador Garcia, Sergio Gil-Lopez, Daniel Molina, V. Richard Benjamins, Raja Chatila, and Francisco Herrera. 2019. Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. Information Fusion 58 (12 2019). Google ScholarDigital Library
- Ester Bernadó-Mansilla and Josep M. Garrell-Guiu. 2003. Accuracy-Based Learning Classifier Systems: Models, Analysis and Applications to Classification Tasks. Evolutionary Computation 11, 3 (sep 2003), 209--238. Google ScholarDigital Library
- M.V. Butz, T. Kovacs, P.L. Lanzi, and S.W. Wilson. 2004. Toward a theory of generalization and learning in XCS. IEEE Transactions on Evolutionary Computation 8, 1 (Feb 2004), 28--46. Google ScholarDigital Library
- Martin V Butz and Stewart W Wilson. 2002. An algorithmic description of XCS. Soft Computing 6, 3--4 (2002), 144--153.Google ScholarCross Ref
- Geoffrey E Hinton and Ruslan R Salakhutdinov. 2006. Reducing the dimensionality of data with neural networks. science 313, 5786 (2006), 504--507.Google Scholar
- John H Holland. 1986. The possibilities of general-purpose learning algorithms applied to parallel rule-based systems. Machine learning, an artificial intelligence approach 2 (1986), 593--623.Google Scholar
- Muhammad Irfan, Zheng Jiangbin, Muhammad Iqbal, and Muhammad Hassan Arif. 2021. Enhancing learning classifier systems through convolutional autoencoder to classify underwater images. Soft Computing 25 (2021), 10423--10440.Google ScholarDigital Library
- Ji-Yoon Kim and Sung-Bae Cho. 2019. Exploiting deep convolutional neural networks for a neural-based learning classifier system. Neurocomputing 354 (2019), 61--70. Recent Advancements in Hybrid Artificial Intelligence Systems. Google ScholarDigital Library
- Diederik P Kingma and Max Welling. 2013. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013).Google Scholar
- Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 1998. Gradient-based learning applied to document recognition. Proc. IEEE 86, 11 (1998), 2278--2324.Google ScholarCross Ref
- Yi Liu, Will N Browne, and Bing Xue. 2019. Absumption to complement subsumption in learning classifier systems. In Proceedings of the genetic and evolutionary computation conference. 410--418.Google ScholarDigital Library
- Yoshiki Nakamura, Motoki Horiuchi, and Masaya Nakata. 2021. Convergence Analysis of Rule-Generality on the XCS Classifier System. In Proceedings of the Genetic and Evolutionary Computation Conference (Lille, France) (GECCO '21). Association for Computing Machinery, New York, NY, USA, 332--339. Google ScholarDigital Library
- Masaya Nakata, Will Browne, Tomoki Hamagami, and Keiki Takadama. 2017. Theoretical XCS Parameter Settings of Learning Accurate Classifiers. In Proceedings of the Genetic and Evolutionary Computation Conference (Berlin, Germany) (GECCO '17). Association for Computing Machinery, New York, NY, USA, 473--480. Google ScholarDigital Library
- Hiroki Shiraishi, Masakazu Tadokoro, Yohei Hayamizu, Yukiko Fukumoto, Hiroyuki Sato, and Keiki Takadama. 2021. Increasing Accuracy and Interpretability of High-Dimensional Rules for Learning Classifier System. In 2021 IEEE Congress on Evolutionary Computation (CEC). 311--318. Google ScholarDigital Library
- Kihyuk Sohn, Honglak Lee, and Xinchen Yan. 2015. Learning Structured Output Representation using Deep Conditional Generative Models. In Advances in Neural Information Processing Systems, C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett (Eds.), Vol. 28. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2015/file/8d55a249e6baa5c06772297520da2051-Paper.pdfGoogle Scholar
- Christopher Stone and Larry Bull. 2003. For Real! XCS with Continuous-Valued Inputs. Evolutionary Computation 11, 3 (2003), 299--336.Google ScholarDigital Library
- Masakazu Tadokoro, Satoshi Hasegawa, Takato Tatsumi, Hiroyuki Sato, and Keiki Takadama. 2019. Knowledge Extraction from XCSR Based on Dimensionality Reduction and Deep Generative Models. In 2019 IEEE Congress on Evolutionary Computation (CEC). IEEE, 1883--1890.Google ScholarDigital Library
- Masakazu Tadokoro, Satoshi Hasegawa, Takato Tatsumi, Hiroyuki Sato, and Keiki Takadama. 2020. Local Covering: Adaptive Rule Generation Method Using Existing Rules for XCS. In 2020 IEEE Congress on Evolutionary Computation (CEC). 1--8. Google ScholarDigital Library
- Masakazu Tadokoro, Hiroyuki Sato, and Keiki Takadama. 2021. XCS with Weight-based Matching in VAE Latent Space and Additional Learning of High-Dimensional Data. In 2021 IEEE Congress on Evolutionary Computation (CEC). 304--310. Google ScholarDigital Library
- Ryan J Urbanowicz and Will N Browne. 2017. Introduction to learning classifier systems. Springer.Google ScholarDigital Library
- Stewart W Wilson. 1995. Classifier fitness based on accuracy. Evolutionary computation 3, 2 (1995), 149--175.Google Scholar
- Stewart W Wilson. 1998. Generalization in the XCS Classifier System. Proc. Genetic Programming (1998).Google Scholar
- Stewart W Wilson. 1999. Get real! XCS with continuous-valued inputs. In International Workshop on Learning Classifier Systems. Springer, 209--219.Google ScholarDigital Library
- Naoya Yatsu, Hiroki Shiraishi, Hiroyuki Sato, and Keiki Takadama. 2022. XCSR with VAE using Gaussian Distribution Matching: From Point to Area Matching in Latent Space for Less-overlapped Rule Generation in Observation Space. In 2022 IEEE Congress on Evolutionary Computation (CEC). 1--8. Google ScholarDigital Library
Index Terms
- Exploring High-dimensional Rules Indirectly via Latent Space Through a Dimensionality Reduction for XCS
Recommendations
Inheritance vs. Expansion: Generalization Degree of Nearest Neighbor Rule in Continuous Space as Covering Operator of XCS
Applications of Evolutionary ComputationAbstractThis paper focuses on the covering mechanism which generates a new if-then rule when the input data does not match the rules in the XCS Classifier System (XCS), a rule-based machine learning system, and discusses how the new rule should be ...
Mechanisms to Alleviate Over-Generalization in XCS for Continuous-Valued Input Spaces
AbstractIn the field of rule-based approaches to Machine Learning, the XCS classifier system (XCS) is a well-known representative of the learning classifier systems family. By using a genetic algorithm (GA), the XCS aims at forming rules or so-called ...
On the Effects of Absumption for XCS with Continuous-Valued Inputs
Applications of Evolutionary ComputationAbstractThe rule-based XCS Classifier System (XCS) aims at forming classifiers which are as general as possible to achieve an optimal performance level. A too high generalization pressure may lead to over-general classifiers degrading the performance of ...
Comments