An Empirical Study on the Effect of a Parallel Distributed Implementation of GAssist

  • Ivarsson Patrik
    Department of Computer Science and Intelligent Systems, Osaka Prefecture University
  • Nojima Yusuke
    Department of Computer Science and Intelligent Systems, Osaka Prefecture University
  • Ishibuchi Hisao
    Department of Computer Science and Intelligent Systems, Osaka Prefecture University

Search this article

Abstract

GAssist is a high-performing machine learning algorithm, which obtains a classifier through genetic learning. In this paper we examine the effects of implementing GAssist as a parallel distributed model. In our parallel distributed model, the population of individuals is divided into multiple subpopulations. Training data are also divided into multiple subsets, which are subsequently assigned one per subpopulation. In each subpopulation, a genetic algorithm is performed separately from the other subpopulations. Additionally, we rotate the training data subset used for each subpopulation periodically. In doing this, we avoid over-fitting of each subpopulation to the local training data subset. Through computational experiments we examine the effectiveness of our parallel distributed model with respect to its search ability, generalization ability and computation time.

Journal

Details 詳細情報について

Report a problem

Back to top