Direct back propagation neural dynamic programming-based particle swarm optimisation

Yongzhong Lu, Danping Yan, Jingyu Zhang, David Levy

Research output: Contribution to journalArticle

4 Citations (Scopus)

Abstract

In this paper, we introduce direct back propagation (BP) neural dynamic programming (NDP) into particle swarm optimisation (PSO). Thus, a direct BP NDP inspired PSO algorithm, which we call NDPSO, is proposed. In NDPSO, since direct BP NDP belongs to the class of heuristic dynamic programming algorithms based on model-based adaptive critic designs and often serves as an online learning control paradigm, critic BP neural network is trained to optimise a total reward-to-go objective, namely to balance Bellman's equation, while action BP neural network is used to train the inertia weight, cognitive, and social coefficients so that the critic BP network output can approach an ultimate reward-to-go objective of success. With the collective aid of action-critic BP neural networks, inertia weight, cognitive, and social coefficients become more adaptive. Besides, the NDPSO's mutation mechanism also has greatly improved the dynamic performance of the standard PSO. Empirical experiments are conducted on both unimodal and multimodal benchmark functions. The experimental results demonstrate NDPSO's effectiveness and superiority to many other PSO variants on solving most multimodal problems.
Original languageEnglish
Pages (from-to)367-388
Number of pages22
JournalConnection Science
Volume26
Issue number4
DOIs
Publication statusPublished - 2014

Keywords

  • neural network
  • adaptive critic designs
  • particle swarm optimisation
  • back propagation
  • dynamic programming

Fingerprint Dive into the research topics of 'Direct back propagation neural dynamic programming-based particle swarm optimisation'. Together they form a unique fingerprint.

  • Cite this