Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Boosting Cooperative Coevolution for Large Scale Optimization with a Fine-Grained Computation Resource Allocation Strategy (1802.09703v2)

Published 27 Feb 2018 in cs.NE

Abstract: Cooperative coevolution (CC) has shown great potential in solving large scale optimization problems (LSOPs). However, traditional CC algorithms often waste part of computation resource (CR) as they equally allocate CR among all the subproblems. The recently developed contribution-based CC (CBCC) algorithms improve the traditional ones to a certain extent by adaptively allocating CR according to some heuristic rules. Different from existing works, this study explicitly constructs a mathematical model for the CR allocation (CRA) problem in CC and proposes a novel fine-grained CRA (FCRA) strategy by fully considering both the theoretically optimal solution of the CRA model and the evolution characteristics of CC. FCRA takes a single iteration as a basic CRA unit and always selects the subproblem which is most likely to make the largest contribution to the total fitness improvement to undergo a new iteration, where the contribution of a subproblem at a new iteration is estimated according to its current contribution, current evolution status as well as the estimation for its current contribution. We verified the efficiency of FCRA by combining it with SHADE which is an excellent differential evolution variant but has never been employed in the CC framework. Experimental results on two benchmark suites for LSOPs demonstrate that FCRA significantly outperforms existing CRA strategies and the resultant CC algorithm is highly competitive in solving LSOPs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhigang Ren (16 papers)
  2. Yongsheng Liang (28 papers)
  3. Aimin Zhang (1 paper)
  4. Yang Yang (884 papers)
  5. Zuren Feng (5 papers)
  6. Lin Wang (403 papers)
Citations (31)

Summary

We haven't generated a summary for this paper yet.