Parallelizing Comprehensive Learning Particle Swarm Optimization by Open Computing Language on an Integrated Graphical Processing Unit
Author
dc.contributor.author
Yu, Xiang
Author
dc.contributor.author
Qiao, Yu
Author
dc.contributor.author
Li, Qingpeng
Author
dc.contributor.author
Xu, Gang
Author
dc.contributor.author
Kang, Chuanxiong
Author
dc.contributor.author
Estévez Montero, Claudio
Author
dc.contributor.author
Deng, Chengzhi
Author
dc.contributor.author
Wang, Shengqian
Admission date
dc.date.accessioned
2020-10-29T13:00:30Z
Available date
dc.date.available
2020-10-29T13:00:30Z
Publication date
dc.date.issued
2020
Cita de ítem
dc.identifier.citation
Complexity Volume 2020, Article ID 6589658, 17 pages
es_ES
Identifier
dc.identifier.other
10.1155/2020/6589658
Identifier
dc.identifier.uri
https://repositorio.uchile.cl/handle/2250/177465
Abstract
dc.description.abstract
Comprehensive learning particle swarm optimization (CLPSO) is a powerful metaheuristic for global optimization. This paper studies parallelizing CLPSO by open computing language (OpenCL) on the integrated Intel HD Graphics 520 (IHDG520) graphical processing unit (GPU) with a low clock rate. We implement a coarse-grained all-GPU model that maps each particle to a separate work item. Two enhancement strategies, namely, generating and transferring random numbers from the central processor to the GPU as well as reducing the number of instructions in the kernel, are proposed to shorten the model's execution time. This paper further investigates parallelizing deterministic optimization for implicit stochastic optimization of China's Xiaowan Reservoir. The deterministic optimization is performed on an ensemble of 62 years' historical inflow records with monthly time steps, is solved by CLPSO, and is parallelized by a coarse-grained multipopulation model extended from the all-GPU model. The multipopulation model involves a large number of work items. Because of the capacity limit for a buffer transferring data from the central processor to the GPU and the size of the global memory region, the random number generation strategy is modified by generating a small number of random numbers that can be flexibly exploited by the large number of work items. Experiments conducted on various benchmark functions and the case study demonstrate that our proposed all-GPU and multipopulation parallelization models are appropriate; and the multipopulation model achieves the consumption of significantly less execution time than the corresponding sequential model.
es_ES
Patrocinador
dc.description.sponsorship
National Natural Science Foundation of China (NSFC)
61703199
61866023
61865012
Shaanxi Province Natural Science Foundation Basic Research Project
2020JM-278
Central Universities Fundamental Research Foundation Project
GK202003006