首页 | 本学科首页   官方微博 | 高级检索  
     


FEREBUS: Highly parallelized engine for kriging training
Authors:Nicodemo Di Pasquale  Michael Bane  Stuart J. Davie  Paul L. A. Popelier
Affiliation:1. Manchester Institute of Biotechnology (MIB), 131 Princess Street, Manchester M1 7DN, Great Britain and School of Chemistry, University of Manchester, Manchester, Great Britain;2. Research IT, The University of Manchester and High End Compute, http://highendcompute.co.uk, Manchester
Abstract:FFLUX is a novel force field based on quantum topological atoms, combining multipolar electrostatics with IQA intraatomic and interatomic energy terms. The program FEREBUS calculates the hyperparameters of models produced by the machine learning method kriging. Calculation of kriging hyperparameters ( θ and p ) requires the optimization of the concentrated log‐likelihood urn:x-wiley:01928651:media:jcc24486:jcc24486-math-0001. FEREBUS uses Particle Swarm Optimization (PSO) and Differential Evolution (DE) algorithms to find the maximum of urn:x-wiley:01928651:media:jcc24486:jcc24486-math-0002. PSO and DE are two heuristic algorithms that each use a set of particles or vectors to explore the space in which urn:x-wiley:01928651:media:jcc24486:jcc24486-math-0003 is defined, searching for the maximum. The log‐likelihood is a computationally expensive function, which needs to be calculated several times during each optimization iteration. The cost scales quickly with the problem dimension and speed becomes critical in model generation. We present the strategy used to parallelize FEREBUS, and the optimization of urn:x-wiley:01928651:media:jcc24486:jcc24486-math-0004 through PSO and DE. The code is parallelized in two ways. MPI parallelization distributes the particles or vectors among the different processes, whereas the OpenMP implementation takes care of the calculation of urn:x-wiley:01928651:media:jcc24486:jcc24486-math-0005, which involves the calculation and inversion of a particular matrix, whose size increases quickly with the dimension of the problem. The run time shows a speed‐up of 61 times going from single core to 90 cores with a saving, in one case, of ~98% of the single core time. In fact, the parallelization scheme presented reduces computational time from 2871 s for a single core calculation, to 41 s for 90 cores calculation. © 2016 The Authors. Journal of Computational Chemistry Published by Wiley Periodicals, Inc.
Keywords:kriging  machine learning  OpenMP  MPI  parallellization  particle swarm optimization  differential evolution  QTAIM  force field design  IQA
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号