Skip to main content

Bayesian Optimization with Gaussian Process as surrogate model

Project description

GPGO - Gaussian Process GO

My own implementation of a Bayesian Black box Optimization with Gaussian Process as a surrogate model. It is still in development but it was successfully used to achieve a bottom up optimization of the Dissipative Particle Dynamics force field for a complex system of polymers chains functionalized gold nanoparticles in a water solvent.

Hyperparameters

The Hyperparameters of the GP are optimized by the common technique of maximizing the Log Marginal Likelihood. In this repository this is achieved by using a search grid (although not in an efficient way) or by using the scipy optimizer module (L-BFGS-B, TNC, SLSCP). The analytical gradient is implemented for the Radial Basis Function kernel and it is possible to use the derivate of the Log Marginal Likelihood to optimize the hyperparameters.

Figure-6

Acquisition function

As it is there are two different acquisition function implemented right now:

-Expected Improvement (EI)

-UCB (Upper Confidence Bound)

Maximizing the Acquisition function

In this little package right now there are 3 ways to run an optimization task with Gaussian Processes:

-NAIVE : AkA sampling the acquisition function with a grid of some kind or a quasi random methods as LHS (require smt package)

-BFGS : optimize the Acquisition function by using the L-BFGS-B optimizer

-DIRECT : optimize the Acquisition function by using the DIRECT optimizer (require DIRECT python package)

-PSO : optimize the Acquisition function by a Particle Swarm Optimization genetic algorithm

Figure-7

Made for experiments

Easy to use it with a shell procedure! Load the data and just .suggest_location() to get the next points of your experiment!

Multi Objective

Right now the package contains an implementation of the NSGAII genetic solver that allows to solve multi objective problems. It has also an early version of a Multi Objective Bayesian optimization that uses the NSGAII and optimize the Acquisition function EI or the mean function of the GP. It will follow a more precise implementation of the Hypervolume improvement.

pareto

TODO

-Load data routines

-Good code practice maybe

-Easy routines for LAMMPS (at least DPD and BD)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

BayGPGO-0.3.2.tar.gz (30.9 kB view details)

Uploaded Source

File details

Details for the file BayGPGO-0.3.2.tar.gz.

File metadata

  • Download URL: BayGPGO-0.3.2.tar.gz
  • Upload date:
  • Size: 30.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.22.0 setuptools/45.2.0.post20200210 requests-toolbelt/0.9.1 tqdm/4.46.0 CPython/3.7.6

File hashes

Hashes for BayGPGO-0.3.2.tar.gz
Algorithm Hash digest
SHA256 d1bea558896482f6770ae0bb6293e3f36b371e13e5e0eed57454c5c3422b7d0f
MD5 ff3f305d2d42e1139dccb7509204d8b9
BLAKE2b-256 00b863a4fa07aa8db0099d18fac558dbcdd94a052afd66b8b3f8fee42c920fa7

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page