Quickstart

ParMOO is a parallel multiobjective optimization solver that seeks to exploit simulation-based structure in objective and constraint functions.

To exploit structure, ParMOO models simulations separately from objectives and constraints. In our language:

  • a design variable is an input to the problem, which we can directly control;

  • a simulation is an expensive or time-consuming process, including real-world experimentation, which is treated as a blackbox function of the design variables and evaluated sparingly;

  • an objective is an algebraic function of the design variables and/or simulation outputs, which we would like to optimize; and

  • a constraint is an algebraic function of the design variables and/or simulation outputs, which cannot exceed a specified bound.

Designs, simulations, and objectives

To solve a multiobjective optimization problem (MOOP), we use surrogate models of the simulation outputs, together with the algebraic definition of the objectives and constraints.

ParMOO animation

In order to achieve scalable parallelism, we use libEnsemble to distribute batches of simulation evaluations across parallel resources.

Dependencies

ParMOO has been tested on Unix/Linux and MacOS systems.

ParMOO’s base has the following dependencies:

  • Python 3.8+

  • numpy – for data structures and performant numerical linear algebra

  • scipy – for scientific calculations needed for specific modules

  • pyDOE – for generating experimental designs

  • pandas – for exporting the resulting databases

Additional dependencies are needed to use the additional features in parmoo.extras:

  • libEnsemble – for managing parallel simulation evaluations

And for using the Pareto front visualization library in parmoo.viz:

  • plotly – for generating interactive plots

  • dash – for hosting interactive plots in your browser

  • kaleido – for exporting static plots post-interaction

Installation

The easiest way to install ParMOO is via the Python package index, PyPI (commonly called pip):

pip install < --user > parmoo

where the braces around < --user > indicate that the --user flag is optional.

To install all dependencies (including libEnsemble) use:

pip install < --user > "parmoo[extras]"

You can also clone this project from our GitHub and pip install it in-place, so that you can easily pull the latest version or checkout the develop branch for pre-release features. On Debian-based systems with a bash shell, this looks like:

git clone https://github.com/parmoo/parmoo
cd parmoo
pip install -e .

Alternatively, the latest release of ParMOO (including all required and optional dependencies) can be installed from the conda-forge channel using:

conda install --channel=conda-forge parmoo

Before doing so, it is recommended to create a new conda environment using:

conda create --name channel-name
conda activate channel-name

For detailed instructions, see Advanced Installation.

Testing

If you have pytest with the pytest-cov plugin and flake8 installed, then you can test your installation.

python3 setup.py test

These tests are run regularly using GitHub Actions.

Basic Usage

ParMOO uses numpy in an object-oriented design, based around the MOOP class. To get started, create a MOOP object, using the constructor.

from parmoo import MOOP
from parmoo.optimizers import LocalGPS

my_moop = MOOP(LocalGPS)

To summarize the framework, in each iteration ParMOO models each simulation using a computationally cheap surrogate, then solves one or more scalarizations of the objectives, which are specified by acquisition functions. Read more about this framework at our Learn About MOOPs page. In the above example, optimizers.LocalGPS is the class of optimizers that the my_moop will use to solve the scalarized surrogate problems.

Next, add design variables to the problem as follows using the MOOP.addDesign(*args) method. In this example, we define one continuous and one categorical design variable. Other options include integer, custom, and raw (using raw variables is not recommended except for expert users).

# Add a single continuous design variable in the range [0.0, 1.0]
my_moop.addDesign({'name': "x1", # optional, name
                   'des_type': "continuous", # optional, type of variable
                   'lb': 0.0, # required, lower bound
                   'ub': 1.0, # required, upper bound
                   'tol': 1.0e-8 # optional tolerance
                  })
# Add a second categorical design variable with 3 levels
my_moop.addDesign({'name': "x2", # optional, name
                   'des_type': "categorical", # required, type of variable
                   'levels': ["good", "bad"] # required, category names
                  })

Next, add simulations to the problem as follows using the MOOP.addSimulation(*args) method. In this example, we define a toy simulation sim_func(x).

import numpy as np
from parmoo.searches import LatinHypercube
from parmoo.surrogates import GaussRBF

# Define a toy simulation for the problem, whose outputs are quadratic
def sim_func(x):
   if x["x2"] == "good":
      return np.array([(x["x1"] - 0.2) ** 2, (x["x1"] - 0.8) ** 2])
   else:
      return np.array([99.9, 99.9])
# Add the simulation to the problem
my_moop.addSimulation({'name': "MySim", # Optional name for this simulation
                       'm': 2, # This simulation has 2 outputs
                       'sim_func': sim_func, # Our sample sim from above
                       'search': LatinHypercube, # Use a LH search
                       'surrogate': GaussRBF, # Use a Gaussian RBF surrogate
                       'hyperparams': {}, # Hyperparams passed to internals
                       'sim_db': { # Optional dict of precomputed points
                                  'search_budget': 10 # Set search budget
                                 },
                      })

Now we can add objectives and constraints using MOOP.addObjective(*args) and MOOP.addConstraint(*args). In this example, there are 2 objectives (each corresponding to a single simulation output) and one constraint.

# First objective just returns the first simulation output
def f1(x, s): return s["MySim"][0]
my_moop.addObjective({'name': "f1", 'obj_func': f1})
# Second objective just returns the second simulation output
def f2(x, s): return s["MySim"][1]
my_moop.addObjective({'name': "f2", 'obj_func': f2})
# Add a single constraint, that x[0] >= 0.1
def c1(x, s): return 0.1 - x["x1"]
my_moop.addConstraint({'name': "c1", 'constraint': c1})

Finally, we must add one or more acquisition functions using MOOP.addAcquisition(*args). These are used to scalarize the surrogate problems. The number of acquisition functions typically determines the number of simulation evaluations per batch. This is useful to know if you are using a parallel solver.

from parmoo.acquisitions import RandomConstraint

# Add 3 acquisition functions
for i in range(3):
   my_moop.addAcquisition({'acquisition': RandomConstraint,
                           'hyperparams': {}})

Finally, the MOOP is solved using the MOOP.solve(budget) method, and the results can be viewed using MOOP.getPF().

import pandas as pd

my_moop.solve(5) # Solve with 5 iterations of ParMOO algorithm
results = my_moop.getPF(format="pandas") # Extract the results as pandas df

After executing the above block of code, the results variable points to a pandas dataframe, each of whose rows corresponds to a nondominated objective value in the my_moop object’s final database. You can reference individual columns in the results array by using the name keys that were assigned during my_moop’s construction, or plot the results by using the viz library.

Congratulations, you now know enough to get started solving MOOPs!

Minimal Working Example

Putting it all together, we get the following minimal working example.


import numpy as np
import pandas as pd
from parmoo import MOOP
from parmoo.searches import LatinHypercube
from parmoo.surrogates import GaussRBF
from parmoo.acquisitions import RandomConstraint
from parmoo.optimizers import LocalGPS

# Fix the random seed for reproducibility
np.random.seed(0)

my_moop = MOOP(LocalGPS)

my_moop.addDesign({'name': "x1",
                   'des_type': "continuous",
                   'lb': 0.0, 'ub': 1.0})
my_moop.addDesign({'name': "x2", 'des_type': "categorical",
                   'levels': ["good", "bad"]})

def sim_func(x):
   if x["x2"] == "good":
      return np.array([(x["x1"] - 0.2) ** 2, (x["x1"] - 0.8) ** 2])
   else:
      return np.array([99.9, 99.9])

my_moop.addSimulation({'name': "MySim",
                       'm': 2,
                       'sim_func': sim_func,
                       'search': LatinHypercube,
                       'surrogate': GaussRBF,
                       'hyperparams': {'search_budget': 20}})

def f1(x, s): return s["MySim"][0]
def f2(x, s): return s["MySim"][1]
my_moop.addObjective({'name': "f1", 'obj_func': f1})
my_moop.addObjective({'name': "f2", 'obj_func': f2})

def c1(x, s): return 0.1 - x["x1"]
my_moop.addConstraint({'name': "c1", 'constraint': c1})

for i in range(3):
   my_moop.addAcquisition({'acquisition': RandomConstraint,
                           'hyperparams': {}})

my_moop.solve(5)
results = my_moop.getPF(format="pandas")

# Display solution
print(results)

# Plot results -- must have extra viz dependencies installed
from parmoo.viz import scatter
# The optional arg `output` exports directly to jpg instead of interactive mode
scatter(my_moop, output="jpeg")

The above code saves all (approximate) Pareto optimal solutions in the results variable, and prints the results variable to the standard output:

         x1    x2        f1        f2        c1
0  0.742840  good  0.294675  0.003267 -0.642840
1  0.726092  good  0.276773  0.005462 -0.626092
2  0.605914  good  0.164766  0.037669 -0.505914
3  0.548931  good  0.121753  0.063036 -0.448931
4  0.543499  good  0.117991  0.065793 -0.443499
5  0.401011  good  0.040405  0.159192 -0.301011
6  0.353552  good  0.023578  0.199316 -0.253552
7  0.328402  good  0.016487  0.222404 -0.228402
8  0.269175  good  0.004785  0.281775 -0.169175
9  0.248183  good  0.002322  0.304502 -0.148183

And produces the following figure of the Pareto points:

Scatter plot of the Pareto front after solving demo problem

Next Steps

  • If you want to take advantage of all that ParMOO has to offer, please see Writing a ParMOO Script.

  • If you would like more information on multiobjective optimization terminology and ParMOO’s methodology, see the Learn About MOOPs page.

  • For a full list of basic usage tutorials, see More Tutorials.

  • To start solving MOOPs on parallel hardware, install libEnsemble and see the libEnsemble tutorial.

  • See some of our pre-built solvers in the parmoo_solver_farm.

  • To interactively explore your solutions, install its extra dependencies and use our built-in viz tool.

  • For more advice, consult our FAQs.

Resources

To seek support or report issues, e-mail:

  • parmoo@mcs.anl.gov

Our full documentation is hosted on:

Please read our LICENSE and CONTRIBUTING files.