Skip to content

πŸ“° Must-Read Papers on Offline Model-Based Optimization πŸ”₯

Notifications You must be signed in to change notification settings

mila-iqia/Awesome-Offline-Model-Based-Optimization

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

21 Commits
Β 
Β 
Β 
Β 

Repository files navigation

Awesome-Offline-Model-Based-Optimization

πŸ“° Must-Read Papers on Offline Model-Based Optimization πŸ”₯

This repository collects important papers for our new survey: "Offline Model-Based Optimization: Comprehensive Review" (will release soon).

Awesome Stars Forks

  • πŸ’»: Code
  • πŸ“–: bibtex

Latest Updates

  • [2025/03/04] First Release of Awesome-Offline-Model-Based Optimization

πŸ” Table of Contents

🌟 What is Offline Model-Based Optimization?

In offline optimization, the goal is to discover a new design, denoted by $\boldsymbol{x}^*$, that maximizes the objective(s) $\boldsymbol{f}(\boldsymbol{x})$. This is achieved using an offline dataset $\mathcal{D}$, which consists of $N$ designs paired with their property labels. In particular, the dataset is given by Offline Dataset where each design vector $\boldsymbol{x}_i$ belongs to a design space $\mathcal{X} \subseteq \mathbb{R}^d$, and each property label $\boldsymbol{y}_i \in \mathbb{R}^m$ contains the corresponding $m$ objective values for that design. The function $\boldsymbol{f}: \mathcal{X} \rightarrow \mathbb{R}^m$ maps a design to its $m$-dimensional objective value vector.

In offline single-objective optimization (SOO), only one objective is considered (i.e., $m=1$). For instance, the design $\boldsymbol{x}$ might represent a neural network architecture, with $f(\boldsymbol{x})$ denoting the network's accuracy on a given dataset.

Offline multi-objective optimization (MOO) extends the framework to simultaneously address multiple objectives using the dataset $\mathcal{D}$. In this setting, the goal is to find solutions that balance competing objectives effectively. For instance, when designing a neural architecture, one might seek to achieve both high accuracy and high efficiency.

πŸ”— Benchmark

Task

Synthetic Function

Real-World System

Scientific Design

Machine Learning Model

Evaluation Metric

Usefulness

Diversity

Novelty

Stability

🎯 Surrogate Modeling

Auxiliary Loss

Data-Drive Adaptation

Collaborative Ensembling

Generative Model Integration

πŸ€” Generative Modeling

Variational Autoencoder (VAE)

Generative Adversarial Network (GAN)

Autoregressive Model

Diffusion Model

Flow Matching

Energy-Based Model

Control by Generative Flow Network (GFlowNet)

Releases

No releases published

Packages

No packages published