Skip to content
This repository has been archived by the owner on Jan 16, 2023. It is now read-only.

Latest commit

 

History

History
41 lines (24 loc) · 1.76 KB

BENCHMARKS.md

File metadata and controls

41 lines (24 loc) · 1.76 KB

Benchmark experiments

Summaries of the different benchmark experiments. To re-run experiments reported in our manuscript, open the corresponding Jupyter Notebook file for a step-by-step guide

These benchmarks were run with the package versions included in requirements.txt

Forecasting

See figure_forecasting_benchmarks_figures.ipynb for an overview

  • find_hyperparameters.py computes hyperparameters across all forecasting models separately for each dynamical system
  • compute_benchmarks_no_noise.py uses the best hyperparameters to train and score a models on the test data.
  • compute_benchmarks_noise_coarse.py and compute_benchmarks_noise_fine.py repeat the forecasting experiments in the presence of random noise.

Importance sampling

See figure_importance_sampling.ipynb for an overview

  • importance_sampling.py uses importance sampling to improve training on an LSTM forecasting model

Transfer learning

See figure_transfer_learning.ipynb for an overview

  • surrogate_transfer_learning.py computes the transfer learning benchmark on the UCR database
  • sweep_surrogate_transfer_learning.py recalculates the transfer learning results for different numbers of dynamical systems
  • sweep_surrogate_transfer_learning.py recalculates the transfer learning results for different numbers of dynamical systems
  • random_surrogate_transfer_learning.py calculates a baseline with random timescales
  • baseline_transfer_learning.py calculates a baseline on the raw UCR time series

Symbolic regression

See figure_symbolic_regression_benchmark.ipynb for an overview

  • symbolic_regression_benchmarks.py calculates all of the benchmarks

Neural ODE

See neural_ode_example.ipynb for an overview

  • node_benchmarks.py calculates all of the benchmarks