By Shishir G. Patil, Paras Jain, Prabal Dutta, Ion Stoica, and Joseph E. Gonzalez (Project Website)
See the paper! https://arxiv.org/abs/2207.07697
POET
enables the training of state-of-the-art memory-hungry ML models on smartphones and other edge devices. POET (Private Optimal Energy Training) exploits the twin techniques of integrated tensor rematerialization, and paging-in/out of secondary storage (as detailed in our paper at ICML 2022) to optimize models for training with limited memory. POET's Mixed Integer Linear Formulation (MILP) ensures the solutions are provably optimal!
With POET, we are the first to demonstrate how to train memory-hungry SOTA ML models such as BERT and ResNets on smartphones and tiny ARM Cortex-M devices 💪
Reach out to us at sgp@berkeley.edu, if you have large models that you are trying to train - be it on GPUs, or your commodity edge devices such as laptops, smartphones, raspberry-pis, ARM Cortex M and A class, fitbits, etc.
Clone the repository and install POET:
git clone https://github.com/ShishirPatil/poet.git
cd poet/
pip install -e .
The ILP solver defaults to using the COIN-OR CBC solver when Gurobi isn't available. However, since Gurobi is much faster and presents solutions with tighter constraints, it is recommended to install it when possible.
If you are affiliated with an academic institution, you can acquire a free Gurobi Web License:
-
Create a free Gurobi account here. Make sure to specify the Academic user option.
-
Complete the rest of the Gurobi account creation process, which will include creating a password and verifying your email address.
-
Login to the Gurobi Web License Manager using your new account.
-
Create and download a new Web License file. It will be called
gurobi.lic
. -
Move the
gurobi.lic
file to your home directory (i.e. to~/gurobi.lic
on MacOS/Linux, orC:\Users\YOUR_USERNAME\gurobi.lic
on Windows).
Once you have installed POET and optionally configured Gurobi, you can run the solver via the command line. Here's an example:
python poet/solve.py --model resnet18_cifar --platform a72 --ram-budget 3000000 --runtime-budget 7.6
If you'd like to use the solver API directly, you can do so as follows:
from poet import solve
# ... use the solver API here
solve(
model="resnet18_cifar",
platform="m4",
ram_budget=737719,
runtime_budget=1.5,
time_limit_s=400,
solve_threads=16,
)
From our paper at ICML 2022:
In this work, we show that paging and rematerialization are highly complementary.
By carefully rematerializing cheap operations while paging results of expensive operations
to auxiliary memory such as a flash or an SD card, we can scale effective memory capacity
with minimal energy over- head. By combining these two methods, we demonstrate it is
possible to train models like BERT on mobile-class edge devices. By framing edge training
as an optimization problem, we discover optimal schedules with provable minimal energy
consumption at a given memory budget. While the focus of this paper is edge deployments,
the energy objective is increasingly becoming relevant even for cloud deployments!
If you use POET in your work, please cite us with:
@inproceedings{patil2022poet,
title={POET: Training Neural Networks on Tiny Devices with Integrated Rematerialization and Paging},
author={Patil, Shishir G and Jain, Paras and Dutta, Prabal and Stoica, Ion and Gonzalez, Joseph},
booktitle={International Conference on Machine Learning},
pages={17573--17583},
year={2022},
organization={PMLR}
}