A library for persisting PyTorch program state
Project description
torchsnapshot
This library is currently in Alpha and currently does not have a stable release. The API may change and may not be backward compatible. If you have suggestions for improvements, please open a GitHub issue. We'd love to hear your feedback.
A light-weight library for adding fault tolerance to large-scale PyTorch distributed training workloads.
Install
Requires Python >= 3.7 and PyTorch >= 1.11
From pip:
pip install torchsnapshot
From source:
git clone https://github.com/facebookresearch/torchsnapshot
cd torchsnapshot
pip install -r requirements.txt
python setup.py install
Concepts
- Stateful object - an object that whose state can be obtained via
.state_dict()
and restored via.load_state_dict()
. Most PyTorch components (e.g.Module
,Optimizer
,LRScheduler
) already implement this protocol. - App state - the application state described using multiple stateful objects.
- Snapshot - the persisted app state.
Basic Usage
Describing the application state with multiple stateful objects:
app_state = {"model": model, "optimizer": optimizer}
Taking a snapshot of the application state:
from torchsnapshot import Snapshot
# File System
snapshot = Snapshot.take(path="/foo/bar/baz", app_state=app_state)
# S3
snapshot = Snapshot.take(path="s3://foo/bar", app_state=app_state)
# Google Cloud Storage
snapshot = Snapshot.take(path="gcs://foo/bar", app_state=app_state)
Referencing an existing snapshot:
snapshot = Snapshot(path="foo/bar/baz")
Restoring the application state from a snapshot:
snapshot.restore(app_state=app_state)
See the example directory for more examples.
License
torchsnapshot is BSD licensed, as found in the LICENSE file.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distributions
Hashes for torchsnapshot-nightly-2022.6.14.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6aa72a6a9f54722b9d04f530f92bd6dceb2ec9b0c0f46111a3ce628ded232ec9 |
|
MD5 | 9148f53935289bafa7f3cf34d8aba22a |
|
BLAKE2b-256 | 7e86849ed0d49a6d9f8afdf1aaa236139999dcb8ce35e6f3394284f1c95f4a1a |
Hashes for torchsnapshot_nightly-2022.6.14-py3.7.egg
Algorithm | Hash digest | |
---|---|---|
SHA256 | 72b391b35235367bf5dd11353343b99b1a46bfbc4095baee8855aba869ee53ad |
|
MD5 | 3d8eb8da9cca750022f3f5e6f7d9f6f8 |
|
BLAKE2b-256 | 97c28326e89f8f1362a7b279d95ad0e0a0f2ec8f80fa9601250a8c050ffdafc0 |
Hashes for torchsnapshot_nightly-2022.6.14-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4057caedf91311d91dfdc8ce1d3d14851f6f58322954081bc5a7f52293591f17 |
|
MD5 | 64bff6ac81845fd3e4d61b7a8102c614 |
|
BLAKE2b-256 | 4f791c5beb0cf1016bb9f4751f2ac4213051fa40708a07aafe7a128d6a79663c |