Skip to content

Official repository for the paper "NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks". This repository contains the code for the experiments in the paper.

License

Notifications You must be signed in to change notification settings

BorealisAI/neuzip

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks

This is the official repository for the paper NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks. This repository contains the code for the experiments in the paper.

Installation

From PyPI

We provide a simple way to install the package from PyPI. You can install the package by running the following command:

pip install neuzip

Note that we only upload the source distribution to PyPI. You need to have NVCC correctly installed on your system to compile the package.

From source

You can also install the package from source, which is useful if you want to modify the code.

git clone https://github.com/BorealisAI/neuzip
cd neuzip
pip install -e .

Basic usage

Using neuzip for your PyTorch model is pretty easy. Here is a simple example:

model: torch.nn.Module = # your model
+ manager = neuzip.Manager()
+ model = manager.convert(model)

The compressed model can be used in the same way as the original model while consuming less memory.

Replicating experiments

You can replicate all the experiments in the paper by using the files in the examples/ directory. Each file corresponds to one or more experiments in the paper.

About

Official repository for the paper "NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks". This repository contains the code for the experiments in the paper.

Resources

License

Stars

Watchers

Forks