Skip to content

Commit

Permalink
rm pytorch from deps
Browse files Browse the repository at this point in the history
  • Loading branch information
mobicham committed Dec 20, 2023
1 parent bb422c9 commit 1a7c9cc
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 1 deletion.
2 changes: 2 additions & 0 deletions Readme.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@ This folder contains the code to perform Half-Quadratic Quantization (<b>HQQ</b>
<b>HQQ</b> is a fast and accurate model quantizer that skips the need for calibration data. It's super simple to implement (just a few lines of code for the optimizer). It can crunch through quantizing the Llama2-70B model in only 4 minutes! 🚀

### Installation
First, make sure you have a Pytorch 2 version that matches your CUDA version: https://pytorch.org/

You can install hqq via ```pip install hqq```.

To get the latest version, you can install the core library directly via ```pip install git+https://github.com/mobiusml/hqq.git```.
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,5 +9,5 @@
author_email='[email protected]',
license='Apache 2',
packages=['hqq', 'hqq/core', 'hqq/engine', 'hqq/models', 'hqq/models/hf', 'hqq/models/timm', 'hqq/models/vllm'],
install_requires=['numpy>=1.24.4','tqdm>=4.64.1', 'torch>=2.1.1', 'huggingface_hub', 'accelerate', 'timm', 'transformers>=4.36.1', 'termcolor'], #add vllm/langchain?
install_requires=['numpy>=1.24.4','tqdm>=4.64.1', 'huggingface_hub', 'accelerate', 'timm', 'transformers>=4.36.1', 'termcolor'], #'torch>=2.1.1', add vllm/langchain?
)

0 comments on commit 1a7c9cc

Please sign in to comment.