You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Why does Transformer Lens only support quantized LLaMA models?
Hi everyone,
I'm trying to use the transformer_lens library to study the activations of a quantized Mistral 7B model (unsloth/mistral-7b-instruct-v0.2-bnb-4bit). However, when I try to load it, I encounter a problem.
The problem is that I get an assertion error stating that only LLaMA models can be used in quantized format with this library. This is the error message I receive:
I find it illogical and frustrating that only LLaMA models are compatible with transformer_lens in quantized format. Can anyone explain why this decision was made? Is there a technical reason behind this or any way to work around this issue so that I can use my Mistral 7B model?
I appreciate any guidance or solutions you can provide.
Thanks!
The text was updated successfully, but these errors were encountered:
bryce13950
changed the title
[Question] Why does Transformer Lens only support quantized LLaMA models?
[Proposal] Expand quantization model support
Nov 3, 2024
Why does Transformer Lens only support quantized LLaMA models?
Hi everyone,
I'm trying to use the transformer_lens library to study the activations of a quantized Mistral 7B model (unsloth/mistral-7b-instruct-v0.2-bnb-4bit). However, when I try to load it, I encounter a problem.
This is the code I'm using:
The problem is that I get an assertion error stating that only LLaMA models can be used in quantized format with this library. This is the error message I receive:
I find it illogical and frustrating that only LLaMA models are compatible with transformer_lens in quantized format. Can anyone explain why this decision was made? Is there a technical reason behind this or any way to work around this issue so that I can use my Mistral 7B model?
I appreciate any guidance or solutions you can provide.
Thanks!
The text was updated successfully, but these errors were encountered: