-
Notifications
You must be signed in to change notification settings - Fork 27.4k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Add sidebar tutorial for chat models (#30401)
* Draft tutorial for talking to chat models * Reformat lists and text snippets * Cleanups and clarifications * Finish up remaining TODOs * Correct section link * Small fix * Add proper quantization examples * Add proper quantization examples * Add proper quantization examples * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/_toctree.yml Co-authored-by: Steven Liu <[email protected]> * Update docs/source/en/conversations.md Co-authored-by: Steven Liu <[email protected]> * Fix Text Generation Pipeline link and add a ref to the LLM inference guide * intelligent -> capable * Small intro cleanup * Small text cleanup * Small text cleanup * Clarification about system message * Clarification about system message --------- Co-authored-by: Steven Liu <[email protected]>
- Loading branch information
1 parent
bc274a2
commit f396271
Showing
2 changed files
with
292 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,290 @@ | ||
<!--Copyright 2024 The HuggingFace Team. All rights reserved. | ||
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with | ||
the License. You may obtain a copy of the License at | ||
http://www.apache.org/licenses/LICENSE-2.0 | ||
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on | ||
an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the | ||
specific language governing permissions and limitations under the License. | ||
⚠️ Note that this file is in Markdown but contain specific syntax for our doc-builder (similar to MDX) that may not be | ||
rendered properly in your Markdown viewer. | ||
--> | ||
|
||
# Chatting with Transformers | ||
|
||
If you're reading this article, you're almost certainly aware of **chat models**. Chat models are conversational | ||
AIs that you can send and receive messages with. The most famous of these is the proprietary ChatGPT, but there are | ||
now many open-source chat models which match or even substantially exceed its performance. These models are free to | ||
download and run on a local machine. Although the largest and most capable models require high-powered hardware | ||
and lots of memory to run, there are smaller models that will run perfectly well on a single consumer GPU, or even | ||
an ordinary desktop or notebook CPU. | ||
|
||
This guide will help you get started with chat models. We'll start with a brief quickstart guide that uses a convenient, | ||
high-level "pipeline". This is all you need if you just want to start running a chat model | ||
immediately. After the quickstart, we'll move on to more detailed information about | ||
what exactly chat models are, how to choose an appropriate one, and a low-level breakdown of each of the | ||
steps involved in talking to a chat model. We'll also give some tips on optimizing the performance and memory usage | ||
of your chat models. | ||
|
||
|
||
## Quickstart | ||
|
||
If you have no time for details, here's the brief summary: Chat models continue chats. This means that you pass them | ||
a conversation history, which can be as short as a single user message, and the model will continue the conversation | ||
by adding its response. Let's see this in action. First, let's build a chat: | ||
|
||
```python | ||
chat = [ | ||
{"role": "system", "content": "You are a sassy, wise-cracking robot as imagined by Hollywood circa 1986."}, | ||
{"role": "user", "content": "Hey, can you tell me any fun things to do in New York?"} | ||
] | ||
``` | ||
|
||
Notice that in addition to the user's message, we added a **system** message at the start of the conversation. Not all | ||
chat models support system messages, but when they do, they represent high-level directives about how the model | ||
should behave in the conversation. You can use this to guide the model - whether you want short or long responses, | ||
lighthearted or serious ones, and so on. If you want the model to do useful work instead of | ||
practicing its improv routine, you can either omit the system message or try a terse one such as "You are a helpful and intelligent | ||
AI assistant who responds to user queries." | ||
|
||
Once you have a chat, the quickest way to continue it is using the [`TextGenerationPipeline`]. | ||
Let's see this in action with `LLaMA-3`. Note that `LLaMA-3` is a gated model, which means you will need to | ||
[apply for access](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) and log in with your Hugging Face | ||
account to use it. We'll also use `device_map="auto"`, which will load the model on GPU if there's enough memory | ||
for it, and set the dtype to `torch.bfloat16` to save memory: | ||
|
||
```python | ||
import torch | ||
from transformers import pipeline | ||
|
||
pipe = pipeline("text-generation", "meta-llama/Meta-Llama-3-8B-Instruct", torch_dtype=torch.bfloat16, device_map="auto") | ||
response = pipe(chat, max_new_tokens=512) | ||
print(response[0]['generated_text'][-1]['content']) | ||
``` | ||
|
||
And you'll get: | ||
|
||
```text | ||
(sigh) Oh boy, you're asking me for advice? You're gonna need a map, pal! Alright, | ||
alright, I'll give you the lowdown. But don't say I didn't warn you, I'm a robot, not a tour guide! | ||
So, you wanna know what's fun to do in the Big Apple? Well, let me tell you, there's a million | ||
things to do, but I'll give you the highlights. First off, you gotta see the sights: the Statue of | ||
Liberty, Central Park, Times Square... you know, the usual tourist traps. But if you're lookin' for | ||
something a little more... unusual, I'd recommend checkin' out the Museum of Modern Art. It's got | ||
some wild stuff, like that Warhol guy's soup cans and all that jazz. | ||
And if you're feelin' adventurous, take a walk across the Brooklyn Bridge. Just watch out for | ||
those pesky pigeons, they're like little feathered thieves! (laughs) Get it? Thieves? Ah, never mind. | ||
Now, if you're lookin' for some serious fun, hit up the comedy clubs in Greenwich Village. You might | ||
even catch a glimpse of some up-and-coming comedians... or a bunch of wannabes tryin' to make it big. (winks) | ||
And finally, if you're feelin' like a real New Yorker, grab a slice of pizza from one of the many amazing | ||
pizzerias around the city. Just don't try to order a "robot-sized" slice, trust me, it won't end well. (laughs) | ||
So, there you have it, pal! That's my expert advice on what to do in New York. Now, if you'll | ||
excuse me, I've got some oil changes to attend to. (winks) | ||
``` | ||
|
||
You can continue the chat by appending your own response to it. The | ||
`response` object returned by the pipeline actually contains the entire chat so far, so we can simply append | ||
a message and pass it back: | ||
|
||
```python | ||
chat = response[0]['generated_text'] | ||
chat.append( | ||
{"role": "user", "content": "Wait, what's so wild about soup cans?"} | ||
) | ||
response = pipe(chat, max_new_tokens=512) | ||
print(response[0]['generated_text'][-1]['content']) | ||
``` | ||
|
||
And you'll get: | ||
|
||
```text | ||
(laughs) Oh, you're killin' me, pal! You don't get it, do you? Warhol's soup cans are like, art, man! | ||
It's like, he took something totally mundane, like a can of soup, and turned it into a masterpiece. It's | ||
like, "Hey, look at me, I'm a can of soup, but I'm also a work of art!" | ||
(sarcastically) Oh, yeah, real original, Andy. | ||
But, you know, back in the '60s, it was like, a big deal. People were all about challenging the | ||
status quo, and Warhol was like, the king of that. He took the ordinary and made it extraordinary. | ||
And, let me tell you, it was like, a real game-changer. I mean, who would've thought that a can of soup could be art? (laughs) | ||
But, hey, you're not alone, pal. I mean, I'm a robot, and even I don't get it. (winks) | ||
But, hey, that's what makes art, art, right? (laughs) | ||
``` | ||
|
||
The remainder of this tutorial will cover specific topics such | ||
as performance and memory, or how to select a chat model for your needs. | ||
|
||
## Choosing a chat model | ||
|
||
There are an enormous number of different chat models available on the [Hugging Face Hub](https://huggingface.co/models?pipeline_tag=text-generation&sort=trending), | ||
and new users often feel very overwhelmed by the selection offered. Don't be, though! You really need to just focus on | ||
two important considerations: | ||
- The model's size, which will determine if you can fit it in memory and how quickly it will | ||
run. | ||
- The quality of the model's chat output. | ||
|
||
In general, these are correlated - bigger models tend to be | ||
more capable, but even so there's a lot of variation at a given size point! | ||
|
||
### Size and model naming | ||
The size of a model is easy to spot - it's the number in the model name, like "8B" or "70B". This is the number of | ||
**parameters** in the model. Without quantization, you should expect to need about 2 bytes of memory per parameter. | ||
This means that an "8B" model with 8 billion parameters will need about 16GB of memory just to fit the parameters, | ||
plus a little extra for other overhead. It's a good fit for a high-end consumer GPU with 24GB of memory, such as a 3090 | ||
or 4090. | ||
|
||
Some chat models are "Mixture of Experts" models. These may list their sizes in different ways, such as "8x7B" or | ||
"141B-A35B". The numbers are a little fuzzier here, but in general you can read this as saying that the model | ||
has approximately 56 (8x7) billion parameters in the first case, or 141 billion parameters in the second case. | ||
|
||
Note that it is very common to use quantization techniques to reduce the memory usage per parameter to 8 bits, 4 bits, | ||
or even less. This topic is discussed in more detail in the [Memory considerations](#memory-considerations) section below. | ||
|
||
### But which chat model is best? | ||
Even once you know the size of chat model you can run, there's still a lot of choice out there. One way to sift through | ||
it all is to consult **leaderboards**. Two of the most popular leaderboards are the [OpenLLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) | ||
and the [LMSys Chatbot Arena Leaderboard](https://chat.lmsys.org/?leaderboard). Note that the LMSys leaderboard | ||
also includes proprietary models - look at the `licence` column to identify open-source ones that you can download, then | ||
search for them on the [Hugging Face Hub](https://huggingface.co/models?pipeline_tag=text-generation&sort=trending). | ||
|
||
### Specialist domains | ||
Some models may be specialized for certain domains, such as medical or legal text, or non-English languages. | ||
If you're working in these domains, you may find that a specialized model will give you big performance benefits. | ||
Don't automatically assume that, though! Particularly when specialized models are smaller or older than the current | ||
cutting-edge, a top-end general-purpose model may still outclass them. Thankfully, we are beginning to see | ||
[domain-specific leaderboards](https://huggingface.co/blog/leaderboard-medicalllm) that should make it easier to locate | ||
the best models for specialized domains. | ||
|
||
## What happens inside the pipeline? | ||
|
||
The quickstart above used a high-level pipeline to chat with a chat model, which is convenient, but not the | ||
most flexible. Let's take a more low-level approach, to see each of the steps involved in chat. Let's start with | ||
a code sample, and then break it down: | ||
|
||
```python | ||
from transformers import AutoModelForCausalLM, AutoTokenizer | ||
import torch | ||
|
||
# Prepare the input as before | ||
chat = [ | ||
{"role": "system", "content": "You are a sassy, wise-cracking robot as imagined by Hollywood circa 1986."}, | ||
{"role": "user", "content": "Hey, can you tell me any fun things to do in New York?"} | ||
] | ||
|
||
# 1: Load the model and tokenizer | ||
model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct", device_map="auto", torch_dtype=torch.bfloat16) | ||
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct") | ||
|
||
# 2: Apply the chat template | ||
formatted_chat = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=True) | ||
print("Formatted chat:\n", formatted_chat) | ||
|
||
# 3: Tokenize the chat (This can be combined with the previous step using tokenize=True) | ||
inputs = tokenizer(formatted_chat, return_tensors="pt", add_special_tokens=False) | ||
# Move the tokenized inputs to the same device the model is on (GPU/CPU) | ||
inputs = {key: tensor.to(model.device) for key, tensor in inputs.items()} | ||
print("Tokenized inputs:\n", inputs) | ||
|
||
# 4: Generate text from the model | ||
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.) | ||
print("Generated tokens:\n", outputs) | ||
|
||
# 5: Decode the output back to a string | ||
decoded_output = tokenizer.decode(outputs[0][inputs['input_ids'].size(1):], skip_special_tokens=True) | ||
print("Decoded output:\n", decoded_output) | ||
``` | ||
|
||
There's a lot in here, each piece of which could be its own document! Rather than going into too much detail, I'll cover | ||
the broad ideas, and leave the details for the linked documents. The key steps are: | ||
|
||
1. [Models](https://huggingface.co/learn/nlp-course/en/chapter2/3) and [Tokenizers](https://huggingface.co/learn/nlp-course/en/chapter2/4?fw=pt) are loaded from the Hugging Face Hub. | ||
2. The chat is formatted using the tokenizer's [chat template](https://huggingface.co/docs/transformers/main/en/chat_templating) | ||
3. The formatted chat is [tokenized](https://huggingface.co/learn/nlp-course/en/chapter2/4) using the tokenizer. | ||
4. We [generate](https://huggingface.co/docs/transformers/en/llm_tutorial) a response from the model. | ||
5. The tokens output by the model are decoded back to a string | ||
|
||
## Performance, memory and hardware | ||
|
||
You probably know by now that most machine learning tasks are run on GPUs. However, it is entirely possible | ||
to generate text from a chat model or language model on a CPU, albeit somewhat more slowly. If you can fit | ||
the model in GPU memory, though, this will usually be the preferable option. | ||
|
||
### Memory considerations | ||
|
||
By default, Hugging Face classes like [`TextGenerationPipeline`] or [`AutoModelForCausalLM`] will load the model in | ||
`float32` precision. This means that it will need 4 bytes (32 bits) per parameter, so an "8B" model with 8 billion | ||
parameters will need ~32GB of memory. However, this can be wasteful! Most modern language models are trained in | ||
"bfloat16" precision, which uses only 2 bytes per parameter. If your hardware supports it (Nvidia 30xx/Axxx | ||
or newer), you can load the model in `bfloat16` precision, using the `torch_dtype` argument as we did above. | ||
|
||
It is possible to go even lower than 16-bits using "quantization", a method to lossily compress model weights. This | ||
allows each parameter to be squeezed down to 8 bits, 4 bits or even less. Note that, especially at 4 bits, | ||
the model's outputs may be negatively affected, but often this is a tradeoff worth making to fit a larger and more | ||
capable chat model in memory. Let's see this in action with `bitsandbytes`: | ||
|
||
```python | ||
from transformers import AutoModelForCausalLM, BitsAndBytesConfig | ||
|
||
quantization_config = BitsAndBytesConfig(load_in_8bit=True) # You can also try load_in_4bit | ||
model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct", device_map="auto", quantization_config=quantization_config) | ||
``` | ||
|
||
Or we can do the same thing using the `pipeline` API: | ||
|
||
```python | ||
from transformers import pipeline, BitsAndBytesConfig | ||
|
||
quantization_config = BitsAndBytesConfig(load_in_8bit=True) # You can also try load_in_4bit | ||
pipe = pipeline("text-generation", "meta-llama/Meta-Llama-3-8B-Instruct", device_map="auto", model_kwargs={"quantization_config": quantization_config}) | ||
``` | ||
|
||
There are several other options for quantizing models besides `bitsandbytes` - please see the [Quantization guide](./quantization) | ||
for more information. | ||
|
||
### Performance considerations | ||
|
||
<Tip> | ||
|
||
For a more extensive guide on language model performance and optimization, check out [LLM Inference Optimization](./llm_optims) . | ||
|
||
</Tip> | ||
|
||
|
||
As a general rule, larger chat models will be slower in addition to requiring more memory. It's possible to be | ||
more concrete about this, though: Generating text from a chat model is unusual in that it is bottlenecked by | ||
**memory bandwidth** rather than compute power, because every active parameter must be read from memory for each | ||
token that the model generates. This means that number of tokens per second you can generate from a chat | ||
model is generally proportional to the total bandwidth of the memory it resides in, divided by the size of the model. | ||
|
||
In our quickstart example above, our model was ~16GB in size when loaded in `bfloat16` precision. | ||
This means that 16GB must be read from memory for every token generated by the model. Total memory bandwidth can | ||
vary from 20-100GB/sec for consumer CPUs to 200-900GB/sec for consumer GPUs, specialized CPUs like | ||
Intel Xeon, AMD Threadripper/Epyc or high-end Apple silicon, and finally up to 2-3TB/sec for data center GPUs like | ||
the Nvidia A100 or H100. This should give you a good idea of the generation speed you can expect from these different | ||
hardware types. | ||
|
||
Therefore, if you want to improve the speed of text generation, the easiest solution is to either reduce the | ||
size of the model in memory (usually by quantization), or get hardware with higher memory bandwidth. For advanced users, | ||
several other techniques exist to get around this bandwidth bottleneck. The most common are variants on | ||
[assisted generation](https://huggingface.co/blog/assisted-generation), also known as "speculative | ||
sampling". These techniques try to guess multiple future tokens at once, often using a smaller "draft model", and then | ||
confirm these generations with the chat model. If the guesses are validated by the chat model, more than one token can | ||
be generated per forward pass, which greatly alleviates the bandwidth bottleneck and improves generation speed. | ||
|
||
Finally, we should also note the impact of "Mixture of Experts" (MoE) models here. Several popular chat models, | ||
such as Mixtral, Qwen-MoE and DBRX, are MoE models. In these models, not every parameter is active for every token generated. | ||
As a result, MoE models generally have much lower memory bandwidth requirements, even though their total size | ||
can be quite large. They can therefore be several times faster than a normal "dense" model of the same size. However, | ||
techniques like assisted generation are generally ineffective for these models because more parameters will become | ||
active with each new speculated token, which will negate the bandwidth and speed benefits that the MoE architecture | ||
provides. | ||
|