Skip to content

Commit

Permalink
Added doc for attention_head_size
Browse files Browse the repository at this point in the history
  • Loading branch information
pglorio authored Sep 27, 2024
1 parent 75554d8 commit a109b3f
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions src/transformers/models/zamba/configuration_zamba.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,8 @@ class ZambaConfig(PretrainedConfig):
model has a output word embedding layer.
hidden_size (`int`, *optional*, defaults to 3712):
Dimension of the hidden representations.
attention_hidden_size (`int`, *optional*, defaults to `None`):
Dimension of the hidden representations of the inputs to the Attention layer.
intermediate_size (`int`, *optional*, defaults to 14848):
Dimension of the MLP representations.
num_hidden_layers (`int`, *optional*, defaults to 76):
Expand Down

0 comments on commit a109b3f

Please sign in to comment.