Skip to content

Commit

Permalink
reformat HybridLayer __init__
Browse files Browse the repository at this point in the history
  • Loading branch information
pglorio committed Oct 4, 2024
1 parent 267530d commit 0a90fc7
Showing 1 changed file with 1 addition and 3 deletions.
4 changes: 1 addition & 3 deletions src/transformers/models/zamba/modeling_zamba.py
Original file line number Diff line number Diff line change
Expand Up @@ -978,9 +978,7 @@ def forward(


class HybridLayer(nn.Module):
def __init__(
self, shared_transf: ZambaAttentionDecoderLayer, linear: nn.Linear, mamba: ZambaMambaDecoderLayer
):
def __init__(self, shared_transf: ZambaAttentionDecoderLayer, linear: nn.Linear, mamba: ZambaMambaDecoderLayer):
super().__init__()
self.shared_transf = shared_transf
self.linear = linear
Expand Down

0 comments on commit 0a90fc7

Please sign in to comment.