Skip to content

Commit

Permalink
remove fused_layernorm with apex dependency and include low_precision…
Browse files Browse the repository at this point in the history
…_layernorm
  • Loading branch information
jacobfulano committed Jan 8, 2024
1 parent cb78d90 commit c44effa
Show file tree
Hide file tree
Showing 5 changed files with 5 additions and 5 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ parameters:

# Algorithms
algorithms:
fused_layernorm: {}
low_precision_layernorm: {}

# Task configuration
tasks:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ scheduler:

# Algorithms
algorithms:
fused_layernorm: {}
low_precision_layernorm: {}

# Task configuration
tasks:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,7 @@ parameters:
weight_decay: 1.0e-5 # Amount of weight decay regularization

algorithms:
fused_layernorm: {}
low_precision_layernorm: {}

max_duration: 286720000sp # Subsample the training data for ~275M samples
eval_interval: 2000ba
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,7 @@ parameters:
weight_decay: 1.0e-5 # Amount of weight decay regularization

algorithms:
fused_layernorm: {}
low_precision_layernorm: {}

max_duration: 286720000sp # Subsample the training data for ~275M samples
eval_interval: 2000ba
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ optimizer:
weight_decay: 1.0e-5 # Amount of weight decay regularization

algorithms:
fused_layernorm: {}
low_precision_layernorm: {}

max_duration: 286720000sp # Subsample the training data for ~275M samples
eval_interval: 2000ba
Expand Down

0 comments on commit c44effa

Please sign in to comment.