Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RNNT training on CPU #95

Open
Peach-He opened this issue Jan 20, 2022 · 1 comment
Open

RNNT training on CPU #95

Peach-He opened this issue Jan 20, 2022 · 1 comment

Comments

@Peach-He
Copy link

Appreciate for the job on supporting RNN-T training on CPU (models/language_modeling/pytorch/rnnt/training/cpu), just quick evaluated the training code and found that WER would keep in 1.00 after even training 10+ epoches.
And I found this issue related on loss function used in training HawkAaron/warp-transducer#93
The grad in cpu is incorrect, is this a know issue? Or have we ever gotten the final WER of 0.058 rather than 1.0?

ashahba pushed a commit that referenced this issue Apr 1, 2022
* update TFX dockerfiles requests version

* Fix syntax
@sramakintel
Copy link
Contributor

sramakintel commented Mar 25, 2024

@Peach-He: The RNNT CPU training scripts have been updated recently. Can you try again to see if it resolves your issues? You can refer to latest optimizations here: https://www.intel.com/content/www/us/en/developer/articles/containers/cpu-reference-model-containers.html

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants