diff --git a/_posts/2024-01-01-owsm.md b/_posts/2024-01-01-owsm.md new file mode 100644 index 00000000..688e6563 --- /dev/null +++ b/_posts/2024-01-01-owsm.md @@ -0,0 +1,97 @@ +--- +layout: post +title: Open Whisper-style Speech Models (OWSM) +description: This is the project page for OWSM models. +date: 2024-01-01 00:00:00-0800 +comments: false +--- + +## Overview + +The **O**pen **W**hisper-style **S**peech **M**odels (OWSM, pronounced as "awesome") are a series of speech foundation models developed by [WAVLab](https://www.wavlab.org/) at Carnegie Mellon University. It reproduces Whisper-style training using publicly available data and an open-source toolkit [ESPnet](https://github.com/espnet/espnet). By publicly releasing data preparation scripts, training and inference code, pre-trained model weights and training logs, we aim to promote transparency and open science in large-scale speech pre-training. + +## News + + +## Demo pages + +- Gradio demo: [![Static Badge](https://img.shields.io/badge/OWSM-Demo-orange)](https://pyf98-owsm-v3-demo.hf.space) +- Colab notebook: [![Open All Collab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1zKI3ZY_OtZd6YmVeED6Cxy1QwT1mqv9O?usp=sharing) + + +## Pre-trained models + +We have released various pre-trained models. The training logs are also available for major models. + +
Name | +Encoder | +Parameters | +Data (hours) | +Model Link | +ESPnet Recipe | +
---|---|---|---|---|---|
OWSM v1 | +Transformer | +272M | +38k | +espnet/owsm_v1 | +egs2/owsm_v1/s2t1 | +
OWSM v2 | +Transformer | +712M | +129k | +espnet/owsm_v2 | +egs2/owsm_v2/s2t1 | +
OWSM v2 | +E-Branchformer | +739M | +129k | +espnet/owsm_v2_ebranchformer | +egs2/owsm_v2/s2t1 | +
OWSM v3 | +Transformer | +889M | +180k | +espnet/owsm_v3 | +egs2/owsm_v3/s2t1 | +
OWSM v3.1 | +E-Branchformer | +1.02B | +180k | +espnet/owsm_v3.1_ebf | +TBD | +