Skip to content

Running the F5-TTS by ONNX Runtime standalone with GUI

Notifications You must be signed in to change notification settings

patientx/F5-TTS-ONNX-gui

 
 

Repository files navigation

F5-TTS-ONNX-gui standalone & GPU accelerated

  • Have python & git installed.
  • Clone this repo with git clone https://github.com/patientx/F5-TTS-ONNX-gui

INSTALLATION

*** Automatic Installation ***

  • Run install.bat
  • Default english version of F5-TTS models are automatically downloaded from huggingface into /models/onnx/ folder.
  • If you can't reach huggingface you can use this link as a backup : https://drive.google.com/file/d/1_QNdX-6l8iwDF9c6HxOxW5bpCB6HJGnR/view

*** Manual installation ***

  • Create a virtual enviroment for the app. python -m venv venv
  • Activate the virtual enviroment. venv\scripts\activate
  • Install the necessary packages. pip install -r requirements.txt
  • Download necessary model files.
  • curl -s -L https://huggingface.co/SWivid/F5-TTS/resolve/main/F5TTS_Base/vocab.txt > models\vocab.txt
  • curl --progress-bar -L https://huggingface.co/huggingfacess/F5-TTS-ONNX/resolve/main/F5_Decode.onnx > models\onnx\F5_Decode.onnx
  • curl --progress-bar -L https://huggingface.co/huggingfacess/F5-TTS-ONNX/resolve/main/F5_Preprocess.onnx > models\onnx\F5_Preprocess.onnx
  • curl --progress-bar -L https://huggingface.co/huggingfacess/F5-TTS-ONNX/resolve/main/F5_Transformer.onnx > models\onnx\F5_Transformer.onnx

RUNNING

  • After you put those models , you can run the app by simply running inference.bat , there would be a simple GUI with default values.
  • App now saves the values everytime you change them.
  • If you want to manually run it, get into the folder via commandline , venv\scripts\activate , python F5-TTS-ONNX-Inference.py .
  • Make sure your sample audio is inside audio folder , change the first box to whatever you want the model to say , and the last box to write what your reference audio is saying.
  • Change "generated audio file" 's name if you don't want it overwritten

  • This fork is using onnxruntime to use your GPU. AMD (from 6000 onwards better) and Nvidia GPU's should work at an acceptable speed.

*** Here is an alternative ONNX model which accepts audio files twice longer than default.

  • https://www.mediafire.com/file/dkwdbfswqks414u/f5-tts-onnx-4096.zip/file
  • Download and extract them in the models/onnx/ folder. Change the names in the GUI accordingly. This model uses more vram (not sure the exact amount) but can generate a bit longer sentences.

About

Running the F5-TTS by ONNX Runtime standalone with GUI

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 97.5%
  • Batchfile 2.5%