diff --git a/.github/workflows/pypi-release.yml b/.github/workflows/pypi-release.yml index 2bbcf3cd70..f81f5a7493 100644 --- a/.github/workflows/pypi-release.yml +++ b/.github/workflows/pypi-release.yml @@ -10,7 +10,7 @@ jobs: build-sdist: runs-on: ubuntu-20.04 steps: - - uses: actions/checkout@v3 + - uses: actions/checkout@v4 - name: Verify tag matches version run: | set -ex @@ -19,7 +19,7 @@ jobs: if [[ "v$version" != "$tag" ]]; then exit 1 fi - - uses: actions/setup-python@v2 + - uses: actions/setup-python@v5 with: python-version: 3.9 - run: | @@ -28,7 +28,7 @@ jobs: python -m build - run: | pip install dist/*.tar.gz - - uses: actions/upload-artifact@v2 + - uses: actions/upload-artifact@v4 with: name: sdist path: dist/*.tar.gz @@ -38,8 +38,8 @@ jobs: matrix: python-version: ["3.9", "3.10", "3.11"] steps: - - uses: actions/checkout@v3 - - uses: actions/setup-python@v2 + - uses: actions/checkout@v4 + - uses: actions/setup-python@v5 with: python-version: ${{ matrix.python-version }} - name: Install pip requirements @@ -50,45 +50,38 @@ jobs: run: | python setup.py bdist_wheel --plat-name=manylinux1_x86_64 python -m pip install dist/*-manylinux*.whl - - uses: actions/upload-artifact@v2 + - uses: actions/upload-artifact@v4 with: name: wheel-${{ matrix.python-version }} path: dist/*-manylinux*.whl publish-artifacts: runs-on: ubuntu-20.04 needs: [build-sdist, build-wheels] + environment: + name: release + url: https://pypi.org/p/coqui-tts + permissions: + id-token: write steps: - run: | mkdir dist - - uses: actions/download-artifact@v2 + - uses: actions/download-artifact@v4 with: name: "sdist" path: "dist/" - - uses: actions/download-artifact@v2 + - uses: actions/download-artifact@v4 with: name: "wheel-3.9" path: "dist/" - - uses: actions/download-artifact@v2 + - uses: actions/download-artifact@v4 with: name: "wheel-3.10" path: "dist/" - - uses: actions/download-artifact@v2 + - uses: actions/download-artifact@v4 with: name: "wheel-3.11" path: "dist/" - run: | ls -lh dist/ - - name: Setup PyPI config - run: | - cat << EOF > ~/.pypirc - [pypi] - username=__token__ - password=${{ secrets.PYPI_TOKEN }} - EOF - - uses: actions/setup-python@v2 - with: - python-version: 3.9 - - run: | - python -m pip install twine - - run: | - twine upload --repository pypi dist/* + - name: Publish package distributions to PyPI + uses: pypa/gh-action-pypi-publish@release/v1 diff --git a/.github/workflows/tests.yml b/.github/workflows/tests.yml index 20308aab6c..b056e3073d 100644 --- a/.github/workflows/tests.yml +++ b/.github/workflows/tests.yml @@ -38,14 +38,14 @@ jobs: sudo apt-get install -y --no-install-recommends git make gcc make system-deps - name: Install/upgrade Python setup deps - run: python3 -m pip install --upgrade pip setuptools wheel uv + run: python3 -m pip install --upgrade pip setuptools wheel - name: Replace scarf urls if: contains(fromJSON('["data_tests", "inference_tests", "test_aux", "test_tts", "test_tts2", "test_xtts", "test_zoo0", "test_zoo1", "test_zoo2"]'), matrix.subset) run: | sed -i 's/https:\/\/coqui.gateway.scarf.sh\//https:\/\/github.com\/coqui-ai\/TTS\/releases\/download\//g' TTS/.models.json - name: Install TTS run: | - python3 -m uv pip install --system "TTS[all] @ ." + python3 -m pip install .[all] python3 setup.py egg_info - name: Unit tests run: make ${{ matrix.subset }} diff --git a/CITATION.cff b/CITATION.cff index 6b0c8f19af..28eb65e23c 100644 --- a/CITATION.cff +++ b/CITATION.cff @@ -10,8 +10,8 @@ authors: version: 1.4 doi: 10.5281/zenodo.6334862 license: "MPL-2.0" -url: "https://www.coqui.ai" -repository-code: "https://github.com/coqui-ai/TTS" +url: "https://github.com/eginhard/coqui-tts" +repository-code: "https://github.com/eginhard/coqui-tts" keywords: - machine learning - deep learning diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index a83b8c8296..8a0fe3904a 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -2,7 +2,7 @@ Welcome to the 🐸TTS! -This repository is governed by [the Contributor Covenant Code of Conduct](https://github.com/coqui-ai/TTS/blob/main/CODE_OF_CONDUCT.md). +This repository is governed by [the Contributor Covenant Code of Conduct](https://github.com/eginhard/coqui-tts/blob/main/CODE_OF_CONDUCT.md). ## Where to start. We welcome everyone who likes to contribute to 🐸TTS. @@ -15,13 +15,13 @@ If you like to contribute code, squash a bug but if you don't know where to star You can pick something out of our road map. We keep the progess of the project in this simple issue thread. It has new model proposals or developmental updates etc. -- [Github Issues Tracker](https://github.com/coqui-ai/TTS/issues) +- [Github Issues Tracker](https://github.com/eginhard/coqui-tts/issues) This is a place to find feature requests, bugs. Issues with the ```good first issue``` tag are good place for beginners to take on. -- ✨**PR**✨ [pages](https://github.com/coqui-ai/TTS/pulls) with the ```🚀new version``` tag. +- ✨**PR**✨ [pages](https://github.com/eginhard/coqui-tts/pulls) with the ```🚀new version``` tag. We list all the target improvements for the next version. You can pick one of them and start contributing. @@ -46,14 +46,14 @@ Let us know if you encounter a problem along the way. The following steps are tested on an Ubuntu system. -1. Fork 🐸TTS[https://github.com/coqui-ai/TTS] by clicking the fork button at the top right corner of the project page. +1. Fork 🐸TTS[https://github.com/eginhard/coqui-tts] by clicking the fork button at the top right corner of the project page. 2. Clone 🐸TTS and add the main repo as a new remote named ```upstream```. ```bash - $ git clone git@github.com:/TTS.git - $ cd TTS - $ git remote add upstream https://github.com/coqui-ai/TTS.git + $ git clone git@github.com:/coqui-tts.git + $ cd coqui-tts + $ git remote add upstream https://github.com/eginhard/coqui-tts.git ``` 3. Install 🐸TTS for development. @@ -105,7 +105,7 @@ The following steps are tested on an Ubuntu system. ```bash $ git fetch upstream - $ git rebase upstream/master + $ git rebase upstream/main # or for the development version $ git rebase upstream/dev ``` @@ -124,7 +124,7 @@ The following steps are tested on an Ubuntu system. 13. Let's discuss until it is perfect. 💪 - We might ask you for certain changes that would appear in the ✨**PR**✨'s page under 🐸TTS[https://github.com/coqui-ai/TTS/pulls]. + We might ask you for certain changes that would appear in the ✨**PR**✨'s page under 🐸TTS[https://github.com/eginhard/coqui-tts/pulls]. 14. Once things look perfect, We merge it to the ```dev``` branch and make it ready for the next version. @@ -132,14 +132,14 @@ The following steps are tested on an Ubuntu system. If you prefer working within a Docker container as your development environment, you can do the following: -1. Fork 🐸TTS[https://github.com/coqui-ai/TTS] by clicking the fork button at the top right corner of the project page. +1. Fork 🐸TTS[https://github.com/eginhard/coqui-tts] by clicking the fork button at the top right corner of the project page. 2. Clone 🐸TTS and add the main repo as a new remote named ```upsteam```. ```bash - $ git clone git@github.com:/TTS.git - $ cd TTS - $ git remote add upstream https://github.com/coqui-ai/TTS.git + $ git clone git@github.com:/coqui-tts.git + $ cd coqui-tts + $ git remote add upstream https://github.com/eginhard/coqui-tts.git ``` 3. Build the Docker Image as your development environment (it installs all of the dependencies for you): diff --git a/README.md b/README.md index 11b1ddb701..782b48ab69 100644 --- a/README.md +++ b/README.md @@ -26,7 +26,7 @@ ______________________________________________________________________ [![Discord](https://img.shields.io/discord/1037326658807533628?color=%239B59B6&label=chat%20on%20discord)](https://discord.gg/5eXr5seRrv) [![License]()](https://opensource.org/licenses/MPL-2.0) [![PyPI version](https://badge.fury.io/py/TTS.svg)](https://badge.fury.io/py/TTS) -[![Covenant](https://camo.githubusercontent.com/7d620efaa3eac1c5b060ece5d6aacfcc8b81a74a04d05cd0398689c01c4463bb/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f436f6e7472696275746f72253230436f76656e616e742d76322e3025323061646f707465642d6666363962342e737667)](https://github.com/eginhard/coqui-tts/blob/master/CODE_OF_CONDUCT.md) +[![Covenant](https://camo.githubusercontent.com/7d620efaa3eac1c5b060ece5d6aacfcc8b81a74a04d05cd0398689c01c4463bb/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f436f6e7472696275746f72253230436f76656e616e742d76322e3025323061646f707465642d6666363962342e737667)](https://github.com/eginhard/coqui-tts/blob/main/CODE_OF_CONDUCT.md) [![Downloads](https://pepy.tech/badge/tts)](https://pepy.tech/project/tts) [![DOI](https://zenodo.org/badge/265612440.svg)](https://zenodo.org/badge/latestdoi/265612440) @@ -62,7 +62,7 @@ Please use our dedicated channels for questions and discussion. Help is much mor | 💾 **Installation** | [TTS/README.md](https://github.com/eginhard/coqui-tts/tree/dev#installation)| | 👩‍💻 **Contributing** | [CONTRIBUTING.md](https://github.com/eginhard/coqui-tts/blob/main/CONTRIBUTING.md)| | 📌 **Road Map** | [Main Development Plans](https://github.com/coqui-ai/TTS/issues/378) -| 🚀 **Released Models** | [TTS Releases](https://github.com/coqui-ai/TTS/releases) and [Experimental Models](https://github.com/coqui-ai/TTS/wiki/Experimental-Released-Models)| +| 🚀 **Released Models** | [Standard models](https://github.com/eginhard/coqui-tts/blob/dev/TTS/.models.json) and [Fairseq models in ~1100 languages](https://github.com/eginhard/coqui-tts#example-text-to-speech-using-fairseq-models-in-1100-languages-)| | 📰 **Papers** | [TTS Papers](https://github.com/erogol/TTS-papers)| ## Features @@ -135,7 +135,7 @@ You can also help us implement more models. If you are only interested in [synthesizing speech](https://coqui-tts.readthedocs.io/en/latest/inference.html) with the released 🐸TTS models, installing from PyPI is the easiest option. ```bash -pip install TTS +pip install coqui-tts ``` If you plan to code or train models, clone 🐸TTS and install it locally. @@ -152,7 +152,9 @@ $ make system-deps # intended to be used on Ubuntu (Debian). Let us know if you $ make install ``` -If you are on Windows, 👑@GuyPaddock wrote installation instructions [here](https://stackoverflow.com/questions/66726331/how-can-i-run-mozilla-tts-coqui-tts-training-with-cuda-on-a-windows-system). +If you are on Windows, 👑@GuyPaddock wrote installation instructions +[here](https://stackoverflow.com/questions/66726331/how-can-i-run-mozilla-tts-coqui-tts-training-with-cuda-on-a-windows-system) +(note that these are out of date, e.g. you need to have at least Python 3.9). ## Docker Image diff --git a/TTS/VERSION b/TTS/VERSION index 2157409059..a723ece79b 100644 --- a/TTS/VERSION +++ b/TTS/VERSION @@ -1 +1 @@ -0.22.0 +0.22.1 diff --git a/TTS/server/README.md b/TTS/server/README.md index 9536e0d55a..f5df08011b 100644 --- a/TTS/server/README.md +++ b/TTS/server/README.md @@ -1,8 +1,8 @@ # :frog: TTS demo server Before you use the server, make sure you -[install](https://github.com/coqui-ai/TTS/tree/dev#install-tts)) :frog: TTS +[install](https://github.com/eginhard/coqui-tts/tree/dev#install-tts)) :frog: TTS properly and install the additional dependencies with `pip install -TTS[server]`. Then, you can follow the steps below. +coqui-tts[server]`. Then, you can follow the steps below. **Note:** If you install :frog:TTS using ```pip```, you can also use the ```tts-server``` end point on the terminal. diff --git a/TTS/server/server.py b/TTS/server/server.py index d117494060..01bd79a137 100644 --- a/TTS/server/server.py +++ b/TTS/server/server.py @@ -12,7 +12,7 @@ try: from flask import Flask, render_template, render_template_string, request, send_file except ImportError as e: - raise ImportError("Server requires requires flask, use `pip install TTS[server]`.") from e + raise ImportError("Server requires requires flask, use `pip install coqui-tts[server]`.") from e from TTS.config import load_config from TTS.utils.manage import ModelManager diff --git a/TTS/server/templates/index.html b/TTS/server/templates/index.html index 6354d3919d..f5f547c7bf 100644 --- a/TTS/server/templates/index.html +++ b/TTS/server/templates/index.html @@ -30,7 +30,7 @@ - Fork me on GitHub @@ -151,4 +151,4 @@ - \ No newline at end of file + diff --git a/TTS/tts/models/xtts.py b/TTS/tts/models/xtts.py index e42288fe83..1c73c42ce9 100644 --- a/TTS/tts/models/xtts.py +++ b/TTS/tts/models/xtts.py @@ -695,12 +695,12 @@ def inference_stream( def forward(self): raise NotImplementedError( - "XTTS has a dedicated trainer, please check the XTTS docs: https://tts.readthedocs.io/en/dev/models/xtts.html#training" + "XTTS has a dedicated trainer, please check the XTTS docs: https://coqui-tts.readthedocs.io/en/dev/models/xtts.html#training" ) def eval_step(self): raise NotImplementedError( - "XTTS has a dedicated trainer, please check the XTTS docs: https://tts.readthedocs.io/en/dev/models/xtts.html#training" + "XTTS has a dedicated trainer, please check the XTTS docs: https://coqui-tts.readthedocs.io/en/dev/models/xtts.html#training" ) @staticmethod @@ -789,5 +789,5 @@ def load_checkpoint( def train_step(self): raise NotImplementedError( - "XTTS has a dedicated trainer, please check the XTTS docs: https://tts.readthedocs.io/en/dev/models/xtts.html#training" + "XTTS has a dedicated trainer, please check the XTTS docs: https://coqui-tts.readthedocs.io/en/dev/models/xtts.html#training" ) diff --git a/TTS/tts/utils/text/phonemizers/__init__.py b/TTS/tts/utils/text/phonemizers/__init__.py index 744ccb3e70..446f288302 100644 --- a/TTS/tts/utils/text/phonemizers/__init__.py +++ b/TTS/tts/utils/text/phonemizers/__init__.py @@ -63,7 +63,7 @@ def get_phonemizer_by_name(name: str, **kwargs) -> BasePhonemizer: return ZH_CN_Phonemizer(**kwargs) if name == "ja_jp_phonemizer": if JA_JP_Phonemizer is None: - raise ValueError(" ❗ You need to install JA phonemizer dependencies. Try `pip install TTS[ja]`.") + raise ValueError(" ❗ You need to install JA phonemizer dependencies. Try `pip install coqui-tts[ja]`.") return JA_JP_Phonemizer(**kwargs) if name == "ko_kr_phonemizer": return KO_KR_Phonemizer(**kwargs) diff --git a/docs/source/faq.md b/docs/source/faq.md index fa48c4a9fb..14be9d4c9c 100644 --- a/docs/source/faq.md +++ b/docs/source/faq.md @@ -3,7 +3,7 @@ We tried to collect common issues and questions we receive about 🐸TTS. It is ## Errors with a pre-trained model. How can I resolve this? - Make sure you use the right commit version of 🐸TTS. Each pre-trained model has its corresponding version that needs to be used. It is defined on the model table. -- If it is still problematic, post your problem on [Discussions](https://github.com/coqui-ai/TTS/discussions). Please give as many details as possible (error message, your TTS version, your TTS model and config.json etc.) +- If it is still problematic, post your problem on [Discussions](https://github.com/eginhard/coqui-tts/discussions). Please give as many details as possible (error message, your TTS version, your TTS model and config.json etc.) - If you feel like it's a bug to be fixed, then prefer Github issues with the same level of scrutiny. ## What are the requirements of a good 🐸TTS dataset? @@ -16,7 +16,7 @@ We tried to collect common issues and questions we receive about 🐸TTS. It is - If you need faster models, consider SpeedySpeech, GlowTTS or AlignTTS. Keep in mind that SpeedySpeech requires a pre-trained Tacotron or Tacotron2 model to compute text-to-speech alignments. ## How can I train my own `tts` model? -0. Check your dataset with notebooks in [dataset_analysis](https://github.com/coqui-ai/TTS/tree/master/notebooks/dataset_analysis) folder. Use [this notebook](https://github.com/coqui-ai/TTS/blob/master/notebooks/dataset_analysis/CheckSpectrograms.ipynb) to find the right audio processing parameters. A better set of parameters results in a better audio synthesis. +0. Check your dataset with notebooks in [dataset_analysis](https://github.com/eginhard/coqui-tts/tree/main/notebooks/dataset_analysis) folder. Use [this notebook](https://github.com/eginhard/coqui-tts/blob/main/notebooks/dataset_analysis/CheckSpectrograms.ipynb) to find the right audio processing parameters. A better set of parameters results in a better audio synthesis. 1. Write your own dataset `formatter` in `datasets/formatters.py` or format your dataset as one of the supported datasets, like LJSpeech. A `formatter` parses the metadata file and converts a list of training samples. diff --git a/docs/source/inference.md b/docs/source/inference.md index 2c57f6182c..0b05965f46 100644 --- a/docs/source/inference.md +++ b/docs/source/inference.md @@ -4,7 +4,7 @@ First, you need to install TTS. We recommend using PyPi. You need to call the command below: ```bash -$ pip install TTS +$ pip install coqui-tts ``` After the installation, 2 terminal commands are available. @@ -14,7 +14,7 @@ After the installation, 2 terminal commands are available. 3. In 🐍Python. - `from TTS.api import TTS` ## On the Commandline - `tts` -![cli.gif](https://github.com/coqui-ai/TTS/raw/main/images/tts_cli.gif) +![cli.gif](https://github.com/eginhard/coqui-tts/raw/main/images/tts_cli.gif) After the installation, 🐸TTS provides a CLI interface for synthesizing speech using pre-trained models. You can either use your own model or the release models under 🐸TTS. @@ -81,11 +81,11 @@ tts --model_name "voice_conversion///" ## On the Demo Server - `tts-server` - -![server.gif](https://github.com/coqui-ai/TTS/raw/main/images/demo_server.gif) + +![server.gif](https://github.com/eginhard/coqui-tts/raw/main/images/demo_server.gif) You can boot up a demo 🐸TTS server to run an inference with your models (make -sure to install the additional dependencies with `pip install TTS[server]`). +sure to install the additional dependencies with `pip install coqui-tts[server]`). Note that the server is not optimized for performance but gives you an easy way to interact with the models. diff --git a/docs/source/installation.md b/docs/source/installation.md index 8aaec01c9e..92743a9db4 100644 --- a/docs/source/installation.md +++ b/docs/source/installation.md @@ -1,6 +1,6 @@ # Installation -🐸TTS supports python >=3.9 <3.12.0 and tested on Ubuntu 18.10, 19.10, 20.10. +🐸TTS supports python >=3.9 <3.12.0 and was tested on Ubuntu 20.04 and 22.04. ## Using `pip` @@ -9,13 +9,13 @@ You can install from PyPI as follows: ```bash -pip install TTS # from PyPI +pip install coqui-tts # from PyPI ``` Or install from Github: ```bash -pip install git+https://github.com/coqui-ai/TTS # from Github +pip install git+https://github.com/eginhard/coqui-tts # from Github ``` ## Installing From Source @@ -23,11 +23,13 @@ pip install git+https://github.com/coqui-ai/TTS # from Github This is recommended for development and more control over 🐸TTS. ```bash -git clone https://github.com/coqui-ai/TTS/ -cd TTS +git clone https://github.com/eginhard/coqui-tts +cd coqui-tts make system-deps # only on Linux systems. make install ``` ## On Windows -If you are on Windows, 👑@GuyPaddock wrote installation instructions [here](https://stackoverflow.com/questions/66726331/ +If you are on Windows, 👑@GuyPaddock wrote installation instructions +[here](https://stackoverflow.com/questions/66726331/) (note that these are out +of date, e.g. you need to have at least Python 3.9) diff --git a/docs/source/main_classes/trainer_api.md b/docs/source/main_classes/trainer_api.md index 876e09e5b6..335294aa4d 100644 --- a/docs/source/main_classes/trainer_api.md +++ b/docs/source/main_classes/trainer_api.md @@ -1,3 +1,3 @@ # Trainer API -We made the trainer a separate project on https://github.com/coqui-ai/Trainer +We made the trainer a separate project on https://github.com/eginhard/coqui-trainer diff --git a/docs/source/models/xtts.md b/docs/source/models/xtts.md index b979d04f6e..014b161669 100644 --- a/docs/source/models/xtts.md +++ b/docs/source/models/xtts.md @@ -3,9 +3,6 @@ ⓍTTS has important model changes that make cross-language voice cloning and multi-lingual speech generation super easy. There is no need for an excessive amount of training data that spans countless hours. -This is the same model that powers [Coqui Studio](https://coqui.ai/), and [Coqui API](https://docs.coqui.ai/docs), however we apply -a few tricks to make it faster and support streaming inference. - ### Features - Voice cloning. - Cross-language voice cloning. @@ -32,21 +29,20 @@ Stay tuned as we continue to add support for more languages. If you have any lan This model is licensed under [Coqui Public Model License](https://coqui.ai/cpml). ### Contact -Come and join in our 🐸Community. We're active on [Discord](https://discord.gg/fBC58unbKE) and [Twitter](https://twitter.com/coqui_ai). -You can also mail us at info@coqui.ai. +Come and join in our 🐸Community. We're active on [Discord](https://discord.gg/fBC58unbKE) and [Github](https://github.com/eginhard/coqui-tts/discussions). ### Inference #### 🐸TTS Command line -You can check all supported languages with the following command: +You can check all supported languages with the following command: ```console tts --model_name tts_models/multilingual/multi-dataset/xtts_v2 \ --list_language_idx ``` -You can check all Coqui available speakers with the following command: +You can check all Coqui available speakers with the following command: ```console tts --model_name tts_models/multilingual/multi-dataset/xtts_v2 \ @@ -280,7 +276,7 @@ To make the `XTTS_v2` fine-tuning more accessible for users that do not have goo The Colab Notebook is available [here](https://colab.research.google.com/drive/1GiI4_X724M8q2W-zZ-jXo7cWTV7RfaH-?usp=sharing). -To learn how to use this Colab Notebook please check the [XTTS fine-tuning video](). +To learn how to use this Colab Notebook please check the [XTTS fine-tuning video](https://www.youtube.com/watch?v=8tpDiiouGxc). If you are not able to acess the video you need to follow the steps: @@ -294,7 +290,7 @@ If you are not able to acess the video you need to follow the steps: ##### Run demo locally To run the demo locally you need to do the following steps: -1. Install 🐸 TTS following the instructions available [here](https://tts.readthedocs.io/en/dev/installation.html#installation). +1. Install 🐸 TTS following the instructions available [here](https://coqui-tts.readthedocs.io/en/latest/installation.html). 2. Install the Gradio demo requirements with the command `python3 -m pip install -r TTS/demos/xtts_ft_demo/requirements.txt` 3. Run the Gradio demo using the command `python3 TTS/demos/xtts_ft_demo/xtts_demo.py` 4. Follow the steps presented in the [tutorial video](https://www.youtube.com/watch?v=8tpDiiouGxc&feature=youtu.be) to be able to fine-tune and test the fine-tuned model. diff --git a/docs/source/tutorial_for_nervous_beginners.md b/docs/source/tutorial_for_nervous_beginners.md index db753e801b..dda2abbc36 100644 --- a/docs/source/tutorial_for_nervous_beginners.md +++ b/docs/source/tutorial_for_nervous_beginners.md @@ -5,14 +5,14 @@ User friendly installation. Recommended only for synthesizing voice. ```bash -$ pip install TTS +$ pip install coqui-tts ``` Developer friendly installation. ```bash -$ git clone https://github.com/coqui-ai/TTS -$ cd TTS +$ git clone https://github.com/eginhard/coqui-tts +$ cd coqui-tts $ pip install -e . ``` @@ -109,15 +109,15 @@ $ tts -h # see the help $ tts --list_models # list the available models. ``` -![cli.gif](https://github.com/coqui-ai/TTS/raw/main/images/tts_cli.gif) +![cli.gif](https://github.com/eginhard/coqui-tts/raw/main/images/tts_cli.gif) You can call `tts-server` to start a local demo server that you can open on your favorite web browser and 🗣️ (make sure to install the additional -dependencies with `pip install TTS[server]`). +dependencies with `pip install coqui-tts[server]`). ```bash $ tts-server -h # see the help $ tts-server --list_models # list the available models. ``` -![server.gif](https://github.com/coqui-ai/TTS/raw/main/images/demo_server.gif) +![server.gif](https://github.com/eginhard/coqui-tts/raw/main/images/demo_server.gif) diff --git a/notebooks/Tutorial_1_use-pretrained-TTS.ipynb b/notebooks/Tutorial_1_use-pretrained-TTS.ipynb index 87d04c499d..3c2e9de924 100644 --- a/notebooks/Tutorial_1_use-pretrained-TTS.ipynb +++ b/notebooks/Tutorial_1_use-pretrained-TTS.ipynb @@ -41,7 +41,7 @@ "outputs": [], "source": [ "! pip install -U pip\n", - "! pip install TTS" + "! pip install coqui-tts" ] }, { diff --git a/notebooks/Tutorial_2_train_your_first_TTS_model.ipynb b/notebooks/Tutorial_2_train_your_first_TTS_model.ipynb index 0f580a85b6..c4186670c9 100644 --- a/notebooks/Tutorial_2_train_your_first_TTS_model.ipynb +++ b/notebooks/Tutorial_2_train_your_first_TTS_model.ipynb @@ -32,7 +32,7 @@ "source": [ "## Install Coqui TTS\n", "! pip install -U pip\n", - "! pip install TTS" + "! pip install coqui-tts" ] }, { @@ -44,7 +44,7 @@ "\n", "### **First things first**: we need some data.\n", "\n", - "We're training a Text-to-Speech model, so we need some _text_ and we need some _speech_. Specificially, we want _transcribed speech_. The speech must be divided into audio clips and each clip needs transcription. More details about data requirements such as recording characteristics, background noise and vocabulary coverage can be found in the [🐸TTS documentation](https://tts.readthedocs.io/en/latest/formatting_your_dataset.html).\n", + "We're training a Text-to-Speech model, so we need some _text_ and we need some _speech_. Specificially, we want _transcribed speech_. The speech must be divided into audio clips and each clip needs transcription. More details about data requirements such as recording characteristics, background noise and vocabulary coverage can be found in the [🐸TTS documentation](https://coqui-tts.readthedocs.io/en/latest/formatting_your_dataset.html).\n", "\n", "If you have a single audio file and you need to **split** it into clips. It is also important to use a lossless audio file format to prevent compression artifacts. We recommend using **wav** file format.\n", "\n", diff --git a/recipes/bel-alex73/README.md b/recipes/bel-alex73/README.md index ad378dd998..6075d3102d 100644 --- a/recipes/bel-alex73/README.md +++ b/recipes/bel-alex73/README.md @@ -39,7 +39,7 @@ Docker container was created for simplify local running. You can run `docker-pre ## Training - with GPU -You need to upload Coqui-TTS(/mycomputer/TTS/) and storage/ directory(/mycomputer/storage/) to some computer with GPU. We don't need cv-corpus/ and fanetyka/ directories for training. Install gcc, then run `pip install -e .[all,dev,notebooks]` to prepare modules. GlowTTS and HifiGan models should be learned separately based on /storage/filtered_dataset only, i.e. they are not dependent from each other. below means list of GPU ids from zero("0,1,2,3" for systems with 4 GPU). See details on the https://tts.readthedocs.io/en/latest/tutorial_for_nervous_beginners.html(multi-gpu training). +You need to upload Coqui-TTS(/mycomputer/TTS/) and storage/ directory(/mycomputer/storage/) to some computer with GPU. We don't need cv-corpus/ and fanetyka/ directories for training. Install gcc, then run `pip install -e .[all,dev,notebooks]` to prepare modules. GlowTTS and HifiGan models should be learned separately based on /storage/filtered_dataset only, i.e. they are not dependent from each other. below means list of GPU ids from zero("0,1,2,3" for systems with 4 GPU). See details on the https://coqui-tts.readthedocs.io/en/latest/tutorial_for_nervous_beginners.html (multi-gpu training). Current setup created for 24GiB GPU. You need to change batch_size if you have more or less GPU memory. Also, you can try to set lr(learning rate) to lower value in the end of training GlowTTS. diff --git a/setup.py b/setup.py index 2465f1a6b0..a25b7674f6 100644 --- a/setup.py +++ b/setup.py @@ -79,12 +79,14 @@ def pip_install(package_name): ) ] setup( - name="TTS", + name="coqui-tts", version=version, - url="https://github.com/coqui-ai/TTS", + url="https://github.com/eginhard/coqui-tts", author="Eren Gölge", author_email="egolge@coqui.ai", - description="Deep learning for Text to Speech by Coqui.", + maintainer="Enno Hermann", + maintainer_email="enno.hermann@gmail.com", + description="Deep learning for Text to Speech.", long_description=README, long_description_content_type="text/markdown", license="MPL-2.0", @@ -101,10 +103,10 @@ def pip_install(package_name): ] }, project_urls={ - "Documentation": "https://github.com/coqui-ai/TTS/wiki", - "Tracker": "https://github.com/coqui-ai/TTS/issues", - "Repository": "https://github.com/coqui-ai/TTS", - "Discussions": "https://github.com/coqui-ai/TTS/discussions", + "Documentation": "https://coqui-tts.readthedocs.io", + "Tracker": "https://github.com/eginhard/coqui-tts/issues", + "Repository": "https://github.com/eginhard/coqui-tts", + "Discussions": "https://github.com/eginhard/coqui-tts/discussions", }, cmdclass={ "build_py": build_py,