| --- |
| license: cc-by-nc-4.0 |
| pipeline_tag: audio-to-audio |
| library_name: f5-tts |
| extra_gated_prompt: "You agree to not use the model to generate, share, or promote content that is illegal, harmful, deceptive, or intended to impersonate real individuals without their informed consent." |
| extra_gated_fields: |
| Affiliation: text |
| Country: country |
| I agree to use this model for non-commercial use ONLY: checkbox |
| --- |
| |
| # EZ-VC: Easy Zero-shot Any-to-Any Voice Conversion |
|
|
| [](https://github.com/EZ-VC/EZ-VC) |
| [](https://aclanthology.org/2025.findings-emnlp.1077/) |
| [](https://ez-vc.github.io/EZ-VC-Demo/) |
| [](https://asr.iitm.ac.in/) |
| <!-- <img src="https://github.com/user-attachments/assets/12d7749c-071a-427c-81bf-b87b91def670" alt="Watermark" style="width: 40px; height: auto"> --> |
|
|
|
|
| ### Our paper has been published in the Findings of EMNLP 2025! |
|
|
| ## Installation |
|
|
| ### Create a separate environment if needed |
|
|
| ```bash |
| # Create a python 3.10 conda env (you could also use virtualenv) |
| conda create -n ez-vc python=3.10 |
| conda activate ez-vc |
| ``` |
|
|
| ### Local installation |
|
|
| ```bash |
| git clone https://github.com/EZ-VC/EZ-VC |
| cd EZ-VC |
| git submodule update --init --recursive |
| pip install -e . |
| |
| # Install espnet for xeus (Exactly this version) |
| pip install 'espnet @ git+https://github.com/wanchichen/espnet.git@ssl' |
| ``` |
|
|
| ## Inference |
|
|
| We have provided a Jupyter notebook for inference in "src/f5_tts/infer/infer.ipynb". |
| |
| Open [Inference notebook](https://github.com/EZ-VC/EZ-VC/blob/main/src/f5_tts/infer/infer.ipynb). |
| |
| Run all. |
| |
| The converted audio will be available at the last cell. |
| |
| |
| ## Acknowledgements |
| |
| - [F5-TTS](https://arxiv.org/abs/2410.06885) for opensourcing their code which has made EZ-VC possible. |
| |
| ## Citation |
| If our work and codebase is useful for you, please cite as: |
| ``` |
| @inproceedings{joglekar-etal-2025-ez, |
| title = "{EZ}-{VC}: Easy Zero-shot Any-to-Any Voice Conversion", |
| author = "Joglekar, Advait and |
| Singh, Divyanshu and |
| Bhatia, Rooshil Rohit and |
| Umesh, Srinivasan", |
| editor = "Christodoulopoulos, Christos and |
| Chakraborty, Tanmoy and |
| Rose, Carolyn and |
| Peng, Violet", |
| booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025", |
| month = nov, |
| year = "2025", |
| address = "Suzhou, China", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2025.findings-emnlp.1077/", |
| doi = "10.18653/v1/2025.findings-emnlp.1077", |
| pages = "19768--19774", |
| ISBN = "979-8-89176-335-7", |
| abstract = "Voice Conversion research in recent times has increasingly focused on improving the zero-shot capabilities of existing methods. Despite remarkable advancements, current architectures still tend to struggle in zero-shot cross-lingual settings. They are also often unable to generalize for speakers of unseen languages and accents. In this paper, we adopt a simple yet effective approach that combines discrete speech representations from self-supervised models with a non-autoregressive Diffusion-Transformer based conditional flow matching speech decoder. We show that this architecture allows us to train a voice-conversion model in a purely textless, self-supervised fashion. Our technique works without requiring multiple encoders to disentangle speech features. Our model also manages to excel in zero-shot cross-lingual settings even for unseen languages. We provide our code, model checkpoint and demo samples here: https://github.com/ez-vc/ez-vc" |
| } |
| ``` |
| ## License |
| |
| Our code is released under MIT License. The pre-trained models are licensed under the CC-BY-NC license. Sorry for any inconvenience this may cause. |