🎤 SoulX-Singer
Official inference code for
SoulX-Singer: Towards High-Quality Zero-Shot Singing Voice Synthesis
🎵 Overview
SoulX-Singer is a high-fidelity, zero-shot singing voice synthesis model that enables users to generate realistic singing voices for unseen singers. It supports melody-conditioned (F0 contour) and score-conditioned (MIDI notes) control for precise pitch, rhythm, and expression.
SoulX-Singer-SVC is a singing voice conversion (SVC) model finetuned from SoulX-Singer. Singing Voice Conversion aims to transform a source singing recording into the target singer’s voice while preserving the original melody, rhythm, and lyrical content. Based on the strong generative capability of SoulX-Singer, SoulX-Singer-SVC enables high-quality singing voice conversion directly from raw singing audio, without requiring lyric or MIDI transcriptions.
✨ Key Features
SoulX-Singer
- 🎤 Zero-Shot Singing – Generate high-fidelity voices for unseen singers, no fine-tuning needed.
- 🎵 Flexible Control Modes – Melody (F0) and Score (MIDI) conditioning.
- 📚 Large-Scale Dataset – 42,000+ hours of aligned vocals, lyrics, notes across Mandarin, English, Cantonese.
- 🧑🎤 Timbre Cloning – Preserve singer identity across languages, styles, and edited lyrics.
- ✏️ Singing Voice Editing – Modify lyrics while keeping natural prosody.
- 🌐 Cross-Lingual Synthesis – High-fidelity synthesis by disentangling timbre from content.
SoulX-Singer-SVC
- 🎙️ Zero-Shot Timbre and Style Transfer – Transfer singer identity and style to unseen voices without per-speaker fine-tuning.
- 🌍 Language-Agnostic Conversion – Works across multilingual singing content.
- 🔄 Transcription-Free Audio-to-Audio Conversion – Convert target singing directly without lyrics transcription or MIDI inputs.
🎬 Demo Examples
Singing Voice Synthesis (SVS)
Singing Voice Conversion (SVC)
📰 News
- [2026-03-16] SoulX-Singer-SVC is released, and SoulX-Singer Online Demo has been updated to support singing voice conversion (SVC).
- [2026-02-12] SoulX-Singer Eval Dataset is now available on Hugging Face Datasets.
- [2026-02-09] SoulX-Singer Online Demo is live on Hugging Face Spaces — try singing voice synthesis in your browser.
- [2026-02-08] MIDI Editor is available on Hugging Face Spaces.
- [2026-02-06] SoulX-Singer inference code and models released.
🚀 Quick Start
1. Clone Repository
git clone https://github.com/Soul-AILab/SoulX-Singer.git cd SoulX-Singer
2. Set Up Environment
1. Install Conda (if not already installed): https://docs.conda.io/en/latest/miniconda.html
2. Create and activate a Conda environment:
conda create -n soulxsinger -y python=3.10
conda activate soulxsinger
3. Install dependencies:
pip install -r requirements.txt
⚠️ If you are in mainland China, use a PyPI mirror:
pip install -r requirements.txt -i https://mirrors.aliyun.com/pypi/simple/ --trusted-host=mirrors.aliyun.com
3. Download Pretrained Models
Install Hugging Face Hub if needed:
pip install -U huggingface_hub
Download the SVS, SVC model and preprocessing models:
pip install -U huggingface_hub # Download the SoulX-Singer SVS and SVC model hf download Soul-AILab/SoulX-Singer --local-dir pretrained_models/SoulX-Singer # Download models required for preprocessing hf download Soul-AILab/SoulX-Singer-Preprocess --local-dir pretrained_models/SoulX-Singer-Preprocess
4. Run the Demo
Run the SVS inference demo
bash example/infer.sh
This script relies on metadata generated from the preprocessing pipeline, including vocal separation and transcription. Users should follow the steps in preprocess to prepare the necessary metadata before running the demo with their own data.
⚠️ Important Note The metadata produced by the automatic preprocessing pipeline may not perfectly align the singing audio with the corresponding lyrics and musical notes. For best synthesis quality, we strongly recommend manually correcting the alignment using the 🎼 Midi-Editor.
How to use the Midi-Editor:
Run the SVC inference demo
bash example/infer_svc.sh
This example performs audio-to-audio SVC, converting the target singing into the prompt timbre using waveform and F0 inputs.
To prepare your own SVC data, run example/preprocess.sh with midi_transcribe=False.
🌐 WebUI
You can launch the interactive interface for SVS (Synthesised from lyrics and MIDI transcriptions) with:
python webui.py
For SVC WebUI (audio-to-audio conversion):
python webui_svc.py
🚧 Roadmap
- 🖥️ Web-based UI for easy and interactive inference
- 🌐 Online MIDI Editor deployment on Hugging Face Spaces
- 🌐 Online demo deployment on Hugging Face Spaces
- 📊 Release the SoulX-Singer-Eval benchmark
- 🎹 Inference support for user-friendly MIDI-based input
- 📚 Comprehensive tutorials and usage documentation
- 🎵 Support for wav-to-wav singing voice conversion (without transcription)
🙏 Acknowledgements
Special thanks to the following open-source projects:
- F5-TTS
- Amphion
- Music Source Separation Training
- Lead Vocal Separation
- Vocal Dereverberation
- RMVPE Paraformer
- Parakeet-tdt-0.6b-v2
- ROSVOT
📄 License
We use the Apache 2.0 license. Researchers and developers are free to use the codes and model weights of our SoulX-Singer. Check the license at LICENSE for more details.
⚠️ Usage Disclaimer
SoulX-Singer is intended for academic research, educational purposes, and legitimate applications such as personalized singing synthesis and assistive technologies.
Please note:
- 🎤 Respect intellectual property, privacy, and personal consent when generating singing content.
- 🚫 Do not use the model to impersonate individuals without authorization or to create deceptive audio.
- ⚠️ The developers assume no liability for any misuse of this model.
We advocate for the responsible development and use of AI and encourage the community to uphold safety and ethical principles. For ethics or misuse concerns, please contact us.
📄 Citation
If you use SoulX-Singer in your research, please cite:
@misc{soulxsinger, title={SoulX-Singer: Towards High-Quality Zero-Shot Singing Voice Synthesis}, author={Jiale Qian and Hao Meng and Tian Zheng and Pengcheng Zhu and Haopeng Lin and Yuhang Dai and Hanke Xie and Wenxiao Cao and Ruixuan Shang and Jun Wu and Hongmei Liu and Hanlin Wen and Jian Zhao and Zhonglin Jiang and Yong Chen and Shunshun Yin and Ming Tao and Jianguo Wei and Lei Xie and Xinsheng Wang}, year={2026}, eprint={2602.07803}, archivePrefix={arXiv}, primaryClass={eess.AS}, url={https://arxiv.org/abs/2602.07803}, }
📬 Contact Us
We welcome your feedback, questions, and collaboration:
-
Email: qianjiale@soulapp.cn | menghao@soulapp.cn | wangxinsheng@soulapp.cn
-
Join discussions: WeChat or Soul APP groups for technical discussions and updates: