Describe the bug
Getting a Type error while running xtts_v2 inside the provided docker image.
TypeError: expected str, bytes or os.PathLike object, not NoneType
To Reproduce
docker run --rm -it -p 5002:5002 --gpus all --entrypoint /bin/bash ghcr.io/coqui-ai/tts
python3 TTS/server/server.py --model_name tts_models/multilingual/multi-dataset/xtts_v2 --use_cuda true
Expected behavior
This model should run as expected and have an api exposed to use this as a server.
Logs
root@552be:~# python3 TTS/server/server.py --model_name tts_models/multilingual/multi-dataset/xtts_v2 --use_cuda true
> tts_models/multilingual/multi-dataset/xtts_v2 is already downloaded.
Traceback (most recent call last):
File "/root/TTS/server/server.py", line 104, in <module>
synthesizer = Synthesizer(
File "/root/TTS/utils/synthesizer.py", line 93, in __init__
self._load_tts(tts_checkpoint, tts_config_path, use_cuda)
File "/root/TTS/utils/synthesizer.py", line 183, in _load_tts
self.tts_config = load_config(tts_config_path)
File "/root/TTS/config/__init__.py", line 82, in load_config
ext = os.path.splitext(config_path)[1]
File "/usr/lib/python3.10/posixpath.py", line 118, in splitext
p = os.fspath(p)
TypeError: expected str, bytes or os.PathLike object, not NoneType
Environment
{
"CUDA": {
"GPU": [
"NVIDIA GeForce GTX 1650"
],
"available": true,
"version": "11.8"
},
"Packages": {
"PyTorch_debug": false,
"PyTorch_version": "2.1.1+cu118",
"TTS": "0.22.0",
"numpy": "1.22.0"
},
"System": {
"OS": "Linux",
"architecture": [
"64bit",
""
],
"processor": "x86_64",
"python": "3.10.12",
"version": "#1 SMP PREEMPT_DYNAMIC Thu Jun 5 18:30:46 UTC 2025"
}
}
Additional context
No response
Describe the bug
Getting a Type error while running xtts_v2 inside the provided docker image.
TypeError: expected str, bytes or os.PathLike object, not NoneType
To Reproduce
docker run --rm -it -p 5002:5002 --gpus all --entrypoint /bin/bash ghcr.io/coqui-ai/tts
python3 TTS/server/server.py --model_name tts_models/multilingual/multi-dataset/xtts_v2 --use_cuda true
Expected behavior
This model should run as expected and have an api exposed to use this as a server.
Logs
Environment
{ "CUDA": { "GPU": [ "NVIDIA GeForce GTX 1650" ], "available": true, "version": "11.8" }, "Packages": { "PyTorch_debug": false, "PyTorch_version": "2.1.1+cu118", "TTS": "0.22.0", "numpy": "1.22.0" }, "System": { "OS": "Linux", "architecture": [ "64bit", "" ], "processor": "x86_64", "python": "3.10.12", "version": "#1 SMP PREEMPT_DYNAMIC Thu Jun 5 18:30:46 UTC 2025" } }Additional context
No response