fish-speech
fish-speech copied to clipboard
[BUG]Can't open inference server
Describe the bug Can't open inference server.
To Reproduce
- Run install_env.bat with USE_MIRROR=false and INSTALL_TYPE=stable
- Change API_FLAGS.txt and enable "--infer", then Run start.bat
- Go to the inference tab and click the "open inference server"
- Go to the web address http://127.0.0.1:7860
Expected behavior The inference web UI should be shown at http://127.0.0.1:7860
Actual behavior No inference web UI is show. The inference service is not running
Screenshots / log It seems that it complains that the triton can't find the CUDA lib. However, according to nvidia's doc: https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/getting_started/quickstart.html#run-on-cpu-only-system. The triton should be able to run without GPU as well.
The python stacktrace appeared twice in the following log. The first happens when starting the webui. I can see that webpage. But When I go to the "inference" tab on the page, and click "open inference server", it shows the same stacktrace and the inference server webpage is not shown.
Start WebUI Inference...
Debug: flags = --listen 0.0.0.0:8000 --llama-checkpoint-path "checkpoints/fish-speech-1.2" --decoder-checkpoint-path "checkpoints/fish-speech-1.2/firefly-gan-vq-fsq-4x1024-42hz-generator.pth" --decoder-config-name firefly_gan_vq
Traceback (most recent call last):
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\runpy.py", line 86, in _run_code
exec(code, run_globals)
File "C:\Users\username\Downloads\fish-speech\tools\webui.py", line 23, in <module>
from tools.api import decode_vq_tokens, encode_reference
File "C:\Users\username\Downloads\fish-speech\tools\api.py", line 34, in <module>
from fish_speech.models.vqgan.modules.firefly import FireflyArchitecture
File "C:\Users\username\Downloads\fish-speech\fish_speech\models\vqgan\__init__.py", line 1, in <module>
from .lit_module import VQGAN
File "C:\Users\username\Downloads\fish-speech\fish_speech\models\vqgan\lit_module.py", line 5, in <module>
import lightning as L
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\__init__.py", line 19, in <module>
from lightning.fabric.fabric import Fabric # noqa: E402
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\__init__.py", line 30, in <module>
from lightning.fabric.fabric import Fabric # noqa: E402
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\fabric.py", line 46, in <module>
from lightning.fabric.loggers import Logger
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\loggers\__init__.py", line 15, in <module>
from lightning.fabric.loggers.tensorboard import TensorBoardLogger # noqa: F401
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\loggers\tensorboard.py", line 31, in <module>
from lightning.fabric.wrappers import _unwrap_objects
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\wrappers.py", line 38, in <module>
from torch._dynamo import OptimizedModule
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\__init__.py", line 2, in <module>
from . import convert_frame, eval_frame, resume_execution
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\convert_frame.py", line 41, in <module>
from . import config, exc, trace_rules
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\exc.py", line 11, in <module>
from .utils import counters
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\utils.py", line 1031, in <module>
if has_triton_package():
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\utils\_triton.py", line 8, in has_triton_package
import triton
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\__init__.py", line 8, in <module>
from .runtime import (
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\__init__.py", line 1, in <module>
from .autotuner import (Autotuner, Config, Heuristics, OutOfResources, autotune, heuristics)
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\autotuner.py", line 7, in <module>
from ..testing import do_bench
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\testing.py", line 7, in <module>
from . import language as tl
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\language\__init__.py", line 6, in <module>
from .standard import (
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\language\standard.py", line 3, in <module>
from ..runtime.jit import jit
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\jit.py", line 10, in <module>
from ..runtime.driver import driver
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\driver.py", line 1, in <module>
from ..backends import backends
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\__init__.py", line 50, in <module>
backends = _discover_backends()
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\__init__.py", line 43, in _discover_backends
compiler = _load_module(name, os.path.join(root, name, 'compiler.py'))
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\__init__.py", line 12, in _load_module
spec.loader.exec_module(module)
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\nvidia\compiler.py", line 3, in <module>
from triton.backends.nvidia.driver import CudaUtils
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\nvidia\driver.py", line 18, in <module>
library_dir += [os.path.join(os.environ.get("CUDA_PATH"), "lib", "x64")]
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\ntpath.py", line 104, in join
path = os.fspath(path)
TypeError: expected str, bytes or os.PathLike object, not NoneType
Next launch the page...
['', 'C:\\Users\\username\\Downloads\\fish-speech\\fish_speech\\webui', 'C:\\Users\\username\\Downloads\\fish-speech', 'C:\\Users\\username\\Downloads\\fish-speech\\fishenv\\env\\python310.zip', 'C:\\Users\\username\\Downloads\\fish-speech\\fishenv\\env\\DLLs', 'C:\\Users\\username\\Downloads\\fish-speech\\fishenv\\env\\lib', 'C:\\Users\\username\\Downloads\\fish-speech\\fishenv\\env', 'C:\\Users\\username\\Downloads\\fish-speech\\fishenv\\env\\lib\\site-packages', '__editable__.fish_speech-0.1.0.finder.__path_hook__']
You are in C:\Users\username\Downloads\fish-speech
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.
2024-07-07 00:14:39.928 | INFO | __main__:clean_infer_cache:146 - C:\Users\username\AppData\Local\Temp\gradio was not found
Traceback (most recent call last):
File "C:\Users\username\Downloads\fish-speech\tools\webui.py", line 23, in <module>
from tools.api import decode_vq_tokens, encode_reference
File "C:\Users\username\Downloads\fish-speech\tools\api.py", line 34, in <module>
from fish_speech.models.vqgan.modules.firefly import FireflyArchitecture
File "C:\Users\username\Downloads\fish-speech\fish_speech\models\vqgan\__init__.py", line 1, in <module>
from .lit_module import VQGAN
File "C:\Users\username\Downloads\fish-speech\fish_speech\models\vqgan\lit_module.py", line 5, in <module>
import lightning as L
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\__init__.py", line 19, in <module>
from lightning.fabric.fabric import Fabric # noqa: E402
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\__init__.py", line 30, in <module>
from lightning.fabric.fabric import Fabric # noqa: E402
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\fabric.py", line 46, in <module>
from lightning.fabric.loggers import Logger
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\loggers\__init__.py", line 15, in <module>
from lightning.fabric.loggers.tensorboard import TensorBoardLogger # noqa: F401
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\loggers\tensorboard.py", line 31, in <module>
from lightning.fabric.wrappers import _unwrap_objects
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\lightning\fabric\wrappers.py", line 38, in <module>
from torch._dynamo import OptimizedModule
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\__init__.py", line 2, in <module>
from . import convert_frame, eval_frame, resume_execution
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\convert_frame.py", line 41, in <module>
from . import config, exc, trace_rules
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\exc.py", line 11, in <module>
from .utils import counters
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\_dynamo\utils.py", line 1031, in <module>
if has_triton_package():
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\torch\utils\_triton.py", line 8, in has_triton_package
import triton
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\__init__.py", line 8, in <module>
from .runtime import (
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\__init__.py", line 1, in <module>
from .autotuner import (Autotuner, Config, Heuristics, OutOfResources, autotune, heuristics)
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\autotuner.py", line 7, in <module>
from ..testing import do_bench
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\testing.py", line 7, in <module>
from . import language as tl
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\language\__init__.py", line 6, in <module>
from .standard import (
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\language\standard.py", line 3, in <module>
from ..runtime.jit import jit
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\jit.py", line 10, in <module>
from ..runtime.driver import driver
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\runtime\driver.py", line 1, in <module>
from ..backends import backends
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\__init__.py", line 50, in <module>
backends = _discover_backends()
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\__init__.py", line 43, in _discover_backends
compiler = _load_module(name, os.path.join(root, name, 'compiler.py'))
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\__init__.py", line 12, in _load_module
spec.loader.exec_module(module)
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\nvidia\compiler.py", line 3, in <module>
from triton.backends.nvidia.driver import CudaUtils
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\site-packages\triton\backends\nvidia\driver.py", line 18, in <module>
library_dir += [os.path.join(os.environ.get("CUDA_PATH"), "lib", "x64")]
File "C:\Users\username\Downloads\fish-speech\fishenv\env\lib\ntpath.py", line 104, in join
path = os.fspath(path)
TypeError: expected str, bytes or os.PathLike object, not NoneType
Additional context Windows 11. Intel integrated graphics card. Use lastest master code.