Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to use vLLM? #780

Open
yikchunnnn opened this issue Oct 11, 2024 · 3 comments
Open

How to use vLLM? #780

yikchunnnn opened this issue Oct 11, 2024 · 3 comments
Labels
documentation Improvements or additions to documentation

Comments

@yikchunnnn
Copy link

The Readme only mentions 'pip install vllm...'

May I know how shall I proceed to enable vLLM?

Do I need to do anything after installing vLLM?
Or ChatTTS will automatically use vLLM when detecting it available?

Thank you in advance.

@2noise 2noise deleted a comment from rose07 Oct 15, 2024
@fumiama fumiama added the documentation Improvements or additions to documentation label Oct 15, 2024
@fumiama
Copy link
Member

fumiama commented Oct 15, 2024

Now vLLM is still under test&dev and ONLY BASIC INFER is available. It will not be enabled be default and if you are a developer, you can easily find the way to enable it by looking the parameters of Chat.load.

@the-nine-nation
Copy link

Using vllm would lead to an error: ImportError: cannot import name 'LogicalTokenBlock' from 'vllm.block' (/root/miniconda3/envs/py39/lib/python3.9/site-packages/vllm/block.py)
How can i fix it?

@superstring
Copy link

Now vLLM is still under test&dev and ONLY BASIC INFER is available. It will not be enabled be default and if you are a developer, you can easily find the way to enable it by looking the parameters of Chat.load.

Does this mean that zero-shot infer is not supported yet?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

No branches or pull requests

4 participants