Update README.md
Browse files
README.md
CHANGED
@@ -44,7 +44,7 @@ Documentation on installing and using vLLM [can be found here](https://vllm.read
|
|
44 |
- vLLM can be deployed as a server that implements the OpenAI API protocol. This allows vLLM to be used as a drop-in replacement for applications using OpenAI API
|
45 |
|
46 |
```shell
|
47 |
-
python3 -m vllm.entrypoints.openai.api_server --model Copycats/EEVE-Korean-Instruct-10.8B-v1.0-AWQ --quantization awq --dtype
|
48 |
```
|
49 |
|
50 |
#### Querying the model using OpenAI Chat API:
|
|
|
44 |
- vLLM can be deployed as a server that implements the OpenAI API protocol. This allows vLLM to be used as a drop-in replacement for applications using OpenAI API
|
45 |
|
46 |
```shell
|
47 |
+
python3 -m vllm.entrypoints.openai.api_server --model Copycats/EEVE-Korean-Instruct-10.8B-v1.0-AWQ --quantization awq --dtype float16
|
48 |
```
|
49 |
|
50 |
#### Querying the model using OpenAI Chat API:
|