Commit
·
0f7b9e2
1
Parent(s):
af9c7a0
Add flashinfer-python dependency
Browse files- generate-responses.py +2 -1
generate-responses.py
CHANGED
|
@@ -2,6 +2,7 @@
|
|
| 2 |
# requires-python = ">=3.10"
|
| 3 |
# dependencies = [
|
| 4 |
# "datasets",
|
|
|
|
| 5 |
# "huggingface-hub[hf_transfer]",
|
| 6 |
# "torch",
|
| 7 |
# "transformers",
|
|
@@ -223,7 +224,7 @@ def main(
|
|
| 223 |
if max_model_len is not None:
|
| 224 |
vllm_kwargs["max_model_len"] = max_model_len
|
| 225 |
logger.info(f"Using max_model_len={max_model_len}")
|
| 226 |
-
|
| 227 |
llm = LLM(**vllm_kwargs)
|
| 228 |
|
| 229 |
# Load tokenizer for chat template
|
|
|
|
| 2 |
# requires-python = ">=3.10"
|
| 3 |
# dependencies = [
|
| 4 |
# "datasets",
|
| 5 |
+
# "flashinfer-python",
|
| 6 |
# "huggingface-hub[hf_transfer]",
|
| 7 |
# "torch",
|
| 8 |
# "transformers",
|
|
|
|
| 224 |
if max_model_len is not None:
|
| 225 |
vllm_kwargs["max_model_len"] = max_model_len
|
| 226 |
logger.info(f"Using max_model_len={max_model_len}")
|
| 227 |
+
|
| 228 |
llm = LLM(**vllm_kwargs)
|
| 229 |
|
| 230 |
# Load tokenizer for chat template
|