Update app.py
Browse files
app.py
CHANGED
|
@@ -18,10 +18,10 @@ css = """
|
|
| 18 |
border-radius: 6px !important;
|
| 19 |
border-color: #343140 !important;
|
| 20 |
}
|
| 21 |
-
.user {
|
| 22 |
background: #1e1c26 !important;
|
| 23 |
}
|
| 24 |
-
.assistant, .pending {
|
| 25 |
background: #16141c !important;
|
| 26 |
}
|
| 27 |
"""
|
|
@@ -46,6 +46,7 @@ def respond(
|
|
| 46 |
llm = Llama(
|
| 47 |
model_path=f"models/{model}",
|
| 48 |
n_gpu_layers=81,
|
|
|
|
| 49 |
)
|
| 50 |
provider = LlamaCppPythonProvider(llm)
|
| 51 |
|
|
|
|
| 18 |
border-radius: 6px !important;
|
| 19 |
border-color: #343140 !important;
|
| 20 |
}
|
| 21 |
+
.user.dark {
|
| 22 |
background: #1e1c26 !important;
|
| 23 |
}
|
| 24 |
+
.assistant.dark, .pending.dark {
|
| 25 |
background: #16141c !important;
|
| 26 |
}
|
| 27 |
"""
|
|
|
|
| 46 |
llm = Llama(
|
| 47 |
model_path=f"models/{model}",
|
| 48 |
n_gpu_layers=81,
|
| 49 |
+
n_ctx=8192,
|
| 50 |
)
|
| 51 |
provider = LlamaCppPythonProvider(llm)
|
| 52 |
|