Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Render current messages in snippets (#19)
Browse files
src/lib/components/CodeSnippets.svelte
CHANGED
|
@@ -33,11 +33,18 @@
|
|
| 33 |
|
| 34 |
let selectedLanguage: Language = 'javascript';
|
| 35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
function highlight(code: string, language: Language) {
|
| 37 |
return hljs.highlight(code, { language }).value;
|
| 38 |
}
|
| 39 |
|
| 40 |
function getJavascriptSnippets() {
|
|
|
|
| 41 |
const snippets: Snippet[] = [];
|
| 42 |
snippets.push({
|
| 43 |
label: 'Install @huggingface/inference',
|
|
@@ -57,9 +64,7 @@ let out = "";
|
|
| 57 |
|
| 58 |
for await (const chunk of inference.chatCompletionStream({
|
| 59 |
model: "${conversation.model}",
|
| 60 |
-
messages:
|
| 61 |
-
{ role: "user", content: "Tell me a story" },
|
| 62 |
-
],
|
| 63 |
max_tokens: ${conversation.config.maxTokens},
|
| 64 |
temperature: ${conversation.config.temperature},
|
| 65 |
seed: 0,
|
|
@@ -81,9 +86,7 @@ const inference = new HfInference("your access token")
|
|
| 81 |
|
| 82 |
const out = await inference.chatCompletion({
|
| 83 |
model: "${conversation.model}",
|
| 84 |
-
messages:
|
| 85 |
-
{ role: "user", content: "Who are you?" }
|
| 86 |
-
],
|
| 87 |
max_tokens: ${conversation.config.maxTokens},
|
| 88 |
temperature: ${conversation.config.temperature},
|
| 89 |
seed: 0,
|
|
|
|
| 33 |
|
| 34 |
let selectedLanguage: Language = 'javascript';
|
| 35 |
|
| 36 |
+
function getMessages(){
|
| 37 |
+
const placeholder = [{ role: "user", content: "Tell me a story" }];
|
| 38 |
+
const messages = conversation.messages.length ? conversation.messages : placeholder;
|
| 39 |
+
return JSON.stringify(messages, null, 2);
|
| 40 |
+
}
|
| 41 |
+
|
| 42 |
function highlight(code: string, language: Language) {
|
| 43 |
return hljs.highlight(code, { language }).value;
|
| 44 |
}
|
| 45 |
|
| 46 |
function getJavascriptSnippets() {
|
| 47 |
+
const messagesStr = getMessages().replace(/"([^"]+)":/g, '$1:');
|
| 48 |
const snippets: Snippet[] = [];
|
| 49 |
snippets.push({
|
| 50 |
label: 'Install @huggingface/inference',
|
|
|
|
| 64 |
|
| 65 |
for await (const chunk of inference.chatCompletionStream({
|
| 66 |
model: "${conversation.model}",
|
| 67 |
+
messages: ${messagesStr},
|
|
|
|
|
|
|
| 68 |
max_tokens: ${conversation.config.maxTokens},
|
| 69 |
temperature: ${conversation.config.temperature},
|
| 70 |
seed: 0,
|
|
|
|
| 86 |
|
| 87 |
const out = await inference.chatCompletion({
|
| 88 |
model: "${conversation.model}",
|
| 89 |
+
messages: ${messagesStr},
|
|
|
|
|
|
|
| 90 |
max_tokens: ${conversation.config.maxTokens},
|
| 91 |
temperature: ${conversation.config.temperature},
|
| 92 |
seed: 0,
|