Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
inference-playground
/
src
/lib
/components
/InferencePlayground
/InferencePlaygroundGenerationConfig.svelte
| <script context="module" lang="ts"> | |
| export const defaultSystemMessage: { [key: string]: string } = { | |
| "Qwen/QwQ-32B-Preview": | |
| "You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step.", | |
| } as const; | |
| </script> | |
| <script lang="ts"> | |
| import type { Conversation } from "$lib/components/InferencePlayground/types"; | |
| import { GENERATION_CONFIG_KEYS, GENERATION_CONFIG_SETTINGS } from "./generationConfigSettings"; | |
| export let conversation: Conversation; | |
| export let classNames = ""; | |
| const customMaxTokens: { [key: string]: number } = { | |
| "01-ai/Yi-1.5-34B-Chat": 2048, | |
| "HuggingFaceM4/idefics-9b-instruct": 2048, | |
| "deepseek-ai/DeepSeek-Coder-V2-Instruct": 16384, | |
| "bigcode/starcoder": 8192, | |
| "bigcode/starcoderplus": 8192, | |
| "HuggingFaceH4/starcoderbase-finetuned-oasst1": 8192, | |
| "google/gemma-7b": 8192, | |
| "google/gemma-1.1-7b-it": 8192, | |
| "google/gemma-2b": 8192, | |
| "google/gemma-1.1-2b-it": 8192, | |
| "google/gemma-2-27b-it": 8192, | |
| "google/gemma-2-9b-it": 4096, | |
| "google/gemma-2-2b-it": 8192, | |
| "tiiuae/falcon-7b": 8192, | |
| "tiiuae/falcon-7b-instruct": 8192, | |
| "timdettmers/guanaco-33b-merged": 2048, | |
| "mistralai/Mixtral-8x7B-Instruct-v0.1": 32768, | |
| "Qwen/Qwen2.5-72B-Instruct": 32768, | |
| "Qwen/Qwen2.5-Coder-32B-Instruct": 32768, | |
| "meta-llama/Meta-Llama-3-70B-Instruct": 8192, | |
| "CohereForAI/c4ai-command-r-plus-08-2024": 32768, | |
| "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": 32768, | |
| "meta-llama/Llama-2-70b-chat-hf": 8192, | |
| "HuggingFaceH4/zephyr-7b-alpha": 17432, | |
| "HuggingFaceH4/zephyr-7b-beta": 32768, | |
| "mistralai/Mistral-7B-Instruct-v0.1": 32768, | |
| "mistralai/Mistral-7B-Instruct-v0.2": 32768, | |
| "mistralai/Mistral-7B-Instruct-v0.3": 32768, | |
| "mistralai/Mistral-Nemo-Instruct-2407": 32768, | |
| "meta-llama/Meta-Llama-3-8B-Instruct": 8192, | |
| "mistralai/Mistral-7B-v0.1": 32768, | |
| "bigcode/starcoder2-3b": 16384, | |
| "bigcode/starcoder2-15b": 16384, | |
| "HuggingFaceH4/starchat2-15b-v0.1": 16384, | |
| "codellama/CodeLlama-7b-hf": 8192, | |
| "codellama/CodeLlama-13b-hf": 8192, | |
| "codellama/CodeLlama-34b-Instruct-hf": 8192, | |
| "meta-llama/Llama-2-7b-chat-hf": 8192, | |
| "meta-llama/Llama-2-13b-chat-hf": 8192, | |
| "OpenAssistant/oasst-sft-6-llama-30b": 2048, | |
| "TheBloke/vicuna-7B-v1.5-GPTQ": 2048, | |
| "HuggingFaceH4/starchat-beta": 8192, | |
| "bigcode/octocoder": 8192, | |
| "vwxyzjn/starcoderbase-triviaqa": 8192, | |
| "lvwerra/starcoderbase-gsm8k": 8192, | |
| "NousResearch/Hermes-3-Llama-3.1-8B": 16384, | |
| "microsoft/Phi-3.5-mini-instruct": 32768, | |
| "meta-llama/Llama-3.1-70B-Instruct": 32768, | |
| "meta-llama/Llama-3.1-8B-Instruct": 8192, | |
| } as const; | |
| $: modelMaxLength = customMaxTokens[conversation.model.id] ?? conversation.model.tokenizerConfig.model_max_length; | |
| $: maxTokens = Math.min(modelMaxLength ?? GENERATION_CONFIG_SETTINGS["max_tokens"].max, 64_000); | |
| </script> | |
| <div class="flex flex-col gap-y-7 {classNames}"> | |
| {#each GENERATION_CONFIG_KEYS as key} | |
| {@const { label, min, step } = GENERATION_CONFIG_SETTINGS[key]} | |
| {@const max = key === "max_tokens" ? maxTokens : GENERATION_CONFIG_SETTINGS[key].max} | |
| <div> | |
| <div class="flex items-center justify-between"> | |
| <label for="temperature-range" class="mb-2 block text-sm font-medium text-gray-900 dark:text-white" | |
| >{label}</label | |
| > | |
| <input | |
| type="number" | |
| class="w-18 rounded border bg-transparent px-1 py-0.5 text-right text-sm dark:border-gray-700" | |
| {min} | |
| {max} | |
| {step} | |
| bind:value={conversation.config[key]} | |
| /> | |
| </div> | |
| <input | |
| id="temperature-range" | |
| type="range" | |
| {min} | |
| {max} | |
| {step} | |
| bind:value={conversation.config[key]} | |
| class="h-2 w-full cursor-pointer appearance-none rounded-lg bg-gray-200 accent-black dark:bg-gray-700 dark:accent-blue-500" | |
| /> | |
| </div> | |
| {/each} | |
| <div class="mt-2"> | |
| <label class="flex cursor-pointer items-center justify-between"> | |
| <input type="checkbox" bind:checked={conversation.streaming} class="peer sr-only" /> | |
| <span class="text-sm font-medium text-gray-900 dark:text-gray-300">Streaming</span> | |
| <div | |
| class="peer relative h-5 w-9 rounded-full bg-gray-200 after:absolute after:start-[2px] after:top-[2px] after:h-4 after:w-4 after:rounded-full after:border after:border-gray-300 after:bg-white after:transition-all after:content-[''] peer-checked:bg-black peer-checked:after:translate-x-full peer-checked:after:border-white peer-focus:outline-none dark:border-gray-600 dark:bg-gray-700 dark:peer-checked:bg-blue-600" | |
| ></div> | |
| </label> | |
| </div> | |
| </div> | |