Benjamin Consolvo
commited on
Commit
·
97ccf9a
1
Parent(s):
0577ef7
markdown links
Browse files- app.py +3 -15
- model_info.json +0 -18
app.py
CHANGED
|
@@ -5,38 +5,26 @@ import os
|
|
| 5 |
from openai import OpenAI
|
| 6 |
import json
|
| 7 |
|
| 8 |
-
working_dir = os.path.dirname(os.path.abspath(__file__))
|
| 9 |
-
endpoint_data = json.load(open(f"{working_dir}/model_info.json"))
|
| 10 |
-
|
| 11 |
def clear_chat():
|
| 12 |
st.session_state.messages = []
|
| 13 |
|
| 14 |
st.title("Intel® AI for Enterprise Inference")
|
| 15 |
st.header("LLM chatbot")
|
| 16 |
|
| 17 |
-
# Extract the keys (model names) from the JSON data
|
| 18 |
-
# model_names = list(endpoint_data.keys())
|
| 19 |
-
|
| 20 |
-
|
| 21 |
with st.sidebar:
|
| 22 |
-
#Enter openai_api key under "Secrets " in HF settings
|
| 23 |
-
#Enter base_url under "Variables" in HF settings
|
| 24 |
-
api_key = st.session_state.api_key = st.secrets["openai_apikey"]
|
| 25 |
-
base_url = st.session_state.base_url = os.environ.get("base_url")
|
| 26 |
client = OpenAI(api_key=api_key, base_url=base_url)
|
| 27 |
models = client.models.list()
|
| 28 |
model_names = sorted([model.id for model in models]) # Extract 'id' from each model object
|
| 29 |
-
# Specify the default model name
|
| 30 |
default_model_name = "meta-llama/Llama-3.3-70B-Instruct" # Replace with your desired default model name
|
| 31 |
|
| 32 |
-
|
| 33 |
# Use st.session_state to persist the selected model
|
| 34 |
if "selected_model" not in st.session_state:
|
| 35 |
-
# st.session_state.selected_model = model_names[0] # Default to the first model
|
| 36 |
st.session_state.selected_model = default_model_name if default_model_name in model_names else model_names[0]
|
| 37 |
|
| 38 |
modelname = st.selectbox(
|
| 39 |
-
"Select LLM model (Running on Intel® Gaudi®) on Denvr Dataworks",
|
| 40 |
model_names,
|
| 41 |
index=model_names.index(st.session_state.selected_model) if st.session_state.selected_model in model_names else 0,
|
| 42 |
key="selected_model",
|
|
|
|
| 5 |
from openai import OpenAI
|
| 6 |
import json
|
| 7 |
|
|
|
|
|
|
|
|
|
|
| 8 |
def clear_chat():
|
| 9 |
st.session_state.messages = []
|
| 10 |
|
| 11 |
st.title("Intel® AI for Enterprise Inference")
|
| 12 |
st.header("LLM chatbot")
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
with st.sidebar:
|
| 15 |
+
api_key = st.session_state.api_key = st.secrets["openai_apikey"] #Enter openai_api key under "Secrets " in HF settings
|
| 16 |
+
base_url = st.session_state.base_url = os.environ.get("base_url") #Enter base_url under "Variables" in HF settings
|
|
|
|
|
|
|
| 17 |
client = OpenAI(api_key=api_key, base_url=base_url)
|
| 18 |
models = client.models.list()
|
| 19 |
model_names = sorted([model.id for model in models]) # Extract 'id' from each model object
|
|
|
|
| 20 |
default_model_name = "meta-llama/Llama-3.3-70B-Instruct" # Replace with your desired default model name
|
| 21 |
|
|
|
|
| 22 |
# Use st.session_state to persist the selected model
|
| 23 |
if "selected_model" not in st.session_state:
|
|
|
|
| 24 |
st.session_state.selected_model = default_model_name if default_model_name in model_names else model_names[0]
|
| 25 |
|
| 26 |
modelname = st.selectbox(
|
| 27 |
+
"Select an LLM model (Running on Intel® Gaudi®). Hosted on Denvr Dataworks",
|
| 28 |
model_names,
|
| 29 |
index=model_names.index(st.session_state.selected_model) if st.session_state.selected_model in model_names else 0,
|
| 30 |
key="selected_model",
|
model_info.json
DELETED
|
@@ -1,18 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"meta-llama/Llama-3.3-70B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 3 |
-
"meta-llama/Meta-Llama-3.2-1B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 4 |
-
"meta-llama/Meta-Llama-3.2-3B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 5 |
-
"meta-llama/Meta-Llama-3.1-8B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 6 |
-
"meta-llama/Meta-Llama-3.1-70B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 7 |
-
"deepseek-ai/DeepSeek-R1-Distill-Llama-8B": "https://api.inference.denvrdata.com/v1/",
|
| 8 |
-
"deepseek-ai/DeepSeek-R1-Distill-Llama-70B": "https://api.inference.denvrdata.com/v1/",
|
| 9 |
-
"mistralai/Mistral-7B-Instruct-v0.2": "https://api.inference.denvrdata.com/v1/",
|
| 10 |
-
"mistralai/Mistral-7B-Instruct-v0.1": "https://api.inference.denvrdata.com/v1/",
|
| 11 |
-
"mistralai/Mixtral-8x7B-Instruct-v0.1": "https://api.inference.denvrdata.com/v1/",
|
| 12 |
-
"tiiuae/Falcon3-7B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 13 |
-
"tiiuae/Falcon3-10B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 14 |
-
"tiiuae/Falcon3.1-74B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
| 15 |
-
"ALLaM/ALLaM-7B-Instruct-preview": "https://api.inference.denvrdata.com/v1/",
|
| 16 |
-
"inceptionai/jais-adapted-70b-chat": "https://api.inference.denvrdata.com/v1/",
|
| 17 |
-
"codellama/CodeLlama-34b-Instruct-hf": "https://api.inference.denvrdata.com/v1/"
|
| 18 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|