working qwen
Browse files- app.py +3 -1
- app_qwen.py +13 -13
- requirements.txt +1 -2
app.py
CHANGED
|
@@ -24,10 +24,12 @@ from app_showui import demo as demo_showui
|
|
| 24 |
from app_together import demo as demo_together
|
| 25 |
from app_xai import demo as demo_grok
|
| 26 |
from app_openai_voice import demo as demo_openai_voice
|
|
|
|
| 27 |
from utils import get_app
|
| 28 |
|
| 29 |
# Create mapping of providers to their demos
|
| 30 |
PROVIDERS = {
|
|
|
|
| 31 |
"Gemini": demo_gemini,
|
| 32 |
"OpenAI Voice": demo_openai_voice,
|
| 33 |
"Gemini Voice": demo_gemini_voice,
|
|
@@ -56,7 +58,7 @@ PROVIDERS = {
|
|
| 56 |
"NVIDIA": demo_nvidia,
|
| 57 |
}
|
| 58 |
|
| 59 |
-
demo = get_app(models=list(PROVIDERS.keys()), default_model="
|
| 60 |
|
| 61 |
if __name__ == "__main__":
|
| 62 |
demo.queue(api_open=False).launch(show_api=False)
|
|
|
|
| 24 |
from app_together import demo as demo_together
|
| 25 |
from app_xai import demo as demo_grok
|
| 26 |
from app_openai_voice import demo as demo_openai_voice
|
| 27 |
+
from app_qwen import demo as demo_qwen
|
| 28 |
from utils import get_app
|
| 29 |
|
| 30 |
# Create mapping of providers to their demos
|
| 31 |
PROVIDERS = {
|
| 32 |
+
"Qwen" : demo_qwen,
|
| 33 |
"Gemini": demo_gemini,
|
| 34 |
"OpenAI Voice": demo_openai_voice,
|
| 35 |
"Gemini Voice": demo_gemini_voice,
|
|
|
|
| 58 |
"NVIDIA": demo_nvidia,
|
| 59 |
}
|
| 60 |
|
| 61 |
+
demo = get_app(models=list(PROVIDERS.keys()), default_model="Qwen", src=PROVIDERS, dropdown_label="Select Provider")
|
| 62 |
|
| 63 |
if __name__ == "__main__":
|
| 64 |
demo.queue(api_open=False).launch(show_api=False)
|
app_qwen.py
CHANGED
|
@@ -1,24 +1,24 @@
|
|
| 1 |
import os
|
| 2 |
|
| 3 |
-
import
|
| 4 |
|
| 5 |
from utils import get_app
|
| 6 |
|
| 7 |
demo = get_app(
|
| 8 |
models=[
|
| 9 |
-
"qwen
|
| 10 |
-
"qwen
|
| 11 |
-
"qwen
|
| 12 |
-
"qwen
|
| 13 |
-
"
|
| 14 |
-
"
|
| 15 |
-
"
|
| 16 |
-
"
|
| 17 |
-
"
|
| 18 |
-
"
|
| 19 |
-
"
|
| 20 |
],
|
| 21 |
-
default_model="
|
| 22 |
src=ai_gradio.registry,
|
| 23 |
accept_token=not os.getenv("DASHSCOPE_API_KEY"),
|
| 24 |
)
|
|
|
|
| 1 |
import os
|
| 2 |
|
| 3 |
+
import dashscope_gradio
|
| 4 |
|
| 5 |
from utils import get_app
|
| 6 |
|
| 7 |
demo = get_app(
|
| 8 |
models=[
|
| 9 |
+
"qwen-turbo-latest",
|
| 10 |
+
"qwen-turbo",
|
| 11 |
+
"qwen-plus",
|
| 12 |
+
"qwen-max",
|
| 13 |
+
"qwen1.5-110b-chat",
|
| 14 |
+
"qwen1.5-72b-chat",
|
| 15 |
+
"qwen1.5-32b-chat",
|
| 16 |
+
"qwen1.5-14b-chat",
|
| 17 |
+
"qwen1.5-7b-chat",
|
| 18 |
+
"qwq-32b-preview",
|
| 19 |
+
"qvq-72b-preview",
|
| 20 |
],
|
| 21 |
+
default_model="qvq-72b-preview",
|
| 22 |
src=ai_gradio.registry,
|
| 23 |
accept_token=not os.getenv("DASHSCOPE_API_KEY"),
|
| 24 |
)
|
requirements.txt
CHANGED
|
@@ -72,7 +72,7 @@ cryptography==44.0.0
|
|
| 72 |
# via
|
| 73 |
# aiortc
|
| 74 |
# pyopenssl
|
| 75 |
-
dashscope-gradio @ git+https://github.com/AK391/dashscope-gradio.git@
|
| 76 |
# via anychat (pyproject.toml)
|
| 77 |
decorator==5.1.1
|
| 78 |
# via librosa
|
|
@@ -513,4 +513,3 @@ xai-gradio==0.0.2
|
|
| 513 |
# via anychat (pyproject.toml)
|
| 514 |
yarl==1.18.3
|
| 515 |
# via aiohttp
|
| 516 |
-
ai-gradio==0.1.5
|
|
|
|
| 72 |
# via
|
| 73 |
# aiortc
|
| 74 |
# pyopenssl
|
| 75 |
+
dashscope-gradio @ git+https://github.com/AK391/dashscope-gradio.git@afe2a389fd194e71ca4c12e6162f3f0fd3cf534a
|
| 76 |
# via anychat (pyproject.toml)
|
| 77 |
decorator==5.1.1
|
| 78 |
# via librosa
|
|
|
|
| 513 |
# via anychat (pyproject.toml)
|
| 514 |
yarl==1.18.3
|
| 515 |
# via aiohttp
|
|
|