File size: 966 Bytes
4360b77
 
 
 
abc2647
4360b77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f7ebb99
4360b77
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
import gradio as gr
from transformers import pipeline

# Load the model locally
MODEL = "Intelligent-Internet/II-Medical-8B"

print("🔄 Loading medical model locally...")
medical_tutor = pipeline(
    "text-generation",
    model=MODEL,
    device=-1,  # Use CPU
    torch_dtype="auto"
)
print("✅ Model loaded!")

def chat(message, history):
    # Simple medical tutoring prompt
    prompt = f"""You are a medical tutor. Provide educational information about: {message}

Remember: This is for learning purposes only, not medical advice.
Answer:"""
    
    response = medical_tutor(
        prompt,
        max_new_tokens=150,
        temperature=0.7,
        do_sample=True,
        pad_token_id=medical_tutor.tokenizer.eos_token_id
    )[0]['generated_text']
    
    # Extract just the answer part
    answer = response.split("Answer:")[-1].strip()
    return answer

gr.ChatInterface(
    chat,
    title="🩺 Medical Tutor",
    
).launch(server_port=7860)