gpt / gpt4free /g4f /Provider /needs_auth /GithubCopilot.py
LoRDxdd's picture
Add gpt4free API for Hugging Face
a4b70d9
from __future__ import annotations
import json
from aiohttp import ClientSession
from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin, BaseConversation
from ...typing import AsyncResult, Messages, Cookies
from ...requests.raise_for_status import raise_for_status
from ...requests.aiohttp import get_connector
from ...providers.helper import format_prompt, get_last_user_message
from ...cookies import get_cookies
class Conversation(BaseConversation):
conversation_id: str
def __init__(self, conversation_id: str):
self.conversation_id = conversation_id
class GithubCopilot(AsyncGeneratorProvider, ProviderModelMixin):
label = "GitHub Copilot"
url = "https://github.com/copilot"
working = True
needs_auth = True
supports_stream = True
default_model = "gpt-4.1"
models = [
# Fast and cost-efficient
"o3-mini",
"gemini-2.0-flash",
"o4-mini", # Preview
# Versatile and highly intelligent
"gpt-4.1",
"gpt-5-mini", # Preview
"gpt-4o",
"claude-3.5-sonnet",
"gemini-2.5-pro",
"claude-3.7-sonnet",
"claude-4-sonnet",
"o3", # Preview
"gpt-5", # Preview
# Most powerful at complex tasks
"claude-3.7-sonnet-thinking",
"claude-4-opus",
# Preview (requires upgrade)
"claude-3.7-sonnet-pro",
"o1",
# Legacy models (for backward compatibility)
"o1-mini",
"o1-preview"
]
@classmethod
async def create_async_generator(
cls,
model: str,
messages: Messages,
stream: bool = True,
api_key: str = None,
proxy: str = None,
cookies: Cookies = None,
conversation_id: str = None,
conversation: Conversation = None,
return_conversation: bool = True,
**kwargs
) -> AsyncResult:
if not model:
model = cls.default_model
if cookies is None:
cookies = get_cookies("github.com")
async with ClientSession(
connector=get_connector(proxy=proxy),
cookies=cookies,
headers={
'User-Agent': 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:133.0) Gecko/20100101 Firefox/133.0',
'Accept': 'application/json',
'Accept-Language': 'en-US,en;q=0.5',
'Referer': 'https://github.com/copilot',
'Content-Type': 'application/json',
'GitHub-Verified-Fetch': 'true',
'X-Requested-With': 'XMLHttpRequest',
'Origin': 'https://github.com',
'Connection': 'keep-alive',
'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-origin',
'Priority': 'u=1'
}
) as session:
headers = {}
if api_key is None:
async with session.post("https://github.com/github-copilot/chat/token") as response:
await raise_for_status(response, "Get token")
api_key = (await response.json()).get("token")
headers = {
"Authorization": f"GitHub-Bearer {api_key}",
}
if conversation is not None:
conversation_id = conversation.conversation_id
if conversation_id is None:
async with session.post(
"https://api.individual.githubcopilot.com/github/chat/threads",
headers=headers
) as response:
await raise_for_status(response)
conversation_id = (await response.json()).get("thread_id")
if return_conversation:
yield Conversation(conversation_id)
content = get_last_user_message(messages)
else:
content = format_prompt(messages)
json_data = {
"content": content,
"intent": "conversation",
"references": [],
"context": [],
"currentURL": f"https://github.com/copilot/c/{conversation_id}",
"streaming": stream,
"confirmations": [],
"customInstructions": [],
"model": api_model,
"mode": "immersive"
}
async with session.post(
f"https://api.individual.githubcopilot.com/github/chat/threads/{conversation_id}/messages",
json=json_data,
headers=headers
) as response:
await raise_for_status(response, f"Send message with model {model}")
if stream:
async for line in response.content:
if line.startswith(b"data: "):
try:
data = json.loads(line[6:])
if data.get("type") == "content":
content = data.get("body", "")
if content:
yield content
except json.JSONDecodeError:
continue
else:
full_content = ""
async for line in response.content:
if line.startswith(b"data: "):
try:
data = json.loads(line[6:])
if data.get("type") == "content":
full_content += data.get("body", "")
except json.JSONDecodeError:
continue
if full_content:
yield full_content