~ | add `g4f.Provider.GeekGpt`

This commit is contained in:
abc 2023-10-16 14:34:00 +01:00
parent 8bce9ace2d
commit 5b240665fb
2 changed files with 85 additions and 1 deletions

81
g4f/Provider/Geekgpt.py Normal file
View File

@ -0,0 +1,81 @@
from __future__ import annotations
import requests, json
from .base_provider import BaseProvider
from ..typing import CreateResult, Messages
from json import dumps
class GeekGpt(BaseProvider):
url = 'https://chat.geekgpt.org'
supports_stream = True
working = True
supports_gpt_35_turbo = True
supports_gpt_4 = True
@classmethod
def create_completion(cls,
model: str,
messages: Messages,
stream: bool, **kwargs) -> CreateResult:
json_data = {
'messages': messages,
'model': model,
'temperature': kwargs.get('temperature', 0.9),
'presence_penalty': kwargs.get('presence_penalty', 0),
'top_p': kwargs.get('top_p', 1),
'frequency_penalty': kwargs.get('frequency_penalty', 0),
'stream': True
}
data = dumps(json_data, separators=(',', ':'))
headers = {
'authority': 'ai.fakeopen.com',
'accept': '*/*',
'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
'authorization': 'Bearer pk-this-is-a-real-free-pool-token-for-everyone',
'content-type': 'application/json',
'origin': 'https://chat.geekgpt.org',
'referer': 'https://chat.geekgpt.org/',
'sec-ch-ua': '"Chromium";v="118", "Google Chrome";v="118", "Not=A?Brand";v="99"',
'sec-ch-ua-mobile': '?0',
'sec-ch-ua-platform': '"macOS"',
'sec-fetch-dest': 'empty',
'sec-fetch-mode': 'cors',
'sec-fetch-site': 'cross-site',
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/118.0.0.0 Safari/537.36',
}
response = requests.post("https://ai.fakeopen.com/v1/chat/completions",
headers=headers, data=data, stream=True)
response.raise_for_status()
for chunk in response.iter_lines():
if b'content' in chunk:
json_data = chunk.decode().replace("data: ", "")
if json_data == "[DONE]":
break
try:
content = json.loads(json_data)["choices"][0]["delta"].get("content")
except Exception as e:
raise RuntimeError(f'error | {e} :', json_data)
if content:
yield content
@classmethod
@property
def params(cls):
params = [
('model', 'str'),
('messages', 'list[dict[str, str]]'),
('stream', 'bool'),
('temperature', 'float'),
]
param = ', '.join([': '.join(p) for p in params])
return f'g4f.provider.{cls.__name__} supports: ({param})'

View File

@ -33,6 +33,7 @@ from .Vercel import Vercel
from .Ylokh import Ylokh
from .You import You
from .Yqcloud import Yqcloud
from .Geekgpt import GeekGpt
from .base_provider import BaseProvider, AsyncProvider, AsyncGeneratorProvider
from .retry_provider import RetryProvider
@ -103,6 +104,7 @@ class ProviderUtils:
'Ylokh': Ylokh,
'You': You,
'Yqcloud': Yqcloud,
'Geekgpt': GeekGpt,
'BaseProvider': BaseProvider,
'AsyncProvider': AsyncProvider,
@ -171,5 +173,6 @@ __all__ = [
'Equing',
'FastGpt',
'Wuguokai',
'V50'
'V50',
'GeekGpt'
]