mirror of https://github.com/xtekky/gpt4free.git
Add unittests for async client (#1830)
* Add unittests for async client * Add pollyfill for anext * Update integration tests
This commit is contained in:
parent
35179fe843
commit
0b712c2bde
|
@ -4,6 +4,7 @@ from .backend import *
|
||||||
from .main import *
|
from .main import *
|
||||||
from .model import *
|
from .model import *
|
||||||
from .client import *
|
from .client import *
|
||||||
|
from .async_client import *
|
||||||
from .include import *
|
from .include import *
|
||||||
from .integration import *
|
from .integration import *
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,56 @@
|
||||||
|
import unittest
|
||||||
|
|
||||||
|
from g4f.client import AsyncClient, ChatCompletion, ChatCompletionChunk
|
||||||
|
from .mocks import AsyncGeneratorProviderMock, ModelProviderMock, YieldProviderMock
|
||||||
|
|
||||||
|
DEFAULT_MESSAGES = [{'role': 'user', 'content': 'Hello'}]
|
||||||
|
|
||||||
|
class AsyncTestPassModel(unittest.IsolatedAsyncioTestCase):
|
||||||
|
|
||||||
|
async def test_response(self):
|
||||||
|
client = AsyncClient(provider=AsyncGeneratorProviderMock)
|
||||||
|
response = await client.chat.completions.create(DEFAULT_MESSAGES, "")
|
||||||
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
|
self.assertEqual("Mock", response.choices[0].message.content)
|
||||||
|
|
||||||
|
async def test_pass_model(self):
|
||||||
|
client = AsyncClient(provider=ModelProviderMock)
|
||||||
|
response = await client.chat.completions.create(DEFAULT_MESSAGES, "Hello")
|
||||||
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
|
self.assertEqual("Hello", response.choices[0].message.content)
|
||||||
|
|
||||||
|
async def test_max_tokens(self):
|
||||||
|
client = AsyncClient(provider=YieldProviderMock)
|
||||||
|
messages = [{'role': 'user', 'content': chunk} for chunk in ["How ", "are ", "you", "?"]]
|
||||||
|
response = await client.chat.completions.create(messages, "Hello", max_tokens=1)
|
||||||
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
|
self.assertEqual("How ", response.choices[0].message.content)
|
||||||
|
response = await client.chat.completions.create(messages, "Hello", max_tokens=2)
|
||||||
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
|
self.assertEqual("How are ", response.choices[0].message.content)
|
||||||
|
|
||||||
|
async def test_max_stream(self):
|
||||||
|
client = AsyncClient(provider=YieldProviderMock)
|
||||||
|
messages = [{'role': 'user', 'content': chunk} for chunk in ["How ", "are ", "you", "?"]]
|
||||||
|
response = client.chat.completions.create(messages, "Hello", stream=True)
|
||||||
|
async for chunk in response:
|
||||||
|
self.assertIsInstance(chunk, ChatCompletionChunk)
|
||||||
|
if chunk.choices[0].delta.content is not None:
|
||||||
|
self.assertIsInstance(chunk.choices[0].delta.content, str)
|
||||||
|
messages = [{'role': 'user', 'content': chunk} for chunk in ["You ", "You ", "Other", "?"]]
|
||||||
|
response = client.chat.completions.create(messages, "Hello", stream=True, max_tokens=2)
|
||||||
|
response = [chunk async for chunk in response]
|
||||||
|
self.assertEqual(len(response), 3)
|
||||||
|
for chunk in response:
|
||||||
|
if chunk.choices[0].delta.content is not None:
|
||||||
|
self.assertEqual(chunk.choices[0].delta.content, "You ")
|
||||||
|
|
||||||
|
async def test_stop(self):
|
||||||
|
client = AsyncClient(provider=YieldProviderMock)
|
||||||
|
messages = [{'role': 'user', 'content': chunk} for chunk in ["How ", "are ", "you", "?"]]
|
||||||
|
response = await client.chat.completions.create(messages, "Hello", stop=["and"])
|
||||||
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
|
self.assertEqual("How are you?", response.choices[0].message.content)
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
unittest.main()
|
|
@ -8,7 +8,7 @@ except ImportError:
|
||||||
has_nest_asyncio = False
|
has_nest_asyncio = False
|
||||||
|
|
||||||
from g4f.client import Client, ChatCompletion
|
from g4f.client import Client, ChatCompletion
|
||||||
from g4f.Provider import Bing, OpenaiChat
|
from g4f.Provider import Bing, OpenaiChat, DuckDuckGo
|
||||||
|
|
||||||
DEFAULT_MESSAGES = [{"role": "system", "content": 'Response in json, Example: {"success: true"}'},
|
DEFAULT_MESSAGES = [{"role": "system", "content": 'Response in json, Example: {"success: true"}'},
|
||||||
{"role": "user", "content": "Say success true in json"}]
|
{"role": "user", "content": "Say success true in json"}]
|
||||||
|
@ -19,11 +19,19 @@ class TestProviderIntegration(unittest.TestCase):
|
||||||
self.skipTest("nest_asyncio is not installed")
|
self.skipTest("nest_asyncio is not installed")
|
||||||
|
|
||||||
def test_bing(self):
|
def test_bing(self):
|
||||||
|
self.skipTest("Not stable")
|
||||||
client = Client(provider=Bing)
|
client = Client(provider=Bing)
|
||||||
response = client.chat.completions.create(DEFAULT_MESSAGES, "", response_format={"type": "json_object"})
|
response = client.chat.completions.create(DEFAULT_MESSAGES, "", response_format={"type": "json_object"})
|
||||||
self.assertIsInstance(response, ChatCompletion)
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
self.assertIn("success", json.loads(response.choices[0].message.content))
|
self.assertIn("success", json.loads(response.choices[0].message.content))
|
||||||
|
|
||||||
|
def test_duckduckgo(self):
|
||||||
|
self.skipTest("Not working")
|
||||||
|
client = Client(provider=DuckDuckGo)
|
||||||
|
response = client.chat.completions.create(DEFAULT_MESSAGES, "", response_format={"type": "json_object"})
|
||||||
|
self.assertIsInstance(response, ChatCompletion)
|
||||||
|
self.assertIn("success", json.loads(response.choices[0].message.content))
|
||||||
|
|
||||||
def test_openai(self):
|
def test_openai(self):
|
||||||
client = Client(provider=OpenaiChat)
|
client = Client(provider=OpenaiChat)
|
||||||
response = client.chat.completions.create(DEFAULT_MESSAGES, "", response_format={"type": "json_object"})
|
response = client.chat.completions.create(DEFAULT_MESSAGES, "", response_format={"type": "json_object"})
|
||||||
|
|
|
@ -11,7 +11,7 @@ except ImportError:
|
||||||
from ..typing import Messages, CreateResult
|
from ..typing import Messages, CreateResult
|
||||||
from .base_provider import AbstractProvider
|
from .base_provider import AbstractProvider
|
||||||
from ..requests import raise_for_status
|
from ..requests import raise_for_status
|
||||||
from ..errors import MissingRequirementsError, RateLimitError, ResponseStatusError
|
from ..errors import MissingRequirementsError
|
||||||
|
|
||||||
class Vercel(AbstractProvider):
|
class Vercel(AbstractProvider):
|
||||||
url = 'https://chat.vercel.ai'
|
url = 'https://chat.vercel.ai'
|
||||||
|
|
|
@ -16,6 +16,13 @@ from ..errors import NoImageResponseError
|
||||||
from ..image import ImageResponse as ImageProviderResponse
|
from ..image import ImageResponse as ImageProviderResponse
|
||||||
from ..providers.base_provider import AsyncGeneratorProvider
|
from ..providers.base_provider import AsyncGeneratorProvider
|
||||||
|
|
||||||
|
try:
|
||||||
|
anext
|
||||||
|
except NameError:
|
||||||
|
async def anext(iter):
|
||||||
|
async for chunk in iter:
|
||||||
|
return chunk
|
||||||
|
|
||||||
async def iter_response(
|
async def iter_response(
|
||||||
response: AsyncIterator[str],
|
response: AsyncIterator[str],
|
||||||
stream: bool,
|
stream: bool,
|
||||||
|
|
|
@ -1262,7 +1262,7 @@ if (SpeechRecognition) {
|
||||||
|
|
||||||
function may_stop() {
|
function may_stop() {
|
||||||
if (microLabel.classList.contains("recognition")) {
|
if (microLabel.classList.contains("recognition")) {
|
||||||
//recognition.stop();
|
recognition.stop();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1272,15 +1272,12 @@ if (SpeechRecognition) {
|
||||||
recognition.onstart = function() {
|
recognition.onstart = function() {
|
||||||
microLabel.classList.add("recognition");
|
microLabel.classList.add("recognition");
|
||||||
startValue = messageInput.value;
|
startValue = messageInput.value;
|
||||||
messageInput.placeholder = "";
|
|
||||||
lastDebounceTranscript = "";
|
lastDebounceTranscript = "";
|
||||||
timeoutHandle = window.setTimeout(may_stop, 10000);
|
timeoutHandle = window.setTimeout(may_stop, 10000);
|
||||||
};
|
};
|
||||||
recognition.onend = function() {
|
recognition.onend = function() {
|
||||||
microLabel.classList.remove("recognition");
|
microLabel.classList.remove("recognition");
|
||||||
messageInput.value = messageInput.placeholder;
|
messageInput.focus();
|
||||||
messageInput.placeholder = "Ask a question";
|
|
||||||
//messageInput.focus();
|
|
||||||
};
|
};
|
||||||
recognition.onresult = function(event) {
|
recognition.onresult = function(event) {
|
||||||
if (!event.results) {
|
if (!event.results) {
|
||||||
|
@ -1298,9 +1295,9 @@ if (SpeechRecognition) {
|
||||||
lastDebounceTranscript = transcript;
|
lastDebounceTranscript = transcript;
|
||||||
}
|
}
|
||||||
if (transcript) {
|
if (transcript) {
|
||||||
messageInput.placeholder = `${startValue ? startValue+"\n" : ""}${transcript.trim()}`;
|
messageInput.value = `${startValue ? startValue+"\n" : ""}${transcript.trim()}`;
|
||||||
if (isFinal) {
|
if (isFinal) {
|
||||||
startValue = messageInput.placeholder;
|
startValue = messageInput.value;
|
||||||
}
|
}
|
||||||
messageInput.style.height = messageInput.scrollHeight + "px";
|
messageInput.style.height = messageInput.scrollHeight + "px";
|
||||||
messageInput.scrollTop = messageInput.scrollHeight;
|
messageInput.scrollTop = messageInput.scrollHeight;
|
||||||
|
|
|
@ -34,15 +34,13 @@ class StreamResponse:
|
||||||
"""Asynchronously parse the JSON response content."""
|
"""Asynchronously parse the JSON response content."""
|
||||||
return json.loads(await self.inner.acontent(), **kwargs)
|
return json.loads(await self.inner.acontent(), **kwargs)
|
||||||
|
|
||||||
async def iter_lines(self) -> AsyncGenerator[bytes, None]:
|
def iter_lines(self) -> AsyncGenerator[bytes, None]:
|
||||||
"""Asynchronously iterate over the lines of the response."""
|
"""Asynchronously iterate over the lines of the response."""
|
||||||
async for line in self.inner.aiter_lines():
|
return self.inner.aiter_lines()
|
||||||
yield line
|
|
||||||
|
|
||||||
async def iter_content(self) -> AsyncGenerator[bytes, None]:
|
def iter_content(self) -> AsyncGenerator[bytes, None]:
|
||||||
"""Asynchronously iterate over the response content."""
|
"""Asynchronously iterate over the response content."""
|
||||||
async for chunk in self.inner.aiter_content():
|
return self.inner.aiter_content()
|
||||||
yield chunk
|
|
||||||
|
|
||||||
async def __aenter__(self):
|
async def __aenter__(self):
|
||||||
"""Asynchronously enter the runtime context for the response object."""
|
"""Asynchronously enter the runtime context for the response object."""
|
||||||
|
|
Loading…
Reference in New Issue