From 802cf1ce21ebf90d46c6b8f2916e327b7bc2c9fc Mon Sep 17 00:00:00 2001 From: H Lohaus Date: Sun, 7 Apr 2024 00:07:58 +0200 Subject: [PATCH] Update Llama2.py --- g4f/Provider/Llama2.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/g4f/Provider/Llama2.py b/g4f/Provider/Llama2.py index 6a94eea1..04b5aee0 100644 --- a/g4f/Provider/Llama2.py +++ b/g4f/Provider/Llama2.py @@ -3,6 +3,7 @@ from __future__ import annotations from aiohttp import ClientSession from ..typing import AsyncResult, Messages +from ..requests.raise_for_status import raise_for_status from .base_provider import AsyncGeneratorProvider, ProviderModelMixin @@ -67,8 +68,10 @@ class Llama2(AsyncGeneratorProvider, ProviderModelMixin): } started = False async with session.post(f"{cls.url}/api", json=data, proxy=proxy) as response: - response.raise_for_status() + await raise_for_status(response) async for chunk in response.content.iter_any(): + if not chunk: + continue if not started: chunk = chunk.lstrip() started = True