|
19 | 19 | from openai import OpenAI, AsyncOpenAI, APIResponseValidationError
|
20 | 20 | from openai._client import OpenAI, AsyncOpenAI
|
21 | 21 | from openai._models import BaseModel, FinalRequestOptions
|
22 |
| -from openai._response import APIResponse, AsyncAPIResponse |
23 | 22 | from openai._constants import RAW_RESPONSE_HEADER
|
24 | 23 | from openai._streaming import Stream, AsyncStream
|
25 | 24 | from openai._exceptions import OpenAIError, APIStatusError, APITimeoutError, APIResponseValidationError
|
@@ -665,33 +664,6 @@ def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str
|
665 | 664 | calculated = client._calculate_retry_timeout(remaining_retries, options, headers)
|
666 | 665 | assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType]
|
667 | 666 |
|
668 |
| - @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) |
669 |
| - @pytest.mark.respx(base_url=base_url) |
670 |
| - def test_streaming_response(self) -> None: |
671 |
| - response = self.client.post( |
672 |
| - "/chat/completions", |
673 |
| - body=dict( |
674 |
| - messages=[ |
675 |
| - { |
676 |
| - "role": "user", |
677 |
| - "content": "Say this is a test", |
678 |
| - } |
679 |
| - ], |
680 |
| - model="gpt-3.5-turbo", |
681 |
| - ), |
682 |
| - cast_to=APIResponse[bytes], |
683 |
| - options={"headers": {RAW_RESPONSE_HEADER: "stream"}}, |
684 |
| - ) |
685 |
| - |
686 |
| - assert not cast(Any, response.is_closed) |
687 |
| - assert _get_open_connections(self.client) == 1 |
688 |
| - |
689 |
| - for _ in response.iter_bytes(): |
690 |
| - ... |
691 |
| - |
692 |
| - assert cast(Any, response.is_closed) |
693 |
| - assert _get_open_connections(self.client) == 0 |
694 |
| - |
695 | 667 | @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout)
|
696 | 668 | @pytest.mark.respx(base_url=base_url)
|
697 | 669 | def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> None:
|
@@ -1372,33 +1344,6 @@ async def test_parse_retry_after_header(self, remaining_retries: int, retry_afte
|
1372 | 1344 | calculated = client._calculate_retry_timeout(remaining_retries, options, headers)
|
1373 | 1345 | assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType]
|
1374 | 1346 |
|
1375 |
| - @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) |
1376 |
| - @pytest.mark.respx(base_url=base_url) |
1377 |
| - async def test_streaming_response(self) -> None: |
1378 |
| - response = await self.client.post( |
1379 |
| - "/chat/completions", |
1380 |
| - body=dict( |
1381 |
| - messages=[ |
1382 |
| - { |
1383 |
| - "role": "user", |
1384 |
| - "content": "Say this is a test", |
1385 |
| - } |
1386 |
| - ], |
1387 |
| - model="gpt-3.5-turbo", |
1388 |
| - ), |
1389 |
| - cast_to=AsyncAPIResponse[bytes], |
1390 |
| - options={"headers": {RAW_RESPONSE_HEADER: "stream"}}, |
1391 |
| - ) |
1392 |
| - |
1393 |
| - assert not cast(Any, response.is_closed) |
1394 |
| - assert _get_open_connections(self.client) == 1 |
1395 |
| - |
1396 |
| - async for _ in response.iter_bytes(): |
1397 |
| - ... |
1398 |
| - |
1399 |
| - assert cast(Any, response.is_closed) |
1400 |
| - assert _get_open_connections(self.client) == 0 |
1401 |
| - |
1402 | 1347 | @mock.patch("openai._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout)
|
1403 | 1348 | @pytest.mark.respx(base_url=base_url)
|
1404 | 1349 | async def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter) -> None:
|
|
0 commit comments