aboutsummaryrefslogtreecommitdiff
path: root/.venv/lib/python3.12/site-packages/openai/_client.py
diff options
context:
space:
mode:
authorS. Solomon Darnell2025-03-28 21:52:21 -0500
committerS. Solomon Darnell2025-03-28 21:52:21 -0500
commit4a52a71956a8d46fcb7294ac71734504bb09bcc2 (patch)
treeee3dc5af3b6313e921cd920906356f5d4febc4ed /.venv/lib/python3.12/site-packages/openai/_client.py
parentcc961e04ba734dd72309fb548a2f97d67d578813 (diff)
downloadgn-ai-4a52a71956a8d46fcb7294ac71734504bb09bcc2.tar.gz
two version of R2R are hereHEADmaster
Diffstat (limited to '.venv/lib/python3.12/site-packages/openai/_client.py')
-rw-r--r--.venv/lib/python3.12/site-packages/openai/_client.py583
1 files changed, 583 insertions, 0 deletions
diff --git a/.venv/lib/python3.12/site-packages/openai/_client.py b/.venv/lib/python3.12/site-packages/openai/_client.py
new file mode 100644
index 00000000..18d96da9
--- /dev/null
+++ b/.venv/lib/python3.12/site-packages/openai/_client.py
@@ -0,0 +1,583 @@
+# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
+
+from __future__ import annotations
+
+import os
+from typing import Any, Union, Mapping
+from typing_extensions import Self, override
+
+import httpx
+
+from . import _exceptions
+from ._qs import Querystring
+from ._types import (
+ NOT_GIVEN,
+ Omit,
+ Timeout,
+ NotGiven,
+ Transport,
+ ProxiesTypes,
+ RequestOptions,
+)
+from ._utils import (
+ is_given,
+ is_mapping,
+ get_async_library,
+)
+from ._version import __version__
+from .resources import files, images, models, batches, embeddings, completions, moderations
+from ._streaming import Stream as Stream, AsyncStream as AsyncStream
+from ._exceptions import OpenAIError, APIStatusError
+from ._base_client import (
+ DEFAULT_MAX_RETRIES,
+ SyncAPIClient,
+ AsyncAPIClient,
+)
+from .resources.beta import beta
+from .resources.chat import chat
+from .resources.audio import audio
+from .resources.uploads import uploads
+from .resources.responses import responses
+from .resources.fine_tuning import fine_tuning
+from .resources.vector_stores import vector_stores
+
+__all__ = ["Timeout", "Transport", "ProxiesTypes", "RequestOptions", "OpenAI", "AsyncOpenAI", "Client", "AsyncClient"]
+
+
+class OpenAI(SyncAPIClient):
+ completions: completions.Completions
+ chat: chat.Chat
+ embeddings: embeddings.Embeddings
+ files: files.Files
+ images: images.Images
+ audio: audio.Audio
+ moderations: moderations.Moderations
+ models: models.Models
+ fine_tuning: fine_tuning.FineTuning
+ vector_stores: vector_stores.VectorStores
+ beta: beta.Beta
+ batches: batches.Batches
+ uploads: uploads.Uploads
+ responses: responses.Responses
+ with_raw_response: OpenAIWithRawResponse
+ with_streaming_response: OpenAIWithStreamedResponse
+
+ # client options
+ api_key: str
+ organization: str | None
+ project: str | None
+
+ websocket_base_url: str | httpx.URL | None
+ """Base URL for WebSocket connections.
+
+ If not specified, the default base URL will be used, with 'wss://' replacing the
+ 'http://' or 'https://' scheme. For example: 'http://example.com' becomes
+ 'wss://example.com'
+ """
+
+ def __init__(
+ self,
+ *,
+ api_key: str | None = None,
+ organization: str | None = None,
+ project: str | None = None,
+ base_url: str | httpx.URL | None = None,
+ websocket_base_url: str | httpx.URL | None = None,
+ timeout: Union[float, Timeout, None, NotGiven] = NOT_GIVEN,
+ max_retries: int = DEFAULT_MAX_RETRIES,
+ default_headers: Mapping[str, str] | None = None,
+ default_query: Mapping[str, object] | None = None,
+ # Configure a custom httpx client.
+ # We provide a `DefaultHttpxClient` class that you can pass to retain the default values we use for `limits`, `timeout` & `follow_redirects`.
+ # See the [httpx documentation](https://www.python-httpx.org/api/#client) for more details.
+ http_client: httpx.Client | None = None,
+ # Enable or disable schema validation for data returned by the API.
+ # When enabled an error APIResponseValidationError is raised
+ # if the API responds with invalid data for the expected schema.
+ #
+ # This parameter may be removed or changed in the future.
+ # If you rely on this feature, please open a GitHub issue
+ # outlining your use-case to help us decide if it should be
+ # part of our public interface in the future.
+ _strict_response_validation: bool = False,
+ ) -> None:
+ """Construct a new synchronous OpenAI client instance.
+
+ This automatically infers the following arguments from their corresponding environment variables if they are not provided:
+ - `api_key` from `OPENAI_API_KEY`
+ - `organization` from `OPENAI_ORG_ID`
+ - `project` from `OPENAI_PROJECT_ID`
+ """
+ if api_key is None:
+ api_key = os.environ.get("OPENAI_API_KEY")
+ if api_key is None:
+ raise OpenAIError(
+ "The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable"
+ )
+ self.api_key = api_key
+
+ if organization is None:
+ organization = os.environ.get("OPENAI_ORG_ID")
+ self.organization = organization
+
+ if project is None:
+ project = os.environ.get("OPENAI_PROJECT_ID")
+ self.project = project
+
+ self.websocket_base_url = websocket_base_url
+
+ if base_url is None:
+ base_url = os.environ.get("OPENAI_BASE_URL")
+ if base_url is None:
+ base_url = f"https://api.openai.com/v1"
+
+ super().__init__(
+ version=__version__,
+ base_url=base_url,
+ max_retries=max_retries,
+ timeout=timeout,
+ http_client=http_client,
+ custom_headers=default_headers,
+ custom_query=default_query,
+ _strict_response_validation=_strict_response_validation,
+ )
+
+ self._default_stream_cls = Stream
+
+ self.completions = completions.Completions(self)
+ self.chat = chat.Chat(self)
+ self.embeddings = embeddings.Embeddings(self)
+ self.files = files.Files(self)
+ self.images = images.Images(self)
+ self.audio = audio.Audio(self)
+ self.moderations = moderations.Moderations(self)
+ self.models = models.Models(self)
+ self.fine_tuning = fine_tuning.FineTuning(self)
+ self.vector_stores = vector_stores.VectorStores(self)
+ self.beta = beta.Beta(self)
+ self.batches = batches.Batches(self)
+ self.uploads = uploads.Uploads(self)
+ self.responses = responses.Responses(self)
+ self.with_raw_response = OpenAIWithRawResponse(self)
+ self.with_streaming_response = OpenAIWithStreamedResponse(self)
+
+ @property
+ @override
+ def qs(self) -> Querystring:
+ return Querystring(array_format="brackets")
+
+ @property
+ @override
+ def auth_headers(self) -> dict[str, str]:
+ api_key = self.api_key
+ return {"Authorization": f"Bearer {api_key}"}
+
+ @property
+ @override
+ def default_headers(self) -> dict[str, str | Omit]:
+ return {
+ **super().default_headers,
+ "X-Stainless-Async": "false",
+ "OpenAI-Organization": self.organization if self.organization is not None else Omit(),
+ "OpenAI-Project": self.project if self.project is not None else Omit(),
+ **self._custom_headers,
+ }
+
+ def copy(
+ self,
+ *,
+ api_key: str | None = None,
+ organization: str | None = None,
+ project: str | None = None,
+ websocket_base_url: str | httpx.URL | None = None,
+ base_url: str | httpx.URL | None = None,
+ timeout: float | Timeout | None | NotGiven = NOT_GIVEN,
+ http_client: httpx.Client | None = None,
+ max_retries: int | NotGiven = NOT_GIVEN,
+ default_headers: Mapping[str, str] | None = None,
+ set_default_headers: Mapping[str, str] | None = None,
+ default_query: Mapping[str, object] | None = None,
+ set_default_query: Mapping[str, object] | None = None,
+ _extra_kwargs: Mapping[str, Any] = {},
+ ) -> Self:
+ """
+ Create a new client instance re-using the same options given to the current client with optional overriding.
+ """
+ if default_headers is not None and set_default_headers is not None:
+ raise ValueError("The `default_headers` and `set_default_headers` arguments are mutually exclusive")
+
+ if default_query is not None and set_default_query is not None:
+ raise ValueError("The `default_query` and `set_default_query` arguments are mutually exclusive")
+
+ headers = self._custom_headers
+ if default_headers is not None:
+ headers = {**headers, **default_headers}
+ elif set_default_headers is not None:
+ headers = set_default_headers
+
+ params = self._custom_query
+ if default_query is not None:
+ params = {**params, **default_query}
+ elif set_default_query is not None:
+ params = set_default_query
+
+ http_client = http_client or self._client
+ return self.__class__(
+ api_key=api_key or self.api_key,
+ organization=organization or self.organization,
+ project=project or self.project,
+ websocket_base_url=websocket_base_url or self.websocket_base_url,
+ base_url=base_url or self.base_url,
+ timeout=self.timeout if isinstance(timeout, NotGiven) else timeout,
+ http_client=http_client,
+ max_retries=max_retries if is_given(max_retries) else self.max_retries,
+ default_headers=headers,
+ default_query=params,
+ **_extra_kwargs,
+ )
+
+ # Alias for `copy` for nicer inline usage, e.g.
+ # client.with_options(timeout=10).foo.create(...)
+ with_options = copy
+
+ @override
+ def _make_status_error(
+ self,
+ err_msg: str,
+ *,
+ body: object,
+ response: httpx.Response,
+ ) -> APIStatusError:
+ data = body.get("error", body) if is_mapping(body) else body
+ if response.status_code == 400:
+ return _exceptions.BadRequestError(err_msg, response=response, body=data)
+
+ if response.status_code == 401:
+ return _exceptions.AuthenticationError(err_msg, response=response, body=data)
+
+ if response.status_code == 403:
+ return _exceptions.PermissionDeniedError(err_msg, response=response, body=data)
+
+ if response.status_code == 404:
+ return _exceptions.NotFoundError(err_msg, response=response, body=data)
+
+ if response.status_code == 409:
+ return _exceptions.ConflictError(err_msg, response=response, body=data)
+
+ if response.status_code == 422:
+ return _exceptions.UnprocessableEntityError(err_msg, response=response, body=data)
+
+ if response.status_code == 429:
+ return _exceptions.RateLimitError(err_msg, response=response, body=data)
+
+ if response.status_code >= 500:
+ return _exceptions.InternalServerError(err_msg, response=response, body=data)
+ return APIStatusError(err_msg, response=response, body=data)
+
+
+class AsyncOpenAI(AsyncAPIClient):
+ completions: completions.AsyncCompletions
+ chat: chat.AsyncChat
+ embeddings: embeddings.AsyncEmbeddings
+ files: files.AsyncFiles
+ images: images.AsyncImages
+ audio: audio.AsyncAudio
+ moderations: moderations.AsyncModerations
+ models: models.AsyncModels
+ fine_tuning: fine_tuning.AsyncFineTuning
+ vector_stores: vector_stores.AsyncVectorStores
+ beta: beta.AsyncBeta
+ batches: batches.AsyncBatches
+ uploads: uploads.AsyncUploads
+ responses: responses.AsyncResponses
+ with_raw_response: AsyncOpenAIWithRawResponse
+ with_streaming_response: AsyncOpenAIWithStreamedResponse
+
+ # client options
+ api_key: str
+ organization: str | None
+ project: str | None
+
+ websocket_base_url: str | httpx.URL | None
+ """Base URL for WebSocket connections.
+
+ If not specified, the default base URL will be used, with 'wss://' replacing the
+ 'http://' or 'https://' scheme. For example: 'http://example.com' becomes
+ 'wss://example.com'
+ """
+
+ def __init__(
+ self,
+ *,
+ api_key: str | None = None,
+ organization: str | None = None,
+ project: str | None = None,
+ base_url: str | httpx.URL | None = None,
+ websocket_base_url: str | httpx.URL | None = None,
+ timeout: Union[float, Timeout, None, NotGiven] = NOT_GIVEN,
+ max_retries: int = DEFAULT_MAX_RETRIES,
+ default_headers: Mapping[str, str] | None = None,
+ default_query: Mapping[str, object] | None = None,
+ # Configure a custom httpx client.
+ # We provide a `DefaultAsyncHttpxClient` class that you can pass to retain the default values we use for `limits`, `timeout` & `follow_redirects`.
+ # See the [httpx documentation](https://www.python-httpx.org/api/#asyncclient) for more details.
+ http_client: httpx.AsyncClient | None = None,
+ # Enable or disable schema validation for data returned by the API.
+ # When enabled an error APIResponseValidationError is raised
+ # if the API responds with invalid data for the expected schema.
+ #
+ # This parameter may be removed or changed in the future.
+ # If you rely on this feature, please open a GitHub issue
+ # outlining your use-case to help us decide if it should be
+ # part of our public interface in the future.
+ _strict_response_validation: bool = False,
+ ) -> None:
+ """Construct a new async AsyncOpenAI client instance.
+
+ This automatically infers the following arguments from their corresponding environment variables if they are not provided:
+ - `api_key` from `OPENAI_API_KEY`
+ - `organization` from `OPENAI_ORG_ID`
+ - `project` from `OPENAI_PROJECT_ID`
+ """
+ if api_key is None:
+ api_key = os.environ.get("OPENAI_API_KEY")
+ if api_key is None:
+ raise OpenAIError(
+ "The api_key client option must be set either by passing api_key to the client or by setting the OPENAI_API_KEY environment variable"
+ )
+ self.api_key = api_key
+
+ if organization is None:
+ organization = os.environ.get("OPENAI_ORG_ID")
+ self.organization = organization
+
+ if project is None:
+ project = os.environ.get("OPENAI_PROJECT_ID")
+ self.project = project
+
+ self.websocket_base_url = websocket_base_url
+
+ if base_url is None:
+ base_url = os.environ.get("OPENAI_BASE_URL")
+ if base_url is None:
+ base_url = f"https://api.openai.com/v1"
+
+ super().__init__(
+ version=__version__,
+ base_url=base_url,
+ max_retries=max_retries,
+ timeout=timeout,
+ http_client=http_client,
+ custom_headers=default_headers,
+ custom_query=default_query,
+ _strict_response_validation=_strict_response_validation,
+ )
+
+ self._default_stream_cls = AsyncStream
+
+ self.completions = completions.AsyncCompletions(self)
+ self.chat = chat.AsyncChat(self)
+ self.embeddings = embeddings.AsyncEmbeddings(self)
+ self.files = files.AsyncFiles(self)
+ self.images = images.AsyncImages(self)
+ self.audio = audio.AsyncAudio(self)
+ self.moderations = moderations.AsyncModerations(self)
+ self.models = models.AsyncModels(self)
+ self.fine_tuning = fine_tuning.AsyncFineTuning(self)
+ self.vector_stores = vector_stores.AsyncVectorStores(self)
+ self.beta = beta.AsyncBeta(self)
+ self.batches = batches.AsyncBatches(self)
+ self.uploads = uploads.AsyncUploads(self)
+ self.responses = responses.AsyncResponses(self)
+ self.with_raw_response = AsyncOpenAIWithRawResponse(self)
+ self.with_streaming_response = AsyncOpenAIWithStreamedResponse(self)
+
+ @property
+ @override
+ def qs(self) -> Querystring:
+ return Querystring(array_format="brackets")
+
+ @property
+ @override
+ def auth_headers(self) -> dict[str, str]:
+ api_key = self.api_key
+ return {"Authorization": f"Bearer {api_key}"}
+
+ @property
+ @override
+ def default_headers(self) -> dict[str, str | Omit]:
+ return {
+ **super().default_headers,
+ "X-Stainless-Async": f"async:{get_async_library()}",
+ "OpenAI-Organization": self.organization if self.organization is not None else Omit(),
+ "OpenAI-Project": self.project if self.project is not None else Omit(),
+ **self._custom_headers,
+ }
+
+ def copy(
+ self,
+ *,
+ api_key: str | None = None,
+ organization: str | None = None,
+ project: str | None = None,
+ websocket_base_url: str | httpx.URL | None = None,
+ base_url: str | httpx.URL | None = None,
+ timeout: float | Timeout | None | NotGiven = NOT_GIVEN,
+ http_client: httpx.AsyncClient | None = None,
+ max_retries: int | NotGiven = NOT_GIVEN,
+ default_headers: Mapping[str, str] | None = None,
+ set_default_headers: Mapping[str, str] | None = None,
+ default_query: Mapping[str, object] | None = None,
+ set_default_query: Mapping[str, object] | None = None,
+ _extra_kwargs: Mapping[str, Any] = {},
+ ) -> Self:
+ """
+ Create a new client instance re-using the same options given to the current client with optional overriding.
+ """
+ if default_headers is not None and set_default_headers is not None:
+ raise ValueError("The `default_headers` and `set_default_headers` arguments are mutually exclusive")
+
+ if default_query is not None and set_default_query is not None:
+ raise ValueError("The `default_query` and `set_default_query` arguments are mutually exclusive")
+
+ headers = self._custom_headers
+ if default_headers is not None:
+ headers = {**headers, **default_headers}
+ elif set_default_headers is not None:
+ headers = set_default_headers
+
+ params = self._custom_query
+ if default_query is not None:
+ params = {**params, **default_query}
+ elif set_default_query is not None:
+ params = set_default_query
+
+ http_client = http_client or self._client
+ return self.__class__(
+ api_key=api_key or self.api_key,
+ organization=organization or self.organization,
+ project=project or self.project,
+ websocket_base_url=websocket_base_url or self.websocket_base_url,
+ base_url=base_url or self.base_url,
+ timeout=self.timeout if isinstance(timeout, NotGiven) else timeout,
+ http_client=http_client,
+ max_retries=max_retries if is_given(max_retries) else self.max_retries,
+ default_headers=headers,
+ default_query=params,
+ **_extra_kwargs,
+ )
+
+ # Alias for `copy` for nicer inline usage, e.g.
+ # client.with_options(timeout=10).foo.create(...)
+ with_options = copy
+
+ @override
+ def _make_status_error(
+ self,
+ err_msg: str,
+ *,
+ body: object,
+ response: httpx.Response,
+ ) -> APIStatusError:
+ data = body.get("error", body) if is_mapping(body) else body
+ if response.status_code == 400:
+ return _exceptions.BadRequestError(err_msg, response=response, body=data)
+
+ if response.status_code == 401:
+ return _exceptions.AuthenticationError(err_msg, response=response, body=data)
+
+ if response.status_code == 403:
+ return _exceptions.PermissionDeniedError(err_msg, response=response, body=data)
+
+ if response.status_code == 404:
+ return _exceptions.NotFoundError(err_msg, response=response, body=data)
+
+ if response.status_code == 409:
+ return _exceptions.ConflictError(err_msg, response=response, body=data)
+
+ if response.status_code == 422:
+ return _exceptions.UnprocessableEntityError(err_msg, response=response, body=data)
+
+ if response.status_code == 429:
+ return _exceptions.RateLimitError(err_msg, response=response, body=data)
+
+ if response.status_code >= 500:
+ return _exceptions.InternalServerError(err_msg, response=response, body=data)
+ return APIStatusError(err_msg, response=response, body=data)
+
+
+class OpenAIWithRawResponse:
+ def __init__(self, client: OpenAI) -> None:
+ self.completions = completions.CompletionsWithRawResponse(client.completions)
+ self.chat = chat.ChatWithRawResponse(client.chat)
+ self.embeddings = embeddings.EmbeddingsWithRawResponse(client.embeddings)
+ self.files = files.FilesWithRawResponse(client.files)
+ self.images = images.ImagesWithRawResponse(client.images)
+ self.audio = audio.AudioWithRawResponse(client.audio)
+ self.moderations = moderations.ModerationsWithRawResponse(client.moderations)
+ self.models = models.ModelsWithRawResponse(client.models)
+ self.fine_tuning = fine_tuning.FineTuningWithRawResponse(client.fine_tuning)
+ self.vector_stores = vector_stores.VectorStoresWithRawResponse(client.vector_stores)
+ self.beta = beta.BetaWithRawResponse(client.beta)
+ self.batches = batches.BatchesWithRawResponse(client.batches)
+ self.uploads = uploads.UploadsWithRawResponse(client.uploads)
+ self.responses = responses.ResponsesWithRawResponse(client.responses)
+
+
+class AsyncOpenAIWithRawResponse:
+ def __init__(self, client: AsyncOpenAI) -> None:
+ self.completions = completions.AsyncCompletionsWithRawResponse(client.completions)
+ self.chat = chat.AsyncChatWithRawResponse(client.chat)
+ self.embeddings = embeddings.AsyncEmbeddingsWithRawResponse(client.embeddings)
+ self.files = files.AsyncFilesWithRawResponse(client.files)
+ self.images = images.AsyncImagesWithRawResponse(client.images)
+ self.audio = audio.AsyncAudioWithRawResponse(client.audio)
+ self.moderations = moderations.AsyncModerationsWithRawResponse(client.moderations)
+ self.models = models.AsyncModelsWithRawResponse(client.models)
+ self.fine_tuning = fine_tuning.AsyncFineTuningWithRawResponse(client.fine_tuning)
+ self.vector_stores = vector_stores.AsyncVectorStoresWithRawResponse(client.vector_stores)
+ self.beta = beta.AsyncBetaWithRawResponse(client.beta)
+ self.batches = batches.AsyncBatchesWithRawResponse(client.batches)
+ self.uploads = uploads.AsyncUploadsWithRawResponse(client.uploads)
+ self.responses = responses.AsyncResponsesWithRawResponse(client.responses)
+
+
+class OpenAIWithStreamedResponse:
+ def __init__(self, client: OpenAI) -> None:
+ self.completions = completions.CompletionsWithStreamingResponse(client.completions)
+ self.chat = chat.ChatWithStreamingResponse(client.chat)
+ self.embeddings = embeddings.EmbeddingsWithStreamingResponse(client.embeddings)
+ self.files = files.FilesWithStreamingResponse(client.files)
+ self.images = images.ImagesWithStreamingResponse(client.images)
+ self.audio = audio.AudioWithStreamingResponse(client.audio)
+ self.moderations = moderations.ModerationsWithStreamingResponse(client.moderations)
+ self.models = models.ModelsWithStreamingResponse(client.models)
+ self.fine_tuning = fine_tuning.FineTuningWithStreamingResponse(client.fine_tuning)
+ self.vector_stores = vector_stores.VectorStoresWithStreamingResponse(client.vector_stores)
+ self.beta = beta.BetaWithStreamingResponse(client.beta)
+ self.batches = batches.BatchesWithStreamingResponse(client.batches)
+ self.uploads = uploads.UploadsWithStreamingResponse(client.uploads)
+ self.responses = responses.ResponsesWithStreamingResponse(client.responses)
+
+
+class AsyncOpenAIWithStreamedResponse:
+ def __init__(self, client: AsyncOpenAI) -> None:
+ self.completions = completions.AsyncCompletionsWithStreamingResponse(client.completions)
+ self.chat = chat.AsyncChatWithStreamingResponse(client.chat)
+ self.embeddings = embeddings.AsyncEmbeddingsWithStreamingResponse(client.embeddings)
+ self.files = files.AsyncFilesWithStreamingResponse(client.files)
+ self.images = images.AsyncImagesWithStreamingResponse(client.images)
+ self.audio = audio.AsyncAudioWithStreamingResponse(client.audio)
+ self.moderations = moderations.AsyncModerationsWithStreamingResponse(client.moderations)
+ self.models = models.AsyncModelsWithStreamingResponse(client.models)
+ self.fine_tuning = fine_tuning.AsyncFineTuningWithStreamingResponse(client.fine_tuning)
+ self.vector_stores = vector_stores.AsyncVectorStoresWithStreamingResponse(client.vector_stores)
+ self.beta = beta.AsyncBetaWithStreamingResponse(client.beta)
+ self.batches = batches.AsyncBatchesWithStreamingResponse(client.batches)
+ self.uploads = uploads.AsyncUploadsWithStreamingResponse(client.uploads)
+ self.responses = responses.AsyncResponsesWithStreamingResponse(client.responses)
+
+
+Client = OpenAI
+
+AsyncClient = AsyncOpenAI