From 3a67f3d850720e7983c9cc9c469abc2b8224c133 Mon Sep 17 00:00:00 2001 From: BV-Venky Date: Sun, 1 Mar 2026 19:15:06 +0530 Subject: [PATCH] feat: widen openai dependency to support 2.x for litellm compatibility --- pyproject.toml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index b53194486..cf0525aa1 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -45,15 +45,15 @@ dependencies = [ [project.optional-dependencies] anthropic = ["anthropic>=0.21.0,<1.0.0"] gemini = ["google-genai>=1.32.0,<2.0.0"] -litellm = ["litellm>=1.75.9,<2.0.0", "openai>=1.68.0,<1.110.0"] +litellm = ["litellm>=1.75.9,<2.0.0", "openai>=1.68.0,<3.0.0"] llamaapi = ["llama-api-client>=0.1.0,<1.0.0"] mistral = ["mistralai>=1.8.2"] ollama = ["ollama>=0.4.8,<1.0.0"] -openai = ["openai>=1.68.0,<2.0.0"] +openai = ["openai>=1.68.0,<3.0.0"] writer = ["writer-sdk>=2.2.0,<3.0.0"] sagemaker = [ "boto3-stubs[sagemaker-runtime]>=1.26.0,<2.0.0", - "openai>=1.68.0,<2.0.0", # SageMaker uses OpenAI-compatible interface + "openai>=1.68.0,<3.0.0", # SageMaker uses OpenAI-compatible interface ] otel = ["opentelemetry-exporter-otlp-proto-http>=1.30.0,<2.0.0"] docs = [