Azure OpenAI Provider¶
Configure and use Azure OpenAI Service with Ondine.
Setup¶
export AZURE_OPENAI_API_KEY="..."
export AZURE_OPENAI_ENDPOINT="https://your-resource.openai.azure.com/"
export AZURE_OPENAI_API_VERSION="2024-02-15-preview"
Basic Usage¶
from ondine import PipelineBuilder
pipeline = (
PipelineBuilder.create()
.from_csv("data.csv", input_columns=["text"], output_columns=["result"])
.with_prompt("Process: {text}")
.with_llm(
provider="azure_openai",
model="gpt-4",
azure_endpoint="https://your-resource.openai.azure.com/",
azure_deployment="your-deployment-name",
api_version="2024-02-15-preview"
)
.build()
)
result = pipeline.execute()
Configuration¶
The deployment name in Azure OpenAI maps to the model:
.with_llm(
provider="azure_openai",
model="gpt-4", # Your base model
azure_deployment="my-gpt4-deployment", # Your Azure deployment name
azure_endpoint=os.getenv("AZURE_OPENAI_ENDPOINT"),
api_version="2024-02-15-preview"
)
Rate Limits¶
Rate limits in Azure OpenAI are configured per deployment. Adjust concurrency based on your TPM (tokens per minute) limits.