This is an automated email from the ASF dual-hosted git repository.

davsclaus pushed a commit to branch camel-4.18.x
in repository https://gitbox.apache.org/repos/asf/camel.git


The following commit(s) were added to refs/heads/camel-4.18.x by this push:
     new a6b24090c9fe Fix OpenAI docs (#21578)
a6b24090c9fe is described below

commit a6b24090c9fea8cb0f83ae610602ddd16b80042b
Author: Lukas Lowinger <[email protected]>
AuthorDate: Tue Feb 24 11:54:19 2026 +0100

    Fix OpenAI docs (#21578)
---
 components/camel-ai/camel-openai/src/main/docs/openai-component.adoc | 5 ++---
 1 file changed, 2 insertions(+), 3 deletions(-)

diff --git 
a/components/camel-ai/camel-openai/src/main/docs/openai-component.adoc 
b/components/camel-ai/camel-openai/src/main/docs/openai-component.adoc
index a42da382887a..afe483f31765 100644
--- a/components/camel-ai/camel-openai/src/main/docs/openai-component.adoc
+++ b/components/camel-ai/camel-openai/src/main/docs/openai-component.adoc
@@ -256,7 +256,7 @@ The full model response is returned as a String in the 
message body.
 When `streaming=true`, the message body contains an 
`Iterator<ChatCompletionChunk>` suitable for Camel streaming EIPs (such as 
`split()` with `streaming()`).
 
 IMPORTANT:
-* Resource cleanup is handled automatically when the Exchange completes 
(success or failure)
+
 * Conversation memory is **not** automatically updated for streaming responses 
(only for non-streaming responses)
 
 === Structured Outputs
@@ -274,6 +274,7 @@ The JSON schema must be a valid JSON object. Invalid schema 
strings will result
 When `conversationMemory=true`, the component maintains conversation history 
in the `CamelOpenAIConversationHistory` exchange property (configurable via 
`conversationHistoryProperty` option). This history is scoped to a single 
Exchange and allows multi-turn conversations within a route.
 
 IMPORTANT:
+
 * Conversation history is automatically updated with each assistant response 
for **non-streaming** responses only
 * The history is stored as a `List<ChatCompletionMessageParam>` in the 
Exchange property
 * The history persists across multiple calls to the endpoint within the same 
Exchange
@@ -300,7 +301,6 @@ from("direct:with-history")
 This component works with any OpenAI API-compatible endpoint by setting the 
`baseUrl` parameter. This includes:
 
 - OpenAI official API (`https://api.openai.com/v1`)
-- Azure OpenAI (may require additional configuration)
 - Local LLM servers (e.g., Ollama, LM Studio, LocalAI)
 - Third-party OpenAI-compatible providers
 
@@ -320,7 +320,6 @@ When using local or third-party providers, ensure they 
support the chat completi
 | Ollama | `nomic-embed-text` | 768
 | Ollama | `mxbai-embed-large` | 1024
 | Mistral | `mistral-embed` | 1024
-| Azure OpenAI | `text-embedding-ada-002` | 1536
 |===
 
 .Example using Ollama for local embeddings:

Reply via email to