[DOCS] Remove references to the _unified inference API (#124157) (#124254)

(cherry picked from commit 0f58537dc9)
This commit is contained in:
Lisa Cawley 2025-03-06 09:39:39 -08:00 committed by GitHub
parent 533d0a8750
commit b95c59b290
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
4 changed files with 6 additions and 6 deletions

View file

@ -13,9 +13,9 @@ However, if you do not plan to use the {infer} APIs to use these models or if yo
[[chat-completion-inference-api-request]] [[chat-completion-inference-api-request]]
==== {api-request-title} ==== {api-request-title}
`POST /_inference/<inference_id>/_unified` `POST /_inference/<inference_id>/_stream`
`POST /_inference/chat_completion/<inference_id>/_unified` `POST /_inference/chat_completion/<inference_id>/_stream`
[discrete] [discrete]
@ -37,7 +37,7 @@ It only works with the `chat_completion` task type for `openai` and `elastic` {i
[NOTE] [NOTE]
==== ====
* The `chat_completion` task type is only available within the _unified API and only supports streaming. * The `chat_completion` task type is only available within the _stream API and only supports streaming.
* The Chat completion {infer} API and the Stream {infer} API differ in their response structure and capabilities. * The Chat completion {infer} API and the Stream {infer} API differ in their response structure and capabilities.
The Chat completion {infer} API provides more comprehensive customization options through more fields and function calling support. The Chat completion {infer} API provides more comprehensive customization options through more fields and function calling support.
If you use the `openai` service or the `elastic` service, use the Chat completion {infer} API. If you use the `openai` service or the `elastic` service, use the Chat completion {infer} API.

View file

@ -39,7 +39,7 @@ Available task types:
[NOTE] [NOTE]
==== ====
The `chat_completion` task type only supports streaming and only through the `_unified` API. The `chat_completion` task type only supports streaming and only through the `_stream` API.
include::inference-shared.asciidoc[tag=chat-completion-docs] include::inference-shared.asciidoc[tag=chat-completion-docs]
==== ====

View file

@ -38,7 +38,7 @@ Available task types:
[NOTE] [NOTE]
==== ====
The `chat_completion` task type only supports streaming and only through the `_unified` API. The `chat_completion` task type only supports streaming and only through the `_stream` API.
include::inference-shared.asciidoc[tag=chat-completion-docs] include::inference-shared.asciidoc[tag=chat-completion-docs]
==== ====

View file

@ -297,7 +297,7 @@ Rerank the results using the new {infer} endpoint.
[source,py] [source,py]
-------------------------------------------------- --------------------------------------------------
# Pass the query and the search results to the service # Pass the query and the search results to the service
response = client.inference.inference( response = client.inference.rerank(
inference_id="cohere_rerank", inference_id="cohere_rerank",
body={ body={
"query": query, "query": query,