Skip to content

Commit 7809d5c

Browse files
committed
OLS-1790: minor edit
1 parent 2ffd5b0 commit 7809d5c

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

modules/ols-large-language-model-requirements.adoc

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -43,16 +43,16 @@ To use {azure-official} with {ols-official}, you need access to link:https://azu
4343

4444
You can configure {rhelai} as the (Large Language Model) LLM provider.
4545

46-
{ols-long} version 1.0 supports vLLM Server version 0.8.4. When self-hosting a large language model with {rhelai}, you can use vLLM Server as the inference engine for your model deployment.
47-
4846
Because the {rhel} is in a different environment than the {ols-long} deployment, the model deployment must allow access using a secure connection. For more information, see link:https://docs.redhat.com/en/documentation/red_hat_enterprise_linux_ai/1.2/html-single/building_your_rhel_ai_environment/index#creating_secure_endpoint[Optional: Allowing access to a model from a secure endpoint].
4947

48+
{ols-long} version 1.0 supports vLLM Server version 0.8.4. When self-hosting a large language model with {rhelai}, you can use vLLM Server as the inference engine for your model deployment.
49+
5050

5151
[id="rhoai_{context}"]
5252
== {rhoai}
5353

5454
{rhoai} is OpenAI API-compatible, and is configured largely the same as the OpenAI provider.
5555

56-
{ols-long} version 1.0 supports vLLM Server version 0.8.4. When self-hosting a large language model with {rhoai}, you can use vLLM Server as the inference engine for your model deployment.
56+
You need a Large Language Model (LLM) deployed on the single model-serving platform of {rhoai} using the Virtual Large Language Model (vLLM) runtime. If the model deployment is in a different {ocp-short-name} environment than the {ols-long} deployment, the model deployment must include a route to expose it outside the cluster. For more information, see link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2-latest/html/serving_models/serving-large-models_serving-large-models#about-the-single-model-serving-platform_serving-large-models[About the single-model serving platform].
5757

58-
You need a Large Language Model (LLM) deployed on the single model-serving platform of {rhoai} using the Virtual Large Language Model (vLLM) runtime. If the model deployment is in a different {ocp-short-name} environment than the {ols-long} deployment, the model deployment must include a route to expose it outside the cluster. For more information, see link:https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/2-latest/html/serving_models/serving-large-models_serving-large-models#about-the-single-model-serving-platform_serving-large-models[About the single-model serving platform].
58+
{ols-long} version 1.0 supports vLLM Server version 0.8.4. When self-hosting a large language model with {rhoai}, you can use vLLM Server as the inference engine for your model deployment.

0 commit comments

Comments
 (0)