Skip to content

Commit eedc28d

Browse files
shubhaatszabosteveyetanothertw
authored
Update elasticsearch-billing-dimensions.md (#2490)
Updated billing page to 1. Exclude the instructions to set auto scaling o 0 2. Added the 1mn tokens for input and output tokens --------- Co-authored-by: István Zoltán Szabó <[email protected]> Co-authored-by: Vlada Chirmicci <[email protected]>
1 parent d79d337 commit eedc28d

File tree

1 file changed

+1
-4
lines changed

1 file changed

+1
-4
lines changed

deploy-manage/cloud-organization/billing/elasticsearch-billing-dimensions.md

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,7 @@ For detailed {{es-serverless}} project rates, see the [{{es-serverless}} pricing
3030
* **Indexing:** The VCUs used to index incoming documents.
3131
* **Search:** The VCUs used to return search results, with the latency and queries per second (QPS) you require.
3232
* **Machine learning:** The VCUs used to perform inference, NLP tasks, and other ML activities.
33+
* **Tokens:** The Elastic Managed LLM is charged per 1Mn Input and Output tokens. The LLM powers all AI Search features such as Playground and AI Assistant for Search, and is enabled by default.
3334

3435

3536
## Data storage and billing [elasticsearch-billing-information-about-the-search-ai-lake-dimension-gb]
@@ -43,10 +44,6 @@ You can control costs using the following strategies:
4344

4445
* **Search Power setting:** [Search Power](../../deploy/elastic-cloud/project-settings.md#elasticsearch-manage-project-search-power-settings) controls the speed of searches against your data. With Search Power, you can improve search performance by adding more resources for querying, or you can reduce provisioned resources to cut costs.
4546
* **Search boost window**: By limiting the number of days of [time series data](../../../solutions/search/ingest-for-search.md#elasticsearch-ingest-time-series-data) that are available for caching, you can reduce the number of search VCUs required.
46-
* **Machine learning trained model autoscaling:** Configure your trained model deployment to allow it to scale down to zero allocations when there are no active inference requests:
47-
48-
* When starting or updating a trained model deployment, [Enable adaptive resources](../../autoscaling/trained-model-autoscaling.md#enabling-autoscaling-in-kibana-adaptive-resources) and set the VCU usage level to **Low**.
49-
* When using the inference API for {{es}} or ELSER, [enable `adaptive_allocations`](../../autoscaling/trained-model-autoscaling.md#enabling-autoscaling-through-apis-adaptive-allocations).
5047

5148
* **Indexing Strategies:** Consider your indexing strategies and how they might impact overall VCU usage and costs:
5249

0 commit comments

Comments
 (0)