You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: demos/embeddings/README.md
+31-1Lines changed: 31 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -503,10 +503,40 @@ Results will be stored in `results` folder:
503
503
"kg_co2_emissions": null
504
504
}
505
505
```
506
-
507
506
Compare against local HuggingFace execution for reference:
508
507
```console
509
508
mteb run -m thenlper/gte-small -t Banking77Classification --output_folder results
509
+
```
510
+
511
+
# Usage of tokenize endpoint (release 2025.4 or weekly)
512
+
513
+
The `tokenize` endpoint provides a simple API for tokenizing input text using the same tokenizer as the deployed embeddings model. This allows you to see how your text will be split into tokens before feature extraction or inference. The endpoint accepts a string or list of strings and returns the corresponding token IDs and tokenized text.
- pad_to_max_length - whether to pad the sequence to the maximum length. Default is False.
528
+
- max_length - maximum length of the sequence. If None (default), the value will be taken from the IR (where default value from original HF/GGUF model is stored).
529
+
- padding_side - side to pad the sequence, can be ‘left’ or ‘right’. Default is None.
530
+
- add_special_tokens - whether to add special tokens like BOS, EOS, PAD. Default is True.
SPDLOG_LOGGER_DEBUG(embeddings_calculator_logger, "max_position_embeddings nor max_trained_positions included in config.json. Using default value {}", max_context_length);
0 commit comments