Skip to content

Commit 2ee898c

Browse files
authored
chore: indicate to mypy that InferenceProvider.rerank is concrete (#3238)
1 parent da73f1a commit 2ee898c

File tree

6 files changed

+1
-62
lines changed

6 files changed

+1
-62
lines changed

llama_stack/apis/inference/inference.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1170,6 +1170,7 @@ async def rerank(
11701170
:returns: RerankResponse with indices sorted by relevance score (descending).
11711171
"""
11721172
raise NotImplementedError("Reranking is not implemented")
1173+
return # this is so mypy's safe-super rule will consider the method concrete
11731174

11741175
@webmethod(route="/openai/v1/completions", method="POST")
11751176
async def openai_completion(

llama_stack/providers/inline/inference/meta_reference/inference.py

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -33,9 +33,6 @@
3333
InterleavedContent,
3434
LogProbConfig,
3535
Message,
36-
OpenAIChatCompletionContentPartImageParam,
37-
OpenAIChatCompletionContentPartTextParam,
38-
RerankResponse,
3936
ResponseFormat,
4037
SamplingParams,
4138
StopReason,
@@ -445,15 +442,6 @@ async def batch_chat_completion(
445442
results = await self._nonstream_chat_completion(request_batch)
446443
return BatchChatCompletionResponse(batch=results)
447444

448-
async def rerank(
449-
self,
450-
model: str,
451-
query: str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam,
452-
items: list[str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam],
453-
max_num_results: int | None = None,
454-
) -> RerankResponse:
455-
raise NotImplementedError("Reranking is not supported for Meta Reference")
456-
457445
async def _nonstream_chat_completion(
458446
self, request_batch: list[ChatCompletionRequest]
459447
) -> list[ChatCompletionResponse]:

llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers.py

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -12,9 +12,6 @@
1212
InterleavedContent,
1313
LogProbConfig,
1414
Message,
15-
OpenAIChatCompletionContentPartImageParam,
16-
OpenAIChatCompletionContentPartTextParam,
17-
RerankResponse,
1815
ResponseFormat,
1916
SamplingParams,
2017
ToolChoice,
@@ -125,12 +122,3 @@ async def batch_chat_completion(
125122
logprobs: LogProbConfig | None = None,
126123
):
127124
raise NotImplementedError("Batch chat completion is not supported for Sentence Transformers")
128-
129-
async def rerank(
130-
self,
131-
model: str,
132-
query: str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam,
133-
items: list[str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam],
134-
max_num_results: int | None = None,
135-
) -> RerankResponse:
136-
raise NotImplementedError("Reranking is not supported for Sentence Transformers")

llama_stack/providers/remote/inference/llama_openai_compat/llama.py

Lines changed: 0 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -3,11 +3,6 @@
33
#
44
# This source code is licensed under the terms described in the LICENSE file in
55
# the root directory of this source tree.
6-
from llama_stack.apis.inference import (
7-
OpenAIChatCompletionContentPartImageParam,
8-
OpenAIChatCompletionContentPartTextParam,
9-
RerankResponse,
10-
)
116
from llama_stack.log import get_logger
127
from llama_stack.providers.remote.inference.llama_openai_compat.config import LlamaCompatConfig
138
from llama_stack.providers.utils.inference.litellm_openai_mixin import LiteLLMOpenAIMixin
@@ -59,12 +54,3 @@ async def initialize(self):
5954

6055
async def shutdown(self):
6156
await super().shutdown()
62-
63-
async def rerank(
64-
self,
65-
model: str,
66-
query: str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam,
67-
items: list[str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam],
68-
max_num_results: int | None = None,
69-
) -> RerankResponse:
70-
raise NotImplementedError("Reranking is not supported for Llama OpenAI Compat")

llama_stack/providers/remote/inference/ollama/ollama.py

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -37,14 +37,11 @@
3737
Message,
3838
OpenAIChatCompletion,
3939
OpenAIChatCompletionChunk,
40-
OpenAIChatCompletionContentPartImageParam,
41-
OpenAIChatCompletionContentPartTextParam,
4240
OpenAICompletion,
4341
OpenAIEmbeddingsResponse,
4442
OpenAIEmbeddingUsage,
4543
OpenAIMessageParam,
4644
OpenAIResponseFormatParam,
47-
RerankResponse,
4845
ResponseFormat,
4946
SamplingParams,
5047
TextTruncation,
@@ -644,15 +641,6 @@ async def batch_chat_completion(
644641
):
645642
raise NotImplementedError("Batch chat completion is not supported for Ollama")
646643

647-
async def rerank(
648-
self,
649-
model: str,
650-
query: str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam,
651-
items: list[str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam],
652-
max_num_results: int | None = None,
653-
) -> RerankResponse:
654-
raise NotImplementedError("Reranking is not supported for Ollama")
655-
656644

657645
async def convert_message_to_openai_dict_for_ollama(message: Message) -> list[dict]:
658646
async def _convert_content(content) -> dict:

llama_stack/providers/remote/inference/vllm/vllm.py

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -39,15 +39,12 @@
3939
Message,
4040
ModelStore,
4141
OpenAIChatCompletion,
42-
OpenAIChatCompletionContentPartImageParam,
43-
OpenAIChatCompletionContentPartTextParam,
4442
OpenAICompletion,
4543
OpenAIEmbeddingData,
4644
OpenAIEmbeddingsResponse,
4745
OpenAIEmbeddingUsage,
4846
OpenAIMessageParam,
4947
OpenAIResponseFormatParam,
50-
RerankResponse,
5148
ResponseFormat,
5249
SamplingParams,
5350
TextTruncation,
@@ -736,12 +733,3 @@ async def batch_chat_completion(
736733
logprobs: LogProbConfig | None = None,
737734
):
738735
raise NotImplementedError("Batch chat completion is not supported for vLLM")
739-
740-
async def rerank(
741-
self,
742-
model: str,
743-
query: str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam,
744-
items: list[str | OpenAIChatCompletionContentPartTextParam | OpenAIChatCompletionContentPartImageParam],
745-
max_num_results: int | None = None,
746-
) -> RerankResponse:
747-
raise NotImplementedError("Reranking is not supported for vLLM")

0 commit comments

Comments
 (0)