Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions .semversioner/next-release/patch-20250519184041368529.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,4 @@
{
"type": "patch",
"description": "[Bugfix]: max_length is not passed in the prompt template reduce_system_prompt in Global search.py"
}
6 changes: 5 additions & 1 deletion graphrag/query/structured_search/global_search/search.py
Original file line number Diff line number Diff line change
Expand Up @@ -184,6 +184,7 @@ async def search(
reduce_response = await self._reduce_response(
map_responses=map_responses,
query=query,
max_length=self.reduce_max_length,
**self.reduce_llm_params,
)
llm_calls["reduce"] = reduce_response.llm_calls
Expand Down Expand Up @@ -297,6 +298,7 @@ async def _reduce_response(
self,
map_responses: list[SearchResult],
query: str,
max_length: int,
**llm_kwargs,
) -> SearchResult:
"""Combine all intermediate responses from single batches into a final answer to the user query."""
Expand Down Expand Up @@ -371,7 +373,9 @@ async def _reduce_response(
text_data = "\n\n".join(data)

search_prompt = self.reduce_system_prompt.format(
report_data=text_data, response_type=self.response_type
report_data=text_data,
response_type=self.response_type,
max_length=max_length,
)
if self.allow_general_knowledge:
search_prompt += "\n" + self.general_knowledge_inclusion_prompt
Expand Down