@@ -225,6 +225,10 @@ async def request(
225
225
model_request_parameters : ModelRequestParameters ,
226
226
) -> ModelResponse :
227
227
check_allow_model_requests ()
228
+ model_settings , model_request_parameters = self .prepare_request (
229
+ model_settings ,
230
+ model_request_parameters ,
231
+ )
228
232
model_settings = cast (GoogleModelSettings , model_settings or {})
229
233
response = await self ._generate_content (messages , False , model_settings , model_request_parameters )
230
234
return self ._process_response (response )
@@ -236,6 +240,10 @@ async def count_tokens(
236
240
model_request_parameters : ModelRequestParameters ,
237
241
) -> usage .RequestUsage :
238
242
check_allow_model_requests ()
243
+ model_settings , model_request_parameters = self .prepare_request (
244
+ model_settings ,
245
+ model_request_parameters ,
246
+ )
239
247
model_settings = cast (GoogleModelSettings , model_settings or {})
240
248
contents , generation_config = await self ._build_content_and_config (
241
249
messages , model_settings , model_request_parameters
@@ -291,6 +299,10 @@ async def request_stream(
291
299
run_context : RunContext [Any ] | None = None ,
292
300
) -> AsyncIterator [StreamedResponse ]:
293
301
check_allow_model_requests ()
302
+ model_settings , model_request_parameters = self .prepare_request (
303
+ model_settings ,
304
+ model_request_parameters ,
305
+ )
294
306
model_settings = cast (GoogleModelSettings , model_settings or {})
295
307
response = await self ._generate_content (messages , True , model_settings , model_request_parameters )
296
308
yield await self ._process_streamed_response (response , model_request_parameters ) # type: ignore
0 commit comments