@@ -1051,7 +1051,9 @@ def get_transformed_inputs(kwargs):
1051
1051
if "input" in litellm_params :
1052
1052
append_log (state , "Model Input" , litellm_params ["input" ])
1053
1053
else :
1054
- append_log (state , "Model Input" , model_input )
1054
+ append_log (
1055
+ state , "Model Input" , messages_to_str (concrete_block .model , model_input )
1056
+ )
1055
1057
background : Messages = [msg ]
1056
1058
result = msg ["content" ]
1057
1059
append_log (state , "Model Output" , result )
@@ -1093,7 +1095,7 @@ def generate_client_response_streaming(
1093
1095
model_input : Messages ,
1094
1096
) -> Generator [YieldMessage , Any , Message ]:
1095
1097
msg_stream : Generator [Message , Any , None ]
1096
- model_input_str = messages_to_str (model_input )
1098
+ model_input_str = messages_to_str (block . model , model_input )
1097
1099
match block :
1098
1100
case BamModelBlock ():
1099
1101
msg_stream = BamModel .generate_text_stream (
@@ -1148,7 +1150,7 @@ def generate_client_response_single(
1148
1150
model_input : Messages ,
1149
1151
) -> Generator [YieldMessage , Any , Message ]:
1150
1152
msg : Message
1151
- model_input_str = messages_to_str (model_input )
1153
+ model_input_str = messages_to_str (block . model , model_input )
1152
1154
match block :
1153
1155
case BamModelBlock ():
1154
1156
msg = BamModel .generate_text (
@@ -1178,7 +1180,7 @@ def generate_client_response_batching( # pylint: disable=too-many-arguments
1178
1180
# model: str,
1179
1181
model_input : Messages ,
1180
1182
) -> Generator [YieldMessage , Any , Message ]:
1181
- model_input_str = messages_to_str (model_input )
1183
+ model_input_str = messages_to_str (block . model , model_input )
1182
1184
match block :
1183
1185
case BamModelBlock ():
1184
1186
msg = yield ModelCallMessage (
0 commit comments