19
19
def test_prompt ():
20
20
model = llm .get_model ("openrouter/openai/gpt-4o" )
21
21
response = model .prompt ("Two names for a pet pelican, be brief" )
22
- assert str (response ) == snapshot ("Skyler and Splash ." )
22
+ assert str (response ) == snapshot ("Pebbles and Skipper ." )
23
23
response_dict = dict (response .response_json )
24
24
response_dict .pop ("id" ) # differs between requests
25
25
assert response_dict == snapshot (
26
26
{
27
- "content" : "Skyler and Splash ." ,
27
+ "content" : "Pebbles and Skipper ." ,
28
28
"role" : "assistant" ,
29
29
"finish_reason" : "stop" ,
30
30
"usage" : {
31
- "completion_tokens" : 5 ,
31
+ "completion_tokens" : 6 ,
32
32
"prompt_tokens" : 17 ,
33
- "total_tokens" : 22 ,
33
+ "total_tokens" : 23 ,
34
34
"completion_tokens_details" : {"reasoning_tokens" : 0 },
35
35
"prompt_tokens_details" : {"cached_tokens" : 0 },
36
- "cost" : 9.25e-05 ,
36
+ "cost" : 0.0001025 ,
37
37
"is_byok" : False ,
38
38
},
39
39
"object" : "chat.completion.chunk" ,
40
40
"model" : "openai/gpt-4o" ,
41
- "created" : 1754441226 ,
41
+ "created" : 1754441342 ,
42
42
}
43
43
)
44
44
@@ -63,12 +63,12 @@ def test_image_prompt():
63
63
"Describe image in three words" ,
64
64
attachments = [llm .Attachment (content = TINY_PNG )],
65
65
)
66
- assert str (response ) == snapshot ("Red green sequence squares " )
66
+ assert str (response ) == snapshot ("Red green geometric shapes " )
67
67
response_dict = response .response_json
68
68
response_dict .pop ("id" ) # differs between requests
69
69
assert response_dict == snapshot (
70
70
{
71
- "content" : "Red green sequence squares " ,
71
+ "content" : "Red green geometric shapes " ,
72
72
"role" : "assistant" ,
73
73
"finish_reason" : "stop" ,
74
74
"usage" : {
@@ -82,6 +82,6 @@ def test_image_prompt():
82
82
},
83
83
"object" : "chat.completion.chunk" ,
84
84
"model" : "anthropic/claude-3.5-sonnet" ,
85
- "created" : 1754441227 ,
85
+ "created" : 1754441344 ,
86
86
}
87
87
)
0 commit comments