@@ -1185,7 +1185,7 @@ describe.concurrent("HfInference", () => {
1185
1185
"meta-llama/Llama-3.2-3B-Instruct" : "meta-llama/Llama-3.2-3B-Instruct" ,
1186
1186
"meta-llama/Llama-3.3-70B-Instruct" : "meta-llama/Llama-3.3-70B-Instruct" ,
1187
1187
"stabilityai/stable-diffusion-2" : "stabilityai/stable-diffusion-2" ,
1188
- "meta-llama/Llama-3.1-405B-BASE-FP8 " : "meta-llama/Llama-3.1-405B-BASE-FP8 " ,
1188
+ "meta-llama/Llama-3.1-405B" : "meta-llama/Llama-3.1-405B" ,
1189
1189
} ;
1190
1190
1191
1191
it ( "chatCompletion - hyperbolic" , async ( ) => {
@@ -1227,20 +1227,18 @@ describe.concurrent("HfInference", () => {
1227
1227
const res = await client . textToImage ( {
1228
1228
model : "stabilityai/stable-diffusion-2" ,
1229
1229
provider : "hyperbolic" ,
1230
- inputs : " award winning high resolution photo of a giant tortoise",
1230
+ messages : [ { role : "user" , content : " award winning high resolution photo of a giant tortoise" } ] ,
1231
1231
} ) ;
1232
1232
expect ( res ) . toBeInstanceOf ( Blob ) ;
1233
1233
} ) ;
1234
1234
1235
1235
it ( "textGeneration" , async ( ) => {
1236
1236
const res = await client . textGeneration ( {
1237
- model : "meta-llama/Llama-3.1-405B-BASE-FP8 " ,
1237
+ model : "meta-llama/Llama-3.1-405B" ,
1238
1238
provider : "hyperbolic" ,
1239
- inputs : "Paris is" ,
1240
- parameters : {
1241
- temperature : 0 ,
1242
- max_tokens : 10 ,
1243
- } ,
1239
+ messages : [ { role : "user" , content : "Paris is" } ] ,
1240
+ temperature : 0 ,
1241
+ max_tokens : 10 ,
1244
1242
} ) ;
1245
1243
expect ( res ) . toMatchObject ( { generated_text : " city of love" } ) ;
1246
1244
} ) ;
0 commit comments