16
16
*/
17
17
18
18
import { BaseConversationalTask , BaseTextGenerationTask } from "./providerHelper" ;
19
- import type {
20
- ChatCompletionOutput ,
21
- TextGenerationOutput ,
22
- TextGenerationOutputFinishReason ,
23
- } from "@huggingface/tasks" ;
19
+ import type { ChatCompletionOutput , TextGenerationOutput , TextGenerationOutputFinishReason } from "@huggingface/tasks" ;
24
20
import { InferenceOutputError } from "../lib/InferenceOutputError" ;
25
21
import type { BodyParams } from "../types" ;
26
22
import { omit } from "../utils/omit" ;
@@ -54,9 +50,9 @@ export class OvhCloudTextGenerationTask extends BaseTextGenerationTask {
54
50
...omit ( params . args , [ "inputs" , "parameters" ] ) ,
55
51
...( params . args . parameters
56
52
? {
57
- max_tokens : ( params . args . parameters as Record < string , unknown > ) . max_new_tokens ,
58
- ...omit ( params . args . parameters as Record < string , unknown > , "max_new_tokens" ) ,
59
- }
53
+ max_tokens : ( params . args . parameters as Record < string , unknown > ) . max_new_tokens ,
54
+ ...omit ( params . args . parameters as Record < string , unknown > , "max_new_tokens" ) ,
55
+ }
60
56
: undefined ) ,
61
57
prompt : params . args . inputs ,
62
58
} ;
@@ -76,5 +72,4 @@ export class OvhCloudTextGenerationTask extends BaseTextGenerationTask {
76
72
}
77
73
throw new InferenceOutputError ( "Expected OVHcloud text generation response format" ) ;
78
74
}
79
-
80
- }
75
+ }
0 commit comments