@@ -33,12 +33,14 @@ export const snippetConversational = (
33
33
const messages : ChatCompletionInputMessage [ ] = opts ?. messages ?? [
34
34
{ role : "user" , content : "What is the capital of France?" } ,
35
35
] ;
36
+ const messagesStr = formatGenerationMessages ( { messages, sep : ",\n\t" , start : `[\n\t` , end : `\n]` } ) ;
36
37
37
38
const config = {
38
39
temperature : opts ?. temperature ,
39
40
max_tokens : opts ?. max_tokens ?? 500 ,
40
41
top_p : opts ?. top_p ,
41
42
} ;
43
+ const configStr = formatGenerationConfig ( { config, sep : ",\n\t" , start : "" , end : "" , connector : "=" } ) ;
42
44
43
45
if ( streaming ) {
44
46
return [
@@ -48,12 +50,12 @@ export const snippetConversational = (
48
50
49
51
client = InferenceClient(api_key="${ accessToken || "{API_TOKEN}" } ")
50
52
51
- messages = ${ formatGenerationMessages ( { messages , sep : ",\n\t" , start : `[\n\t` , end : `\n]` } ) }
53
+ messages = ${ messagesStr }
52
54
53
55
stream = client.chat.completions.create(
54
56
model="${ model . id } ",
55
57
messages=messages,
56
- ${ formatGenerationConfig ( { config , sep : ",\n\t" , start : "" , end : "" , connector : "=" } ) } ,
58
+ ${ configStr } ,
57
59
stream=True
58
60
)
59
61
@@ -69,12 +71,12 @@ client = OpenAI(
69
71
api_key="${ accessToken || "{API_TOKEN}" } "
70
72
)
71
73
72
- messages = ${ formatGenerationMessages ( { messages , sep : ",\n\t" , start : `[\n\t` , end : `\n]` } ) }
74
+ messages = ${ messagesStr }
73
75
74
76
stream = client.chat.completions.create(
75
77
model="${ model . id } ",
76
78
messages=messages,
77
- ${ formatGenerationConfig ( { config , sep : ",\n\t" , start : "" , end : "" , connector : "=" } ) } ,
79
+ ${ configStr } ,
78
80
stream=True
79
81
)
80
82
@@ -90,12 +92,12 @@ for chunk in stream:
90
92
91
93
client = InferenceClient(api_key="${ accessToken || "{API_TOKEN}" } ")
92
94
93
- messages = ${ formatGenerationMessages ( { messages , sep : ",\n\t" , start : `[\n\t` , end : `\n]` } ) }
95
+ messages = ${ messagesStr }
94
96
95
97
completion = client.chat.completions.create(
96
98
model="${ model . id } ",
97
99
messages=messages,
98
- ${ formatGenerationConfig ( { config , sep : ",\n\t" , start : "" , end : "" , connector : "=" } ) }
100
+ ${ configStr }
99
101
)
100
102
101
103
print(completion.choices[0].message)` ,
@@ -109,12 +111,12 @@ client = OpenAI(
109
111
api_key="${ accessToken || "{API_TOKEN}" } "
110
112
)
111
113
112
- messages = ${ formatGenerationMessages ( { messages , sep : ",\n\t" , start : `[\n\t` , end : `\n]` } ) }
114
+ messages = ${ messagesStr }
113
115
114
116
completion = client.chat.completions.create(
115
117
model="${ model . id } ",
116
118
messages=messages,
117
- ${ formatGenerationConfig ( { config , sep : ",\n\t" , start : "" , end : "" , connector : "=" } ) }
119
+ ${ configStr }
118
120
)
119
121
120
122
print(completion.choices[0].message)` ,
0 commit comments