@@ -20,15 +20,15 @@ class TestCompletions:
20
20
@parametrize
21
21
def test_method_create_overload_1 (self , client : OpenAI ) -> None :
22
22
completion = client .completions .create (
23
- model = "gpt-3.5-turbo-instruct " ,
23
+ model = "string " ,
24
24
prompt = "This is a test." ,
25
25
)
26
26
assert_matches_type (Completion , completion , path = ["response" ])
27
27
28
28
@parametrize
29
29
def test_method_create_with_all_params_overload_1 (self , client : OpenAI ) -> None :
30
30
completion = client .completions .create (
31
- model = "gpt-3.5-turbo-instruct " ,
31
+ model = "string " ,
32
32
prompt = "This is a test." ,
33
33
best_of = 0 ,
34
34
echo = True ,
@@ -52,7 +52,7 @@ def test_method_create_with_all_params_overload_1(self, client: OpenAI) -> None:
52
52
@parametrize
53
53
def test_raw_response_create_overload_1 (self , client : OpenAI ) -> None :
54
54
response = client .completions .with_raw_response .create (
55
- model = "gpt-3.5-turbo-instruct " ,
55
+ model = "string " ,
56
56
prompt = "This is a test." ,
57
57
)
58
58
@@ -64,7 +64,7 @@ def test_raw_response_create_overload_1(self, client: OpenAI) -> None:
64
64
@parametrize
65
65
def test_streaming_response_create_overload_1 (self , client : OpenAI ) -> None :
66
66
with client .completions .with_streaming_response .create (
67
- model = "gpt-3.5-turbo-instruct " ,
67
+ model = "string " ,
68
68
prompt = "This is a test." ,
69
69
) as response :
70
70
assert not response .is_closed
@@ -78,7 +78,7 @@ def test_streaming_response_create_overload_1(self, client: OpenAI) -> None:
78
78
@parametrize
79
79
def test_method_create_overload_2 (self , client : OpenAI ) -> None :
80
80
completion_stream = client .completions .create (
81
- model = "gpt-3.5-turbo-instruct " ,
81
+ model = "string " ,
82
82
prompt = "This is a test." ,
83
83
stream = True ,
84
84
)
@@ -87,7 +87,7 @@ def test_method_create_overload_2(self, client: OpenAI) -> None:
87
87
@parametrize
88
88
def test_method_create_with_all_params_overload_2 (self , client : OpenAI ) -> None :
89
89
completion_stream = client .completions .create (
90
- model = "gpt-3.5-turbo-instruct " ,
90
+ model = "string " ,
91
91
prompt = "This is a test." ,
92
92
stream = True ,
93
93
best_of = 0 ,
@@ -111,7 +111,7 @@ def test_method_create_with_all_params_overload_2(self, client: OpenAI) -> None:
111
111
@parametrize
112
112
def test_raw_response_create_overload_2 (self , client : OpenAI ) -> None :
113
113
response = client .completions .with_raw_response .create (
114
- model = "gpt-3.5-turbo-instruct " ,
114
+ model = "string " ,
115
115
prompt = "This is a test." ,
116
116
stream = True ,
117
117
)
@@ -123,7 +123,7 @@ def test_raw_response_create_overload_2(self, client: OpenAI) -> None:
123
123
@parametrize
124
124
def test_streaming_response_create_overload_2 (self , client : OpenAI ) -> None :
125
125
with client .completions .with_streaming_response .create (
126
- model = "gpt-3.5-turbo-instruct " ,
126
+ model = "string " ,
127
127
prompt = "This is a test." ,
128
128
stream = True ,
129
129
) as response :
@@ -142,15 +142,15 @@ class TestAsyncCompletions:
142
142
@parametrize
143
143
async def test_method_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
144
144
completion = await async_client .completions .create (
145
- model = "gpt-3.5-turbo-instruct " ,
145
+ model = "string " ,
146
146
prompt = "This is a test." ,
147
147
)
148
148
assert_matches_type (Completion , completion , path = ["response" ])
149
149
150
150
@parametrize
151
151
async def test_method_create_with_all_params_overload_1 (self , async_client : AsyncOpenAI ) -> None :
152
152
completion = await async_client .completions .create (
153
- model = "gpt-3.5-turbo-instruct " ,
153
+ model = "string " ,
154
154
prompt = "This is a test." ,
155
155
best_of = 0 ,
156
156
echo = True ,
@@ -174,7 +174,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn
174
174
@parametrize
175
175
async def test_raw_response_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
176
176
response = await async_client .completions .with_raw_response .create (
177
- model = "gpt-3.5-turbo-instruct " ,
177
+ model = "string " ,
178
178
prompt = "This is a test." ,
179
179
)
180
180
@@ -186,7 +186,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncOpenAI) -
186
186
@parametrize
187
187
async def test_streaming_response_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
188
188
async with async_client .completions .with_streaming_response .create (
189
- model = "gpt-3.5-turbo-instruct " ,
189
+ model = "string " ,
190
190
prompt = "This is a test." ,
191
191
) as response :
192
192
assert not response .is_closed
@@ -200,7 +200,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncOpe
200
200
@parametrize
201
201
async def test_method_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
202
202
completion_stream = await async_client .completions .create (
203
- model = "gpt-3.5-turbo-instruct " ,
203
+ model = "string " ,
204
204
prompt = "This is a test." ,
205
205
stream = True ,
206
206
)
@@ -209,7 +209,7 @@ async def test_method_create_overload_2(self, async_client: AsyncOpenAI) -> None
209
209
@parametrize
210
210
async def test_method_create_with_all_params_overload_2 (self , async_client : AsyncOpenAI ) -> None :
211
211
completion_stream = await async_client .completions .create (
212
- model = "gpt-3.5-turbo-instruct " ,
212
+ model = "string " ,
213
213
prompt = "This is a test." ,
214
214
stream = True ,
215
215
best_of = 0 ,
@@ -233,7 +233,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn
233
233
@parametrize
234
234
async def test_raw_response_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
235
235
response = await async_client .completions .with_raw_response .create (
236
- model = "gpt-3.5-turbo-instruct " ,
236
+ model = "string " ,
237
237
prompt = "This is a test." ,
238
238
stream = True ,
239
239
)
@@ -245,7 +245,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncOpenAI) -
245
245
@parametrize
246
246
async def test_streaming_response_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
247
247
async with async_client .completions .with_streaming_response .create (
248
- model = "gpt-3.5-turbo-instruct " ,
248
+ model = "string " ,
249
249
prompt = "This is a test." ,
250
250
stream = True ,
251
251
) as response :
0 commit comments