@@ -20,15 +20,15 @@ class TestTranscriptions:
2020 @parametrize
2121 def test_method_create_overload_1 (self , client : OpenAI ) -> None :
2222 transcription = client .audio .transcriptions .create (
23- file = b"raw file contents " ,
23+ file = b"Example data " ,
2424 model = "gpt-4o-transcribe" ,
2525 )
2626 assert_matches_type (TranscriptionCreateResponse , transcription , path = ["response" ])
2727
2828 @parametrize
2929 def test_method_create_with_all_params_overload_1 (self , client : OpenAI ) -> None :
3030 transcription = client .audio .transcriptions .create (
31- file = b"raw file contents " ,
31+ file = b"Example data " ,
3232 model = "gpt-4o-transcribe" ,
3333 chunking_strategy = "auto" ,
3434 include = ["logprobs" ],
@@ -46,7 +46,7 @@ def test_method_create_with_all_params_overload_1(self, client: OpenAI) -> None:
4646 @parametrize
4747 def test_raw_response_create_overload_1 (self , client : OpenAI ) -> None :
4848 response = client .audio .transcriptions .with_raw_response .create (
49- file = b"raw file contents " ,
49+ file = b"Example data " ,
5050 model = "gpt-4o-transcribe" ,
5151 )
5252
@@ -58,7 +58,7 @@ def test_raw_response_create_overload_1(self, client: OpenAI) -> None:
5858 @parametrize
5959 def test_streaming_response_create_overload_1 (self , client : OpenAI ) -> None :
6060 with client .audio .transcriptions .with_streaming_response .create (
61- file = b"raw file contents " ,
61+ file = b"Example data " ,
6262 model = "gpt-4o-transcribe" ,
6363 ) as response :
6464 assert not response .is_closed
@@ -72,7 +72,7 @@ def test_streaming_response_create_overload_1(self, client: OpenAI) -> None:
7272 @parametrize
7373 def test_method_create_overload_2 (self , client : OpenAI ) -> None :
7474 transcription_stream = client .audio .transcriptions .create (
75- file = b"raw file contents " ,
75+ file = b"Example data " ,
7676 model = "gpt-4o-transcribe" ,
7777 stream = True ,
7878 )
@@ -81,7 +81,7 @@ def test_method_create_overload_2(self, client: OpenAI) -> None:
8181 @parametrize
8282 def test_method_create_with_all_params_overload_2 (self , client : OpenAI ) -> None :
8383 transcription_stream = client .audio .transcriptions .create (
84- file = b"raw file contents " ,
84+ file = b"Example data " ,
8585 model = "gpt-4o-transcribe" ,
8686 stream = True ,
8787 chunking_strategy = "auto" ,
@@ -99,7 +99,7 @@ def test_method_create_with_all_params_overload_2(self, client: OpenAI) -> None:
9999 @parametrize
100100 def test_raw_response_create_overload_2 (self , client : OpenAI ) -> None :
101101 response = client .audio .transcriptions .with_raw_response .create (
102- file = b"raw file contents " ,
102+ file = b"Example data " ,
103103 model = "gpt-4o-transcribe" ,
104104 stream = True ,
105105 )
@@ -111,7 +111,7 @@ def test_raw_response_create_overload_2(self, client: OpenAI) -> None:
111111 @parametrize
112112 def test_streaming_response_create_overload_2 (self , client : OpenAI ) -> None :
113113 with client .audio .transcriptions .with_streaming_response .create (
114- file = b"raw file contents " ,
114+ file = b"Example data " ,
115115 model = "gpt-4o-transcribe" ,
116116 stream = True ,
117117 ) as response :
@@ -132,15 +132,15 @@ class TestAsyncTranscriptions:
132132 @parametrize
133133 async def test_method_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
134134 transcription = await async_client .audio .transcriptions .create (
135- file = b"raw file contents " ,
135+ file = b"Example data " ,
136136 model = "gpt-4o-transcribe" ,
137137 )
138138 assert_matches_type (TranscriptionCreateResponse , transcription , path = ["response" ])
139139
140140 @parametrize
141141 async def test_method_create_with_all_params_overload_1 (self , async_client : AsyncOpenAI ) -> None :
142142 transcription = await async_client .audio .transcriptions .create (
143- file = b"raw file contents " ,
143+ file = b"Example data " ,
144144 model = "gpt-4o-transcribe" ,
145145 chunking_strategy = "auto" ,
146146 include = ["logprobs" ],
@@ -158,7 +158,7 @@ async def test_method_create_with_all_params_overload_1(self, async_client: Asyn
158158 @parametrize
159159 async def test_raw_response_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
160160 response = await async_client .audio .transcriptions .with_raw_response .create (
161- file = b"raw file contents " ,
161+ file = b"Example data " ,
162162 model = "gpt-4o-transcribe" ,
163163 )
164164
@@ -170,7 +170,7 @@ async def test_raw_response_create_overload_1(self, async_client: AsyncOpenAI) -
170170 @parametrize
171171 async def test_streaming_response_create_overload_1 (self , async_client : AsyncOpenAI ) -> None :
172172 async with async_client .audio .transcriptions .with_streaming_response .create (
173- file = b"raw file contents " ,
173+ file = b"Example data " ,
174174 model = "gpt-4o-transcribe" ,
175175 ) as response :
176176 assert not response .is_closed
@@ -184,7 +184,7 @@ async def test_streaming_response_create_overload_1(self, async_client: AsyncOpe
184184 @parametrize
185185 async def test_method_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
186186 transcription_stream = await async_client .audio .transcriptions .create (
187- file = b"raw file contents " ,
187+ file = b"Example data " ,
188188 model = "gpt-4o-transcribe" ,
189189 stream = True ,
190190 )
@@ -193,7 +193,7 @@ async def test_method_create_overload_2(self, async_client: AsyncOpenAI) -> None
193193 @parametrize
194194 async def test_method_create_with_all_params_overload_2 (self , async_client : AsyncOpenAI ) -> None :
195195 transcription_stream = await async_client .audio .transcriptions .create (
196- file = b"raw file contents " ,
196+ file = b"Example data " ,
197197 model = "gpt-4o-transcribe" ,
198198 stream = True ,
199199 chunking_strategy = "auto" ,
@@ -211,7 +211,7 @@ async def test_method_create_with_all_params_overload_2(self, async_client: Asyn
211211 @parametrize
212212 async def test_raw_response_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
213213 response = await async_client .audio .transcriptions .with_raw_response .create (
214- file = b"raw file contents " ,
214+ file = b"Example data " ,
215215 model = "gpt-4o-transcribe" ,
216216 stream = True ,
217217 )
@@ -223,7 +223,7 @@ async def test_raw_response_create_overload_2(self, async_client: AsyncOpenAI) -
223223 @parametrize
224224 async def test_streaming_response_create_overload_2 (self , async_client : AsyncOpenAI ) -> None :
225225 async with async_client .audio .transcriptions .with_streaming_response .create (
226- file = b"raw file contents " ,
226+ file = b"Example data " ,
227227 model = "gpt-4o-transcribe" ,
228228 stream = True ,
229229 ) as response :
0 commit comments