@@ -4,24 +4,18 @@ import com.aallam.openai.api.audio.transcriptionRequest
4
4
import com.aallam.openai.api.audio.translationRequest
5
5
import com.aallam.openai.api.file.FileSource
6
6
import com.aallam.openai.api.model.ModelId
7
- import com.aallam.openai.client.internal.asSource
8
- import io.ktor.client.*
9
- import io.ktor.client.call.*
10
- import io.ktor.client.request.*
7
+ import com.aallam.openai.client.internal.TestFileSystem
8
+ import com.aallam.openai.client.internal.testFilePath
11
9
import kotlin.test.Test
12
10
import kotlin.test.assertEquals
13
11
import kotlin.test.assertTrue
14
12
15
13
class TestAudio : TestOpenAI () {
16
14
17
- private val httpClient = HttpClient ()
18
-
19
15
@Test
20
16
fun transcription () = test {
21
- val speedTalkingUrl = " https://github.com/aallam/sample-data/raw/main/openai/audio/micro-machines.wav"
22
- val audioBytes: ByteArray = httpClient.get(speedTalkingUrl).body()
23
17
val request = transcriptionRequest {
24
- audio = FileSource (name = " micro-machines.wav" , source = audioBytes.asSource() )
18
+ audio = FileSource (path = testFilePath( " audio/ micro-machines.wav" ), fileSystem = TestFileSystem )
25
19
model = ModelId (" whisper-1" )
26
20
}
27
21
@@ -37,10 +31,8 @@ class TestAudio : TestOpenAI() {
37
31
38
32
@Test
39
33
fun transcriptionText () = test {
40
- val speedTalkingUrl = " https://github.com/aallam/sample-data/raw/main/openai/audio/micro-machines.wav"
41
- val audioBytes: ByteArray = httpClient.get(speedTalkingUrl).body()
42
34
val request = transcriptionRequest {
43
- audio = FileSource (name = " micro-machines.wav" , source = audioBytes.asSource() )
35
+ audio = FileSource (path = testFilePath( " audio/ micro-machines.wav" ), fileSystem = TestFileSystem )
44
36
model = ModelId (" whisper-1" )
45
37
responseFormat = " text"
46
38
}
@@ -50,10 +42,8 @@ class TestAudio : TestOpenAI() {
50
42
51
43
@Test
52
44
fun transcriptionJsonVerbose () = test {
53
- val speedTalkingUrl = " https://github.com/aallam/sample-data/raw/main/openai/audio/micro-machines.wav"
54
- val audioBytes: ByteArray = httpClient.get(speedTalkingUrl).body()
55
45
val request = transcriptionRequest {
56
- audio = FileSource (name = " micro-machines.wav" , source = audioBytes.asSource() )
46
+ audio = FileSource (path = testFilePath( " audio/ micro-machines.wav" ), fileSystem = TestFileSystem )
57
47
model = ModelId (" whisper-1" )
58
48
responseFormat = " verbose_json"
59
49
}
@@ -66,10 +56,8 @@ class TestAudio : TestOpenAI() {
66
56
67
57
@Test
68
58
fun translation () = test {
69
- val multilingualUrl = " https://github.com/aallam/sample-data/raw/main/openai/audio/multilingual.wav"
70
- val audioBytes: ByteArray = httpClient.get(multilingualUrl).body()
71
59
val request = translationRequest {
72
- audio = FileSource (name = " multilingual.wav" , source = audioBytes.asSource() )
60
+ audio = FileSource (path = testFilePath( " audio/ multilingual.wav" ), fileSystem = TestFileSystem )
73
61
model = ModelId (" whisper-1" )
74
62
}
75
63
val translation = openAI.translation(request)
@@ -84,10 +72,8 @@ class TestAudio : TestOpenAI() {
84
72
85
73
@Test
86
74
fun translationText () = test {
87
- val multilingualUrl = " https://github.com/aallam/sample-data/raw/main/openai/audio/multilingual.wav"
88
- val audioBytes: ByteArray = httpClient.get(multilingualUrl).body()
89
75
val request = translationRequest {
90
- audio = FileSource (name = " multilingual.wav" , source = audioBytes.asSource() )
76
+ audio = FileSource (path = testFilePath( " audio/ multilingual.wav" ), fileSystem = TestFileSystem )
91
77
model = ModelId (" whisper-1" )
92
78
responseFormat = " text"
93
79
}
@@ -97,10 +83,8 @@ class TestAudio : TestOpenAI() {
97
83
98
84
@Test
99
85
fun translationJsonVerbose () = test {
100
- val multilingualUrl = " https://github.com/aallam/sample-data/raw/main/openai/audio/multilingual.wav"
101
- val audioBytes: ByteArray = httpClient.get(multilingualUrl).body()
102
86
val request = translationRequest {
103
- audio = FileSource (name = " multilingual.wav" , source = audioBytes.asSource() )
87
+ audio = FileSource (path = testFilePath( " audio/ multilingual.wav" ), fileSystem = TestFileSystem )
104
88
model = ModelId (" whisper-1" )
105
89
responseFormat = " verbose_json"
106
90
}
0 commit comments