amine@106
|
1 import os
|
amine@106
|
2 import sys
|
amine@106
|
3 import math
|
amine@107
|
4 from array import array
|
amine@110
|
5 from tempfile import NamedTemporaryFile
|
amine@110
|
6 import filecmp
|
amine@108
|
7 from unittest import TestCase
|
amine@108
|
8 from genty import genty, genty_dataset
|
amine@110
|
9 from auditok.io import (
|
amine@126
|
10 DATA_FORMAT,
|
amine@121
|
11 AudioIOError,
|
amine@110
|
12 AudioParameterError,
|
amine@126
|
13 BufferAudioSource,
|
amine@110
|
14 check_audio_data,
|
amine@128
|
15 _get_audio_parameters,
|
amine@116
|
16 _array_to_bytes,
|
amine@118
|
17 _mix_audio_channels,
|
amine@119
|
18 _extract_selected_channel,
|
amine@126
|
19 _load_raw,
|
amine@129
|
20 _load_wave,
|
amine@131
|
21 _load_with_pydub,
|
amine@120
|
22 from_file,
|
amine@111
|
23 _save_raw,
|
amine@110
|
24 _save_wave,
|
amine@110
|
25 )
|
amine@106
|
26
|
amine@106
|
27
|
amine@106
|
28 if sys.version_info >= (3, 0):
|
amine@106
|
29 PYTHON_3 = True
|
amine@124
|
30 from unittest.mock import patch, Mock
|
amine@106
|
31 else:
|
amine@106
|
32 PYTHON_3 = False
|
amine@124
|
33 from mock import patch, Mock
|
amine@120
|
34
|
amine@120
|
35 AUDIO_PARAMS_SHORT = {"sr": 16000, "sw": 2, "ch": 1}
|
amine@106
|
36
|
amine@106
|
37
|
amine@106
|
38 def _sample_generator(*data_buffers):
|
amine@106
|
39 """
|
amine@106
|
40 Takes a list of many mono audio data buffers and makes a sample generator
|
amine@106
|
41 of interleaved audio samples, one sample from each channel. The resulting
|
amine@106
|
42 generator can be used to build a multichannel audio buffer.
|
amine@106
|
43 >>> gen = _sample_generator("abcd", "ABCD")
|
amine@106
|
44 >>> list(gen)
|
amine@106
|
45 ["a", "A", "b", "B", "c", "C", "d", "D"]
|
amine@106
|
46 """
|
amine@106
|
47 frame_gen = zip(*data_buffers)
|
amine@106
|
48 return (sample for frame in frame_gen for sample in frame)
|
amine@106
|
49
|
amine@106
|
50
|
amine@107
|
51 def _generate_pure_tone(
|
amine@107
|
52 frequency, duration_sec=1, sampling_rate=16000, sample_width=2, volume=1e4
|
amine@107
|
53 ):
|
amine@107
|
54 """
|
amine@107
|
55 Generates a pure tone with the given frequency.
|
amine@107
|
56 """
|
amine@107
|
57 assert frequency <= sampling_rate / 2
|
amine@107
|
58 max_value = (2 ** (sample_width * 8) // 2) - 1
|
amine@107
|
59 if volume > max_value:
|
amine@107
|
60 volume = max_value
|
amine@107
|
61 fmt = DATA_FORMAT[sample_width]
|
amine@107
|
62 total_samples = int(sampling_rate * duration_sec)
|
amine@107
|
63 step = frequency / sampling_rate
|
amine@107
|
64 two_pi_step = 2 * math.pi * step
|
amine@107
|
65 data = array(
|
amine@107
|
66 fmt,
|
amine@107
|
67 (
|
amine@107
|
68 int(math.sin(two_pi_step * i) * volume)
|
amine@107
|
69 for i in range(total_samples)
|
amine@107
|
70 ),
|
amine@107
|
71 )
|
amine@107
|
72 return data
|
amine@107
|
73
|
amine@107
|
74
|
amine@107
|
75 PURE_TONE_DICT = {
|
amine@107
|
76 freq: _generate_pure_tone(freq, 1, 16000, 2) for freq in (400, 800, 1600)
|
amine@107
|
77 }
|
amine@107
|
78 PURE_TONE_DICT.update(
|
amine@107
|
79 {
|
amine@107
|
80 freq: _generate_pure_tone(freq, 0.1, 16000, 2)
|
amine@107
|
81 for freq in (600, 1150, 2400, 7220)
|
amine@107
|
82 }
|
amine@107
|
83 )
|
amine@108
|
84
|
amine@108
|
85
|
amine@108
|
86 @genty
|
amine@108
|
87 class TestIO(TestCase):
|
amine@108
|
88 @genty_dataset(
|
amine@108
|
89 valid_mono=(b"\0" * 113, 1, 1),
|
amine@108
|
90 valid_stereo=(b"\0" * 160, 1, 2),
|
amine@108
|
91 invalid_mono_sw_2=(b"\0" * 113, 2, 1, False),
|
amine@108
|
92 invalid_stereo_sw_1=(b"\0" * 113, 1, 2, False),
|
amine@108
|
93 invalid_stereo_sw_2=(b"\0" * 158, 2, 2, False),
|
amine@108
|
94 )
|
amine@108
|
95 def test_check_audio_data(self, data, sample_width, channels, valid=True):
|
amine@108
|
96
|
amine@108
|
97 if not valid:
|
amine@108
|
98 with self.assertRaises(AudioParameterError):
|
amine@108
|
99 check_audio_data(data, sample_width, channels)
|
amine@108
|
100 else:
|
amine@108
|
101 self.assertIsNone(check_audio_data(data, sample_width, channels))
|
amine@110
|
102
|
amine@110
|
103 @genty_dataset(
|
amine@118
|
104 mono_1byte=([400], 1),
|
amine@118
|
105 stereo_1byte=([400, 600], 1),
|
amine@118
|
106 three_channel_1byte=([400, 600, 2400], 1),
|
amine@118
|
107 mono_2byte=([400], 2),
|
amine@118
|
108 stereo_2byte=([400, 600], 2),
|
amine@118
|
109 three_channel_2byte=([400, 600, 1150], 2),
|
amine@118
|
110 mono_4byte=([400], 4),
|
amine@118
|
111 stereo_4byte=([400, 600], 4),
|
amine@118
|
112 four_channel_2byte=([400, 600, 1150, 7220], 4),
|
amine@118
|
113 )
|
amine@118
|
114 def test_mix_audio_channels(self, frequencies, sample_width):
|
amine@118
|
115 sampling_rate = 16000
|
amine@118
|
116 sample_width = 2
|
amine@118
|
117 channels = len(frequencies)
|
amine@118
|
118 mono_channels = [
|
amine@118
|
119 _generate_pure_tone(
|
amine@118
|
120 freq,
|
amine@118
|
121 duration_sec=0.1,
|
amine@118
|
122 sampling_rate=sampling_rate,
|
amine@118
|
123 sample_width=sample_width,
|
amine@118
|
124 )
|
amine@118
|
125 for freq in frequencies
|
amine@118
|
126 ]
|
amine@118
|
127 fmt = DATA_FORMAT[sample_width]
|
amine@118
|
128 expected = _array_to_bytes(
|
amine@118
|
129 array(
|
amine@118
|
130 fmt,
|
amine@118
|
131 (sum(samples) // channels for samples in zip(*mono_channels)),
|
amine@118
|
132 )
|
amine@118
|
133 )
|
amine@118
|
134 data = _array_to_bytes(array(fmt, _sample_generator(*mono_channels)))
|
amine@118
|
135 mixed = _mix_audio_channels(data, channels, sample_width)
|
amine@118
|
136 self.assertEqual(mixed, expected)
|
amine@118
|
137
|
amine@118
|
138 @genty_dataset(
|
amine@119
|
139 mono_1byte=([400], 1, 0),
|
amine@119
|
140 stereo_1byte_2st_channel=([400, 600], 1, 1),
|
amine@119
|
141 mono_2byte=([400], 2, 0),
|
amine@119
|
142 stereo_2byte_1st_channel=([400, 600], 2, 0),
|
amine@119
|
143 stereo_2byte_2nd_channel=([400, 600], 2, 1),
|
amine@119
|
144 three_channel_2byte_last_negative_idx=([400, 600, 1150], 2, -1),
|
amine@119
|
145 three_channel_2byte_2nd_negative_idx=([400, 600, 1150], 2, -2),
|
amine@119
|
146 three_channel_2byte_1st_negative_idx=([400, 600, 1150], 2, -3),
|
amine@119
|
147 three_channel_4byte_1st=([400, 600, 1150], 4, 0),
|
amine@119
|
148 three_channel_4byte_last_negative_idx=([400, 600, 1150], 4, -1),
|
amine@119
|
149 )
|
amine@119
|
150 def test_extract_selected_channel(
|
amine@119
|
151 self, frequencies, sample_width, use_channel
|
amine@119
|
152 ):
|
amine@119
|
153
|
amine@119
|
154 mono_channels = [
|
amine@119
|
155 _generate_pure_tone(
|
amine@119
|
156 freq,
|
amine@119
|
157 duration_sec=0.1,
|
amine@119
|
158 sampling_rate=16000,
|
amine@119
|
159 sample_width=sample_width,
|
amine@119
|
160 )
|
amine@119
|
161 for freq in frequencies
|
amine@119
|
162 ]
|
amine@119
|
163 channels = len(frequencies)
|
amine@119
|
164 fmt = DATA_FORMAT[sample_width]
|
amine@119
|
165 expected = _array_to_bytes(mono_channels[use_channel])
|
amine@119
|
166 data = _array_to_bytes(array(fmt, _sample_generator(*mono_channels)))
|
amine@119
|
167 selected_channel = _extract_selected_channel(
|
amine@119
|
168 data, channels, sample_width, use_channel
|
amine@119
|
169 )
|
amine@119
|
170 self.assertEqual(selected_channel, expected)
|
amine@119
|
171
|
amine@119
|
172 @genty_dataset(
|
amine@120
|
173 raw_with_audio_format=(
|
amine@120
|
174 "audio",
|
amine@120
|
175 "raw",
|
amine@120
|
176 "_load_raw",
|
amine@120
|
177 AUDIO_PARAMS_SHORT,
|
amine@120
|
178 ),
|
amine@120
|
179 raw_with_extension=(
|
amine@120
|
180 "audio.raw",
|
amine@120
|
181 None,
|
amine@120
|
182 "_load_raw",
|
amine@120
|
183 AUDIO_PARAMS_SHORT,
|
amine@120
|
184 ),
|
amine@120
|
185 wave_with_audio_format=("audio", "wave", "_load_wave"),
|
amine@120
|
186 wav_with_audio_format=("audio", "wave", "_load_wave"),
|
amine@120
|
187 wav_with_extension=("audio.wav", None, "_load_wave"),
|
amine@120
|
188 format_and_extension_both_given=("audio.dat", "wav", "_load_wave"),
|
amine@120
|
189 format_and_extension_both_given_b=("audio.raw", "wave", "_load_wave"),
|
amine@120
|
190 no_format_nor_extension=("audio", None, "_load_with_pydub"),
|
amine@120
|
191 other_formats_ogg=("audio.ogg", None, "_load_with_pydub"),
|
amine@120
|
192 other_formats_webm=("audio", "webm", "_load_with_pydub"),
|
amine@120
|
193 )
|
amine@120
|
194 def test_from_file(
|
amine@120
|
195 self, filename, audio_format, funtion_name, kwargs=None
|
amine@120
|
196 ):
|
amine@120
|
197 funtion_name = "auditok.io." + funtion_name
|
amine@120
|
198 if kwargs is None:
|
amine@120
|
199 kwargs = {}
|
amine@120
|
200 with patch(funtion_name) as patch_function:
|
amine@120
|
201 from_file(filename, audio_format, **kwargs)
|
amine@120
|
202 self.assertTrue(patch_function.called)
|
amine@120
|
203
|
amine@120
|
204 @genty_dataset(
|
amine@111
|
205 mono=("mono_400Hz.raw", (400,)),
|
amine@111
|
206 three_channel=("3channel_400-800-1600Hz.raw", (400, 800, 1600)),
|
amine@111
|
207 )
|
amine@111
|
208 def test_save_raw(self, filename, frequencies):
|
amine@111
|
209 filename = "tests/data/test_16KHZ_{}".format(filename)
|
amine@111
|
210 sample_width = 2
|
amine@111
|
211 fmt = DATA_FORMAT[sample_width]
|
amine@111
|
212 mono_channels = [PURE_TONE_DICT[freq] for freq in frequencies]
|
amine@111
|
213 data = _array_to_bytes(array(fmt, _sample_generator(*mono_channels)))
|
amine@111
|
214 tmpfile = NamedTemporaryFile()
|
amine@111
|
215 _save_raw(tmpfile.name, data)
|
amine@111
|
216 self.assertTrue(filecmp.cmp(tmpfile.name, filename, shallow=False))
|
amine@111
|
217
|
amine@121
|
218 def test_from_file_no_pydub(self):
|
amine@121
|
219 with patch("auditok.io._WITH_PYDUB", False):
|
amine@121
|
220 with self.assertRaises(AudioIOError):
|
amine@121
|
221 from_file("audio", "mp3")
|
amine@121
|
222
|
amine@111
|
223 @genty_dataset(
|
amine@122
|
224 raw_first_channel=("raw", 0, 400),
|
amine@122
|
225 raw_second_channel=("raw", 1, 800),
|
amine@122
|
226 raw_third_channel=("raw", 2, 1600),
|
amine@122
|
227 raw_left_channel=("raw", "left", 400),
|
amine@122
|
228 raw_right_channel=("raw", "right", 800),
|
amine@122
|
229 wav_first_channel=("wav", 0, 400),
|
amine@122
|
230 wav_second_channel=("wav", 1, 800),
|
amine@122
|
231 wav_third_channel=("wav", 2, 1600),
|
amine@122
|
232 wav_left_channel=("wav", "left", 400),
|
amine@122
|
233 wav_right_channel=("wav", "right", 800),
|
amine@122
|
234 )
|
amine@122
|
235 def test_from_file_multichannel_audio(
|
amine@122
|
236 self, audio_format, use_channel, frequency
|
amine@122
|
237 ):
|
amine@122
|
238 expected = PURE_TONE_DICT[frequency]
|
amine@122
|
239 filename = "tests/data/test_16KHZ_3channel_400-800-1600Hz.{}".format(
|
amine@122
|
240 audio_format
|
amine@122
|
241 )
|
amine@122
|
242 sample_width = 2
|
amine@122
|
243 audio_source = from_file(
|
amine@122
|
244 filename,
|
amine@122
|
245 sampling_rate=16000,
|
amine@122
|
246 sample_width=sample_width,
|
amine@122
|
247 channels=3,
|
amine@122
|
248 use_channel=use_channel,
|
amine@122
|
249 )
|
amine@122
|
250 fmt = DATA_FORMAT[sample_width]
|
amine@122
|
251 data = array(fmt, audio_source._buffer)
|
amine@122
|
252 self.assertEqual(data, expected)
|
amine@122
|
253
|
amine@122
|
254 @genty_dataset(
|
amine@123
|
255 raw_mono=("raw", "mono_400Hz", (400,)),
|
amine@123
|
256 raw_3channel=("raw", "3channel_400-800-1600Hz", (400, 800, 1600)),
|
amine@123
|
257 wav_mono=("wav", "mono_400Hz", (400,)),
|
amine@123
|
258 wav_3channel=("wav", "3channel_400-800-1600Hz", (400, 800, 1600)),
|
amine@123
|
259 )
|
amine@123
|
260 def test_from_file_multichannel_audio_mix(
|
amine@123
|
261 self, audio_format, filename_suffix, frequencies
|
amine@123
|
262 ):
|
amine@123
|
263 sampling_rate = 16000
|
amine@123
|
264 sample_width = 2
|
amine@123
|
265 channels = len(frequencies)
|
amine@123
|
266 mono_channels = [PURE_TONE_DICT[freq] for freq in frequencies]
|
amine@123
|
267 channels = len(frequencies)
|
amine@123
|
268 fmt = DATA_FORMAT[sample_width]
|
amine@123
|
269 expected = _array_to_bytes(
|
amine@123
|
270 array(
|
amine@123
|
271 fmt,
|
amine@123
|
272 (sum(samples) // channels for samples in zip(*mono_channels)),
|
amine@123
|
273 )
|
amine@123
|
274 )
|
amine@123
|
275 filename = "tests/data/test_16KHZ_{}.{}".format(
|
amine@123
|
276 filename_suffix, audio_format
|
amine@123
|
277 )
|
amine@123
|
278 audio_source = from_file(
|
amine@123
|
279 filename,
|
amine@123
|
280 use_channel="mix",
|
amine@123
|
281 sampling_rate=sampling_rate,
|
amine@123
|
282 sample_width=2,
|
amine@123
|
283 channels=channels,
|
amine@123
|
284 )
|
amine@123
|
285 mixed = audio_source._buffer
|
amine@123
|
286 self.assertEqual((mixed), expected)
|
amine@123
|
287
|
amine@124
|
288 @patch("auditok.io._WITH_PYDUB", True)
|
amine@124
|
289 @patch("auditok.io.BufferAudioSource")
|
amine@124
|
290 @genty_dataset(
|
amine@124
|
291 ogg_first_channel=("ogg", 0, "from_ogg"),
|
amine@124
|
292 ogg_second_channel=("ogg", 1, "from_ogg"),
|
amine@124
|
293 ogg_mix=("ogg", "mix", "from_ogg"),
|
amine@124
|
294 ogg_default=("ogg", None, "from_ogg"),
|
amine@124
|
295 mp3_left_channel=("mp3", "left", "from_mp3"),
|
amine@124
|
296 mp3_right_channel=("mp3", "right", "from_mp3"),
|
amine@124
|
297 flac_first_channel=("flac", 0, "from_file"),
|
amine@124
|
298 flac_second_channel=("flac", 1, "from_file"),
|
amine@124
|
299 flv_left_channel=("flv", "left", "from_flv"),
|
amine@124
|
300 webm_right_channel=("webm", "right", "from_file"),
|
amine@124
|
301 )
|
amine@124
|
302 def test_from_file_multichannel_audio_compressed(
|
amine@124
|
303 self, audio_format, use_channel, function, *mocks
|
amine@124
|
304 ):
|
amine@124
|
305 filename = "audio.{}".format(audio_format)
|
amine@124
|
306 segment_mock = Mock()
|
amine@124
|
307 segment_mock.sample_width = 2
|
amine@124
|
308 segment_mock.channels = 2
|
amine@124
|
309 segment_mock._data = b"abcd"
|
amine@124
|
310 with patch("auditok.io._extract_selected_channel") as ext_mock:
|
amine@124
|
311 with patch(
|
amine@124
|
312 "auditok.io.AudioSegment.{}".format(function)
|
amine@124
|
313 ) as open_func:
|
amine@124
|
314 open_func.return_value = segment_mock
|
amine@124
|
315 from_file(filename, use_channel=use_channel)
|
amine@124
|
316 self.assertTrue(open_func.called)
|
amine@124
|
317 self.assertTrue(ext_mock.called)
|
amine@124
|
318
|
amine@124
|
319 use_channel = {"left": 0, "right": 1, None: 0}.get(
|
amine@124
|
320 use_channel, use_channel
|
amine@124
|
321 )
|
amine@124
|
322 ext_mock.assert_called_with(
|
amine@124
|
323 segment_mock._data,
|
amine@124
|
324 segment_mock.channels,
|
amine@124
|
325 segment_mock.sample_width,
|
amine@124
|
326 use_channel,
|
amine@124
|
327 )
|
amine@124
|
328
|
amine@124
|
329 with patch("auditok.io._extract_selected_channel") as ext_mock:
|
amine@124
|
330 with patch(
|
amine@124
|
331 "auditok.io.AudioSegment.{}".format(function)
|
amine@124
|
332 ) as open_func:
|
amine@124
|
333 segment_mock.channels = 1
|
amine@124
|
334 open_func.return_value = segment_mock
|
amine@124
|
335 from_file(filename, use_channel=use_channel)
|
amine@124
|
336 self.assertTrue(open_func.called)
|
amine@124
|
337 self.assertFalse(ext_mock.called)
|
amine@124
|
338
|
amine@125
|
339 @patch("auditok.io._WITH_PYDUB", True)
|
amine@125
|
340 @patch("auditok.io.BufferAudioSource")
|
amine@125
|
341 @genty_dataset(
|
amine@125
|
342 ogg=("ogg", "from_ogg"),
|
amine@125
|
343 mp3=("mp3", "from_mp3"),
|
amine@125
|
344 flac=("flac", "from_file"),
|
amine@125
|
345 )
|
amine@125
|
346 def test_from_file_multichannel_audio_mix_compressed(
|
amine@125
|
347 self, audio_format, function, *mocks
|
amine@125
|
348 ):
|
amine@125
|
349 filename = "audio.{}".format(audio_format)
|
amine@125
|
350 segment_mock = Mock()
|
amine@125
|
351 segment_mock.sample_width = 2
|
amine@125
|
352 segment_mock.channels = 2
|
amine@125
|
353 segment_mock._data = b"abcd"
|
amine@125
|
354 with patch("auditok.io._mix_audio_channels") as mix_mock:
|
amine@125
|
355 with patch(
|
amine@125
|
356 "auditok.io.AudioSegment.{}".format(function)
|
amine@125
|
357 ) as open_func:
|
amine@125
|
358 open_func.return_value = segment_mock
|
amine@125
|
359 from_file(filename, use_channel="mix")
|
amine@125
|
360 self.assertTrue(open_func.called)
|
amine@125
|
361 mix_mock.assert_called_with(
|
amine@125
|
362 segment_mock._data,
|
amine@125
|
363 segment_mock.channels,
|
amine@125
|
364 segment_mock.sample_width,
|
amine@125
|
365 )
|
amine@125
|
366
|
amine@123
|
367 @genty_dataset(
|
amine@126
|
368 dafault_first_channel=(None, 400),
|
amine@126
|
369 first_channel=(0, 400),
|
amine@126
|
370 second_channel=(1, 800),
|
amine@126
|
371 third_channel=(2, 1600),
|
amine@126
|
372 negative_first_channel=(-3, 400),
|
amine@126
|
373 negative_second_channel=(-2, 800),
|
amine@126
|
374 negative_third_channel=(-1, 1600),
|
amine@126
|
375 )
|
amine@126
|
376 def test_load_raw(self, use_channel, frequency):
|
amine@126
|
377 filename = "tests/data/test_16KHZ_3channel_400-800-1600Hz.raw"
|
amine@126
|
378 if use_channel is not None:
|
amine@126
|
379 audio_source = _load_raw(
|
amine@126
|
380 filename,
|
amine@126
|
381 sampling_rate=16000,
|
amine@126
|
382 sample_width=2,
|
amine@126
|
383 channels=3,
|
amine@126
|
384 use_channel=use_channel,
|
amine@126
|
385 )
|
amine@126
|
386 else:
|
amine@126
|
387 audio_source = _load_raw(
|
amine@126
|
388 filename, sampling_rate=16000, sample_width=2, channels=3
|
amine@126
|
389 )
|
amine@126
|
390 self.assertIsInstance(audio_source, BufferAudioSource)
|
amine@126
|
391 self.assertEqual(audio_source.sampling_rate, 16000)
|
amine@126
|
392 self.assertEqual(audio_source.sample_width, 2)
|
amine@126
|
393 self.assertEqual(audio_source.channels, 1)
|
amine@126
|
394 # generate a pure sine wave tone of the given frequency
|
amine@126
|
395 expected = PURE_TONE_DICT[frequency]
|
amine@126
|
396 # compre with data read from file
|
amine@126
|
397 fmt = DATA_FORMAT[2]
|
amine@126
|
398 data = array(fmt, audio_source._buffer)
|
amine@126
|
399 self.assertEqual(data, expected)
|
amine@126
|
400
|
amine@126
|
401 @genty_dataset(
|
amine@127
|
402 mono=("mono_400Hz", (400,)),
|
amine@127
|
403 three_channel=("3channel_400-800-1600Hz", (400, 800, 1600)),
|
amine@127
|
404 )
|
amine@127
|
405 def test_load_raw_mix(self, filename_suffix, frequencies):
|
amine@127
|
406 sampling_rate = 16000
|
amine@127
|
407 sample_width = 2
|
amine@127
|
408 channels = len(frequencies)
|
amine@127
|
409 mono_channels = [PURE_TONE_DICT[freq] for freq in frequencies]
|
amine@127
|
410
|
amine@127
|
411 fmt = DATA_FORMAT[sample_width]
|
amine@127
|
412 expected = _array_to_bytes(
|
amine@127
|
413 array(
|
amine@127
|
414 fmt,
|
amine@127
|
415 (sum(samples) // channels for samples in zip(*mono_channels)),
|
amine@127
|
416 )
|
amine@127
|
417 )
|
amine@127
|
418 filename = "tests/data/test_16KHZ_{}.raw".format(filename_suffix)
|
amine@127
|
419 audio_source = _load_raw(
|
amine@127
|
420 filename,
|
amine@127
|
421 use_channel="mix",
|
amine@127
|
422 sampling_rate=sampling_rate,
|
amine@127
|
423 sample_width=2,
|
amine@127
|
424 channels=channels,
|
amine@127
|
425 )
|
amine@127
|
426 mixed = audio_source._buffer
|
amine@127
|
427 self.assertEqual(mixed, expected)
|
amine@127
|
428 self.assertIsInstance(audio_source, BufferAudioSource)
|
amine@127
|
429 self.assertEqual(audio_source.sampling_rate, sampling_rate)
|
amine@127
|
430 self.assertEqual(audio_source.sample_width, sample_width)
|
amine@127
|
431 self.assertEqual(audio_source.channels, 1)
|
amine@127
|
432
|
amine@127
|
433 @genty_dataset(
|
amine@128
|
434 missing_sampling_rate=("sr",),
|
amine@128
|
435 missing_sample_width=("sw",),
|
amine@128
|
436 missing_channels=("ch",),
|
amine@128
|
437 )
|
amine@128
|
438 def test_load_raw_missing_audio_param(self, missing_param):
|
amine@128
|
439 with self.assertRaises(AudioParameterError):
|
amine@128
|
440 params = AUDIO_PARAMS_SHORT.copy()
|
amine@128
|
441 del params[missing_param]
|
amine@128
|
442 srate, swidth, channels, _ = _get_audio_parameters(params)
|
amine@128
|
443 _load_raw("audio", srate, swidth, channels)
|
amine@128
|
444
|
amine@128
|
445 @genty_dataset(
|
amine@129
|
446 dafault_first_channel=(None, 400),
|
amine@129
|
447 first_channel=(0, 400),
|
amine@129
|
448 second_channel=(1, 800),
|
amine@129
|
449 third_channel=(2, 1600),
|
amine@129
|
450 negative_first_channel=(-3, 400),
|
amine@129
|
451 negative_second_channel=(-2, 800),
|
amine@129
|
452 negative_third_channel=(-1, 1600),
|
amine@129
|
453 )
|
amine@129
|
454 def test_load_wave(self, use_channel, frequency):
|
amine@129
|
455 filename = "tests/data/test_16KHZ_3channel_400-800-1600Hz.wav"
|
amine@129
|
456 if use_channel is not None:
|
amine@129
|
457 audio_source = _load_wave(filename, use_channel=use_channel)
|
amine@129
|
458 else:
|
amine@129
|
459 audio_source = _load_wave(filename)
|
amine@129
|
460 self.assertIsInstance(audio_source, BufferAudioSource)
|
amine@129
|
461 self.assertEqual(audio_source.sampling_rate, 16000)
|
amine@129
|
462 self.assertEqual(audio_source.sample_width, 2)
|
amine@129
|
463 self.assertEqual(audio_source.channels, 1)
|
amine@129
|
464 # generate a pure sine wave tone of the given frequency
|
amine@129
|
465 expected = PURE_TONE_DICT[frequency]
|
amine@129
|
466 # compre with data read from file
|
amine@129
|
467 fmt = DATA_FORMAT[2]
|
amine@129
|
468 data = array(fmt, audio_source._buffer)
|
amine@129
|
469 self.assertEqual(data, expected)
|
amine@129
|
470
|
amine@129
|
471 @genty_dataset(
|
amine@130
|
472 mono=("mono_400Hz", (400,)),
|
amine@130
|
473 three_channel=("3channel_400-800-1600Hz", (400, 800, 1600)),
|
amine@130
|
474 )
|
amine@130
|
475 def test_load_wave_mix(self, filename_suffix, frequencies):
|
amine@130
|
476 sampling_rate = 16000
|
amine@130
|
477 sample_width = 2
|
amine@130
|
478 channels = len(frequencies)
|
amine@130
|
479 mono_channels = [PURE_TONE_DICT[freq] for freq in frequencies]
|
amine@130
|
480 fmt = DATA_FORMAT[sample_width]
|
amine@130
|
481 expected = _array_to_bytes(
|
amine@130
|
482 array(
|
amine@130
|
483 fmt,
|
amine@130
|
484 (sum(samples) // channels for samples in zip(*mono_channels)),
|
amine@130
|
485 )
|
amine@130
|
486 )
|
amine@130
|
487 filename = "tests/data/test_16KHZ_{}.wav".format(filename_suffix)
|
amine@130
|
488 audio_source = _load_wave(filename, use_channel="mix")
|
amine@130
|
489 mixed = audio_source._buffer
|
amine@130
|
490 self.assertEqual(mixed, expected)
|
amine@130
|
491 self.assertIsInstance(audio_source, BufferAudioSource)
|
amine@130
|
492 self.assertEqual(audio_source.sampling_rate, sampling_rate)
|
amine@130
|
493 self.assertEqual(audio_source.sample_width, sample_width)
|
amine@130
|
494 self.assertEqual(audio_source.channels, 1)
|
amine@130
|
495
|
amine@131
|
496 @patch("auditok.io._WITH_PYDUB", True)
|
amine@131
|
497 @patch("auditok.io.BufferAudioSource")
|
amine@131
|
498 @genty_dataset(
|
amine@131
|
499 ogg_default_first_channel=("ogg", 2, None, "from_ogg"),
|
amine@131
|
500 ogg_first_channel=("ogg", 1, 0, "from_ogg"),
|
amine@131
|
501 ogg_second_channel=("ogg", 2, 1, "from_ogg"),
|
amine@131
|
502 ogg_mix_channels=("ogg", 3, "mix", "from_ogg"),
|
amine@131
|
503 mp3_left_channel=("mp3", 1, "left", "from_mp3"),
|
amine@131
|
504 mp3_right_channel=("mp3", 2, "right", "from_mp3"),
|
amine@131
|
505 mp3_mix_channels=("mp3", 3, "mix", "from_mp3"),
|
amine@131
|
506 flac_first_channel=("flac", 2, 0, "from_file"),
|
amine@131
|
507 flac_second_channel=("flac", 2, 1, "from_file"),
|
amine@131
|
508 flv_left_channel=("flv", 1, "left", "from_flv"),
|
amine@131
|
509 webm_right_channel=("webm", 2, "right", "from_file"),
|
amine@131
|
510 webm_mix_channels=("webm", 4, "mix", "from_file"),
|
amine@131
|
511 )
|
amine@131
|
512 def test_load_with_pydub(
|
amine@131
|
513 self, audio_format, channels, use_channel, function, *mocks
|
amine@131
|
514 ):
|
amine@131
|
515 filename = "audio.{}".format(audio_format)
|
amine@131
|
516 segment_mock = Mock()
|
amine@131
|
517 segment_mock.sample_width = 2
|
amine@131
|
518 segment_mock.channels = channels
|
amine@131
|
519 segment_mock._data = b"abcdefgh"
|
amine@131
|
520 with patch("auditok.io._extract_selected_channel") as ext_mock:
|
amine@131
|
521 with patch(
|
amine@131
|
522 "auditok.io.AudioSegment.{}".format(function)
|
amine@131
|
523 ) as open_func:
|
amine@131
|
524 open_func.return_value = segment_mock
|
amine@131
|
525 use_channel = {"left": 0, "right": 1, None: 0}.get(
|
amine@131
|
526 use_channel, use_channel
|
amine@131
|
527 )
|
amine@131
|
528 _load_with_pydub(filename, audio_format, use_channel)
|
amine@131
|
529 self.assertTrue(open_func.called)
|
amine@131
|
530 if channels > 1:
|
amine@131
|
531 self.assertTrue(ext_mock.called)
|
amine@131
|
532 ext_mock.assert_called_with(
|
amine@131
|
533 segment_mock._data,
|
amine@131
|
534 segment_mock.channels,
|
amine@131
|
535 segment_mock.sample_width,
|
amine@131
|
536 use_channel,
|
amine@131
|
537 )
|
amine@131
|
538 else:
|
amine@131
|
539 self.assertFalse(ext_mock.called)
|
amine@131
|
540
|
amine@130
|
541 @genty_dataset(
|
amine@110
|
542 mono=("mono_400Hz.wav", (400,)),
|
amine@110
|
543 three_channel=("3channel_400-800-1600Hz.wav", (400, 800, 1600)),
|
amine@110
|
544 )
|
amine@110
|
545 def test_save_wave(self, filename, frequencies):
|
amine@110
|
546 filename = "tests/data/test_16KHZ_{}".format(filename)
|
amine@110
|
547 sampling_rate = 16000
|
amine@110
|
548 sample_width = 2
|
amine@110
|
549 channels = len(frequencies)
|
amine@110
|
550 fmt = DATA_FORMAT[sample_width]
|
amine@110
|
551 mono_channels = [PURE_TONE_DICT[freq] for freq in frequencies]
|
amine@110
|
552 data = _array_to_bytes(array(fmt, _sample_generator(*mono_channels)))
|
amine@110
|
553 tmpfile = NamedTemporaryFile()
|
amine@110
|
554 _save_wave(tmpfile.name, data, sampling_rate, sample_width, channels)
|
amine@110
|
555 self.assertTrue(filecmp.cmp(tmpfile.name, filename, shallow=False))
|