135:
108:
143:
applied to the 5-speaker three-dimensional sound system, as in the listening room technique. The system also convolves the head-related transfer function with the impulse response from the signal recorded by the microphones and the energy is adjusted per the original time frame of the sound signal, and an additional
142:
The reverberation reconstruction involves measuring the sound by a four-point microphone to measure its real delivery delays in different locations. Each microphone measures an impulse response from a time-stretched pulse signal for various time frames with various sound sources. The obtained data is
150:
This technique also improves the directionality, naturalness, and clarity of the reconstructed sound with respect to the original. A drawback of this method is that the assumption of a single sound source—while real-life reverberations include various sounds with overlap—coupled with adding all the
90:
After identifying the direction, other signal processing techniques are used to measure the impulse response over lengths of time to determine the intensity components in different directions. By having both data and combining intensity of sound with direction, a three-dimensional sound field is
115:
In the listening room method, the listener receives the sound either through headphones or through loudspeakers. Headphones introduce enough sound sources for a listener to experience 3D sound with directionality. With loudspeakers, the placement and number of loudspeakers affects the depth of
195:
This method is applicable primarily to ultrasound and to lower sound pressures, often in water and in medical imaging. The method works under the assumption that the wave number of the medium is constant. If the wave number is changing throughout the medium, this method cannot reconstruct the
95:
As a result of this two-step process, the reconstructed three-dimensional sound field contains information not only on the localization of the sound source, but also on the physical aspects of the environment of the original signal source. This is its difference from the results of the sound
125:
can be used on the sound source signal to pan its convolution to each of the loudspeakers depending on their direction and location. This allows the calculation of the energy of signal for each speaker through evaluation of sound in several control points within the listening room.
120:
recommended formation: center, 30° to the left, 110° to the left, 30° to the right, and 110° to the right. This setup is used with several three-dimensional sound systems and reconstruction techniques. As an alternative, the
376:
Ohbuchi; Mizutani; Wakatsuki; Nishimiya; Masuyama (2009). "Reconstruction of Three-Dimensional Sound Field from Two-Dimensional Sound Field Using
Optical Computerized Tomography and Near-Field Acoustical Holography".
192:. Then the wave number of the medium is estimated through analysis of the water temperature. Multiple two-dimensional sound fields are calculated, and the three-dimensional sound field can be reconstructed as well.
163:
signal processing to reconstruct the sound field. These measurements can be done using projections, eliminating the need to use multiple microphones to determine separate impulse responses. These projectors use a
147:
is added to the sound to match the time frame of the impulse response. The convolution and delays are applied to all the sound source data taken and summed for the resulting signal.
99:
After the sound is reconstructed and the spatial cues are available, they need to be delivered to the customer. The different methods to do this are included in this section.
249:
536:
438:
184:
In near-field acoustical holography, light refraction is measured in a two-dimensional area in the medium (this two-dimensional sound field is a
35:
to match natural environments and provide spatial cues of the sound source. They also see applications in creating 3D visualizations on a sound
468:
352:
308:
267:
159:
As the sound waves cause changes in air density, it subsequently causes sound pressure changes. They are measured and then processed using
47:. This technology is used in entertainment to reproduce a live performance through computer speakers. The technology is also used in
116:
reproduction. There are various methods to select the speakers location. A simple model consists of five speakers, placed in the
337:
2013 International Joint
Conference on Awareness Science and Technology & Ubi-Media Computing (ICAST 2013 & UMEDIA 2013)
220:
122:
84:
151:
different values does not improve listeners perception of the size of the room, the perception of distance is not improved.
248:
Oikawa; Goto; Ikeda; Takizawa; Yamasaki (2005). "Sound Field
Measurements Based on Reconstruction from Laser Projections".
210:
473:
431:
424:
225:
169:
67:
To reproduce robust and natural-sounding audio from a three-dimensional audio recording, sound localization and
185:
165:
335:
Tanno; Saiji; Huang (2013). "A new 5-loudspeaker 3D sound system with a reverberation reconstruction method".
134:
72:
251:
Proceedings. (ICASSP '05). IEEE International
Conference on Acoustics, Speech, and Signal Processing, 2005
510:
205:
80:
24:
20:
458:
386:
168:
to measure the refractive index of the medium on the laser path. These measurements are processed by
493:
91:
determined and physical qualities that create the resulting changes in intensity are reconstructed.
36:
498:
402:
358:
314:
273:
291:
Kim; Jee; Park; Yoon; Choi (2004). "The real-time implementation of 3D sound system using DSP".
505:
483:
447:
348:
304:
263:
52:
488:
394:
340:
296:
255:
215:
173:
56:
44:
390:
478:
40:
530:
406:
68:
362:
277:
71:
reconstruction techniques are used. These techniques process sound to reproduce the
463:
318:
344:
300:
259:
107:
189:
160:
28:
398:
48:
416:
293:
IEEE 60th
Vehicular Technology Conference, 2004. VTC2004-Fall. 2004
133:
117:
106:
32:
83:
using multiple microphone arrays, binaural hearing methods, and
39:
to include physical aspects of sound waves including direction,
420:
172:
to reproduce the three-dimensional sound field, and then the
138:
79:
The location of a sound source is determined through
55:. Reconstructing sound fields is also applicable to
188:of the three-dimensional sound field) to produce a
27:technology. These methods of reconstructing three-
111:Loudspeaker location from ITU-R recommendation
432:
196:three-dimensional sound field as accurately.
8:
439:
425:
417:
243:
241:
330:
328:
254:. Vol. 4. pp. iv/661–iv/664.
237:
469:Computational auditory scene analysis
85:HRTF (head-related transfer function)
7:
81:three-dimensional sound localization
379:Japanese Journal of Applied Physics
537:Multidimensional signal processing
295:. Vol. 7. pp. 4798–480.
14:
59:to measure points in ultrasound.
216:Convolution Back Projection(CBP)
180:Near-field acoustical holography
221:Head-related transfer function
123:head-related transfer function
1:
211:Acoustic source localization
130:Reverberation reconstruction
474:Music information retrieval
345:10.1109/ICAwST.2013.6765429
301:10.1109/VETECF.2004.1405005
260:10.1109/ICASSP.2005.1416095
174:convolution back projection
31:sound are used to recreate
553:
226:Tomographic reconstruction
170:Tomographic reconstruction
51:applications to determine
454:
176:is used to visualize it.
53:location of sound sources
166:laser Doppler vibrometer
516:3D sound reconstruction
17:3D sound reconstruction
399:10.1143/JJAP.48.07GC03
139:
112:
96:localization process.
19:is the application of
511:3D sound localization
206:3D sound localization
137:
110:
25:3D sound localization
459:Acoustic fingerprint
339:. pp. 174–179.
494:Speaker recognition
391:2009JaJAP..48gGC03O
499:Speech recognition
140:
113:
524:
523:
506:Sound recognition
484:Speech processing
448:Computer audition
354:978-1-4799-2364-9
310:978-0-7803-8521-4
269:978-0-7803-8874-1
155:Laser projections
544:
489:Speech analytics
441:
434:
427:
418:
411:
410:
373:
367:
366:
332:
323:
322:
288:
282:
281:
245:
552:
551:
547:
546:
545:
543:
542:
541:
527:
526:
525:
520:
450:
445:
415:
414:
375:
374:
370:
355:
334:
333:
326:
311:
290:
289:
285:
270:
247:
246:
239:
234:
202:
182:
157:
132:
105:
65:
57:medical imaging
12:
11:
5:
550:
548:
540:
539:
529:
528:
522:
521:
519:
518:
513:
508:
503:
502:
501:
496:
491:
481:
479:Semantic audio
476:
471:
466:
461:
455:
452:
451:
446:
444:
443:
436:
429:
421:
413:
412:
368:
353:
324:
309:
283:
268:
236:
235:
233:
230:
229:
228:
223:
218:
213:
208:
201:
198:
181:
178:
156:
153:
131:
128:
104:
103:Listening room
101:
93:
92:
88:
64:
61:
23:techniques to
21:reconstruction
13:
10:
9:
6:
4:
3:
2:
549:
538:
535:
534:
532:
517:
514:
512:
509:
507:
504:
500:
497:
495:
492:
490:
487:
486:
485:
482:
480:
477:
475:
472:
470:
467:
465:
462:
460:
457:
456:
453:
449:
442:
437:
435:
430:
428:
423:
422:
419:
408:
404:
400:
396:
392:
388:
384:
380:
372:
369:
364:
360:
356:
350:
346:
342:
338:
331:
329:
325:
320:
316:
312:
306:
302:
298:
294:
287:
284:
279:
275:
271:
265:
261:
257:
253:
252:
244:
242:
238:
231:
227:
224:
222:
219:
217:
214:
212:
209:
207:
204:
203:
199:
197:
193:
191:
187:
186:cross section
179:
177:
175:
171:
167:
162:
154:
152:
148:
146:
136:
129:
127:
124:
119:
109:
102:
100:
97:
89:
86:
82:
78:
77:
76:
74:
70:
69:reverberation
62:
60:
58:
54:
50:
46:
42:
38:
34:
30:
26:
22:
18:
515:
464:Audio mining
382:
378:
371:
336:
292:
286:
250:
194:
183:
158:
149:
144:
141:
114:
98:
94:
66:
16:
15:
29:dimensional
232:References
161:tomography
63:Techniques
407:119815337
385:(7): 07.
45:intensity
531:Category
363:11582154
278:15044296
200:See also
190:hologram
49:military
41:pressure
387:Bibcode
319:9906064
73:spatial
405:
361:
351:
317:
307:
276:
266:
75:cues.
43:, and
33:sounds
403:S2CID
359:S2CID
315:S2CID
274:S2CID
145:delay
118:ITU-R
37:field
349:ISBN
305:ISBN
264:ISBN
395:doi
341:doi
297:doi
256:doi
533::
401:.
393:.
383:48
381:.
357:.
347:.
327:^
313:.
303:.
272:.
262:.
240:^
440:e
433:t
426:v
409:.
397::
389::
365:.
343::
321:.
299::
280:.
258::
87:.
Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.