Virtual Acoustic Reality in Architectural Design

사운드 디자인 / VR 체험

Within the Architecture Engineering and Construction industry (AEC), presenting design ideas through 3D models is now common in the design process and is a valuable tool for communication of ideas and collaboration. 

3D models could also be used in acoustical simulation software to calculate the sound propagation in a space from a source to a receiver. The information of how sound propagates from a source to a receiver is contained in the so-called Room Impulse Response (RIR), which can be used to simulate, through convolution, how a sound, a music track for instance, is perceived by a listener in that room. This process is called “auralization”.

Auralizations have been part of the architectural and acoustical design process for a while now, but they are usually limited to static (no headtracking) binaural representations of the soundfield in one source-receiver location.

Now that new immersive technologies are gaining ground in architecture and commercial and residential cinema space too, and VR/AR applications are finding their way as design tools, simulating sound propagation in the virtual space is crucial to keep the virtual reality experience intact. Sound and acoustics play a fundamental role to recreate realistic experiences in non-static immersive technologies where audio has to continuously adapt to the listener’s head orientation. 

Simulating the acoustics of a space could be a powerful tool to inform design decisions, such as material and shape selection. Sound interacts differently with each surface and informs our brain of our surroundings. Sound absorbing, scattering, and reflecting materials are used by designers to control objective or subjective parameters as loudness, reverberation, speech intelligibility and acoustical capacity of a space. The acoustical comfort of a space supports the function for which the room was designed, and it is important for the health of its users.  When music is reproduced in a room, the audience needs a calibrated and clear response from the room avoiding detrimental reflections or excessive coloring of sound. A fragile relation must be balanced between sound and space.

At Keith Yates Design, we set up a demo of an acoustically untreated home theatre to explore new possibilities in virtual reality with Wwise and Unity, using their powerful, well documented, and smooth integration that allows for a dynamic gaming-like experience of a space. Acoustical simulation of the room is performed with CATT from which soundfield B-Format (i.e. Ambisonics) Room Impulse Response are exported and subsequently imported in Wwise. The room is simulated without any acoustical treatment panel to control the reverberation, sound decay and spatial envelopment, that are included in the final design.

Overview of the system

We envisioned a VR experience of a home theatre as a 3-DOF app. Giving the ability to walk through the room was not essential and would have increased the complexity of the demo itself. Generally, acoustical demos are aimed at comparing seats in a room and consistency of the soundfield in multiple positions.

Visual renderings for sightlines and auralizations are typically two separate outputs obtained from different software and are generally static with no possibility of listening or looking around. We wanted to cross that bridge and take this process one step further.

As the demo starts up, the scene is set inside the home theatre with, for the purpose of this demo, two screen speakers (left and right channel) that can be switched on from a UI menu (Figure 1).

Figure 1 – Screen speakers (left and right) highlighted in red. These are the AkAmbient (simple mode) sound sources.

The player/listener is in the audience seats and can be “teleported” to other seats and experiment the space by listening and looking around.

Playthrough and look-around of the demo. Both left and right screen speakers are playing the stereo tracks. Better experienced with headphones.


Figure 2 – Block diagram of the system.

The impulse response of the source-receiver pair is calculated beforehand and has to be fed into Wwise when preparing the scene. Thus, once source and receiver positions were selected, we used CATT (software used professionally for acoustical simulations in ray tracing) to precompute accurate impulse responses of the room (Figure 3a, 3b), considering the 3D radiation pattern of the selected speakers and their specific aiming at the audience seats (Figure 4).

3_image10Figure 3a – Spatial analysis app of B-Format 1st order room impulse response developed in Matlab by the author. The hedgehog plot shows the module and direction of arrival of sound intensity vectors and their colour indicates the time of arrival. Sound reflections are identified and highlighted with the respective colour in the RIR waveform in Figure 3b.

Figure 3b –First 70ms of the W (pressure) channel of the room impulse response with right loudspeaker as source and receiver at mid row middle seat.   





Figure 4 – Selected speaker model of the room and an example of the radiation pattern data of a flush mounted screen speaker.

Simulations of sound propagation can take hours in CATT as it uses hundreds of thousands of “sound rays” interacting with purpose made models and their surfaces. Each surface has assigned material properties (absorption and scattering) in frequency octave bands. The algorithm used by the software and its tools achieve a simulation accuracy that goes beyond “simplified” methods recently introduced in some audio middleware for gaming. For this reason, we decided not to use Wwise Reflect, or Steam Audio, as we cannot compromise on acoustical accuracy resulting from a specific architectural design. Wwise Reflect, instead, could be very useful for fast prototyping of spaces to analyze early sound reflections and Speaker Boundary Interference Response. We are currently experimenting with it to simulate how direct sound and early reflections interfere with each other at the listening position without having to create an acoustical model for CATT.

Time trace simulation of sound propagation of the left and right speakers. The simulation is truncated at 100ms with “sound rays” shown up to the 5th order of reflection.

Once the room’s acoustical response is “baked” for several source-receiver pairs, we imported the results in Wwise and we used the Convolution Reverb plugin and Wwise Recorder to obtain the convoluted ambisonics tracks.

Convolutions with Convolution Reverb could be done in real time, but this would put a burden on computational resources and memory pools. It may be worth to use this approach if different RIR files have to be switched at run time using a script. However, this is not found to be straightforward and required knowledge of Wwise APIs and specific plugin mechanics. 

Figure 5 – Schematic View and an example of a Mixing Session in Wwise to visualize sound sfx, bus routing, active plugins attached to each bus and active states.

The convoluted ambisonics tracks are assigned to an nth order Ambisonics output bus in Wwise which is a child of a “Binauralizer” bus that takes as input the ambisonics track and outputs a binaural track to the Master Audio Bus. The Binauralizer bus uses the Auro-3D Headphone plugin to render the binaural output, but special care should be taken in setting up the parameters of the plugin if we don’t want to change the RIR. The speaker distance should be set to be greater than the room dimensions so that Auro Headphone’s algorithm would not generate additional early reflections and the Reverb option should be disabled.

Figure 6 – Auro-3D Headphones settings used in the demo.

The ambisonics tracks positioning must be set to 3D spatialization with respect to the emitter 3D position to allow for soundfield rotation. No attenuation should be used as that information is already contained in the simulated IRs while the Spread curve should always be 100% for a correct spatialization of the soundfield around the listener’s head. Wwise 3D Meter is a useful tool to visualize the soundfield at the listener’s position in real time.

Split screen example of Wwise connected to Unity in Profiler layout. The 3D Meter selected as ambisonics bus meter shows the soundfield around the listener in real time. For this example, pink noise bursts are played by the left speaker only for ease of visualization.

Game states and Reverb Zones are used to manage the audio tracks that play at each source-receiver position. Audio tracks are attached to emitters using AkAmbient scripts while Reverb Zone is a Unity Collider with a custom script component attached that triggers the Wwise entry and exit states. Wwise states are linked with the corresponding convoluted ambisonics soundtracks that get pushed down the Wwise routing and are played in the binaural master audio output.

Figure 7 – The green box colliders in the room are positioned where RIR per source-receiver pairs were simulated. Each collider triggers an AkState when the player enters or exits the area.


Fabrizio D’Amelio   Acoustic Engineer

 

 

Keith Yates Design

Keith Yates Design

Keith Yates Design is the world’s leading design & engineering firm specializing in high-end, purpose built private theaters and screening rooms, as well as acoustic design for residential multi-use common areas and high-spec commercial facilities. Since 1991 KYD has focused on envisioning, modeling, optimizing, documenting, testing and obsessively calibrating these exquisite venues for marquee-name directors, cinematographers, technologists and, especially, everyday movie and music lovers seeking to be transported to another place and time.

http://www.keithyates.com

댓글

댓글 달기

이메일 주소는 공개되지 않습니다.

다른 글

게임 오디오 직업 스킬 - 게임 사운드 디자이너로 고용되는 법

20.1.2021 - 작성자: 브라이언 슈밋(BRIAN SCHMIDT)

파이널 판타지 VII 리메이크의 사운드

수많은 사람들의 사랑을 받는 파이널 판타지 VII의 음악은 새로운 세대의 팬들이 경험할 수 있도록 수 년에 걸쳐 새롭게 재고되어 왔습니다. 최근 Square Enix(스퀘어...

4.5.2021 - 작성자: SQUARE ENIX (스퀘어 에닉스)

Wwise에서 Audio Object를 저작하고 프로파일링하는 간단한 9 단계

Wwise에서 새롭게 제공되는 오브젝트 기반 오디오 파이프라인을 둘러보고 싶지만 어디서부터 시작해야 할지 모르시는 분들 계시나요? 그렇다면 Windows용 Wwise에서 Audio...

21.7.2021 - 작성자: 데미안 캐스트바우어 (Damian Kastbauer)

zerocrossing의 SpectralMultiEffect

SpectralMultiEffect는 Wwise용 플러그인입니다. 이 플러그인은 게임에서 상호작용성을 향상시키기 위해 제작되었으며 사운드 디자이너가 실험해보고 오디오를 변화시킬 수...

9.2.2022 - 작성자: 하비에르 아르시니에가스(Javier Arciniegas)

Strata 멀티트랙 SFX 라이브러리를 먼저 사용해본 사람들은 이렇게 말했습니다.

Strata의 시작 지난 40~50년 동안 SFX 라이브러리 제작자들은 거의 동일한 방식으로 콘텐츠를 제작하고 배포했습니다.디자인한 사운드를 녹음, 믹싱, 렌더링하고 테마별...

1.12.2022 - 작성자: 시몽 아슈비 (Simon Ashby)

임펄스 응답 리버브로 현실감 있는 사운드 만들기

임펄스 응답은 실제 공간을 초현실적으로 재구현하는 것으로 유명합니다. 최고 품질의 임펄스 응답을 녹음하는 것은 다소 기술적이며 고급 장비가 필요한 일입니다. 공간적인 품질을 갖춘...

8.9.2023 - 작성자: BOOM Library

다른 글

게임 오디오 직업 스킬 - 게임 사운드 디자이너로 고용되는 법

파이널 판타지 VII 리메이크의 사운드

수많은 사람들의 사랑을 받는 파이널 판타지 VII의 음악은 새로운 세대의 팬들이 경험할 수 있도록 수 년에 걸쳐 새롭게 재고되어 왔습니다. 최근 Square Enix(스퀘어...

Wwise에서 Audio Object를 저작하고 프로파일링하는 간단한 9 단계

Wwise에서 새롭게 제공되는 오브젝트 기반 오디오 파이프라인을 둘러보고 싶지만 어디서부터 시작해야 할지 모르시는 분들 계시나요? 그렇다면 Windows용 Wwise에서 Audio...