Interactive Sound Rendering

Similar documents
INTERACTIVE GEOMETRIC SOUND PROPAGATION AND RENDERING

Efficient Auralization for Moving Sources and Receiver

DAAD: A New Software for Architectural Acoustic Design

The Use of Computer Modeling in Room Acoustics

Spatialized Audio Rendering for Immersive Virtual Environments

CUBE-MAP DATA STRUCTURE FOR INTERACTIVE GLOBAL ILLUMINATION COMPUTATION IN DYNAMIC DIFFUSE ENVIRONMENTS

Go to contents 18 3D Visualization of Building Services in Virtual Environment

List of Publications

Direct and Reflected: Understanding the Truth with Y-S 3

Electromagnetic Wave Simulation Software Poynting

Acoustic Terms, Definitions and General Information

Volumetric Meshes for Real Time Medical Simulations

Room Acoustic Reproduction by Spatial Room Response

Waves Sound and Light

Physical Science Study Guide Unit 7 Wave properties and behaviors, electromagnetic spectrum, Doppler Effect

Doppler Effect Plug-in in Music Production and Engineering

INTERFERENCE OF SOUND WAVES

PRODUCT INFORMATION. Insight+ Uses and Features

GRAFICA - A COMPUTER GRAPHICS TEACHING ASSISTANT. Andreas Savva, George Ioannou, Vasso Stylianou, and George Portides, University of Nicosia Cyprus

MODULE VII LARGE BODY WAVE DIFFRACTION

VIRTUAL SPACE 4D PROJECT: THE ROOM ACOUSTICAL TOOL. J. Keränen, P. Larm, V. Hongisto

Image Processing and Computer Graphics. Rendering Pipeline. Matthias Teschner. Computer Science Department University of Freiburg

Introduction to Computer Graphics

AURA - Analysis Utility for Room Acoustics Merging EASE for sound reinforcement systems and CAESAR for room acoustics

Fluid structure interaction of a vibrating circular plate in a bounded fluid volume: simulation and experiment

Parallel Simplification of Large Meshes on PC Clusters

High-fidelity electromagnetic modeling of large multi-scale naval structures

Lecture Notes, CEng 477

On One Approach to Scientific CAD/CAE Software Developing Process

Finite Element Analysis for Acoustic Behavior of a Refrigeration Compressor

Building Design for Advanced Technology Instruments Sensitive to Acoustical Noise

GEOMETRIC-BASED REVERBERATOR USING ACOUSTIC RENDERING NETWORKS

ABSTRACT FOR THE 1ST INTERNATIONAL WORKSHOP ON HIGH ORDER CFD METHODS

2.5 Physically-based Animation

ROOM ACOUSTIC PREDICTION MODELLING

A NEW PYRAMID TRACER FOR MEDIUM AND LARGE SCALE ACOUSTIC PROBLEMS

Computer Graphics AACHEN AACHEN AACHEN AACHEN. Public Perception of CG. Computer Graphics Research. Methodological Approaches

L20: GPU Architecture and Models

The Design and Implementation of Multimedia Software

THE SIMULATION OF MOVING SOUND SOURCES. John M. Chowning

Electromagnetic Sensor Design: Key Considerations when selecting CAE Software

Convention Paper Presented at the 118th Convention 2005 May Barcelona, Spain

A NEW METHOD OF STORAGE AND VISUALIZATION FOR MASSIVE POINT CLOUD DATASET

Development of Optical Wave Microphone Measuring Sound Waves with No Diaphragm

INTRODUCTION TO RENDERING TECHNIQUES

Doppler. Doppler. Doppler shift. Doppler Frequency. Doppler shift. Doppler shift. Chapter 19

Various Technics of Liquids and Solids Level Measurements. (Part 3)

CAD and Creativity. Contents

SOFTWARE FOR GENERATION OF SPECTRUM COMPATIBLE TIME HISTORY

Module 13 : Measurements on Fiber Optic Systems

A Pattern-Based Approach to. Automated Application Performance Analysis

Aircraft cabin noise synthesis for noise subjective analysis

An Overview of the Finite Element Analysis

GPR Polarization Simulation with 3D HO FDTD

Faculty of Computer Science Computer Graphics Group. Final Diploma Examination

Finite Element Method (ENGC 6321) Syllabus. Second Semester

Course Overview. CSCI 480 Computer Graphics Lecture 1. Administrative Issues Modeling Animation Rendering OpenGL Programming [Angel Ch.

Computer Networks and Internets, 5e Chapter 6 Information Sources and Signals. Introduction

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG

Acousto-optic modulator

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

8 Room and Auditorium Acoustics

STUDY OF DAM-RESERVOIR DYNAMIC INTERACTION USING VIBRATION TESTS ON A PHYSICAL MODEL

Overview Motivation and applications Challenges. Dynamic Volume Computation and Visualization on the GPU. GPU feature requests Conclusions

SOFA an Open Source Framework for Medical Simulation

Synthetic Sensing: Proximity / Distance Sensors

animation animation shape specification as a function of time

Optical modeling of finite element surface displacements using commercial software

Finite Difference Time Domain and BPM: Flexible Algorithm Selection Technology

Optical Design Tools for Backlight Displays

Stream Processing on GPUs Using Distributed Multimedia Middleware

COMP175: Computer Graphics. Lecture 1 Introduction and Display Technologies

Rendering Microgeometry with Volumetric Precomputed Radiance Transfer

System-Level Display Power Reduction Technologies for Portable Computing and Communications Devices

Back to Elements - Tetrahedra vs. Hexahedra

A Cognitive Approach to Vision for a Mobile Robot

High-Speed Thin Client Technology for Mobile Environment: Mobile RVEC

Dispersion diagrams of a water-loaded cylindrical shell obtained from the structural and acoustic responses of the sensor array along the shell

Multiresolution 3D Rendering on Mobile Devices

Off-line Model Simplification for Interactive Rigid Body Dynamics Simulations Satyandra K. Gupta University of Maryland, College Park

Medical Image Processing on the GPU. Past, Present and Future. Anders Eklund, PhD Virginia Tech Carilion Research Institute

Waves: Recording Sound Waves and Sound Wave Interference (Teacher s Guide)

What Audio Engineers Should Know About Human Sound Perception. Part 2. Binaural Effects and Spatial Hearing

Basic Theory of Intermedia Composing with Sounds and Images

Intersection of a Line and a Convex. Hull of Points Cloud

Efficient Storage, Compression and Transmission

Efficient numerical simulation of time-harmonic wave equations

Music technology. Draft GCE A level and AS subject content

Co-simulation of Microwave Networks. Sanghoon Shin, Ph.D. RS Microwave

SGN-1158 Introduction to Signal Processing Test. Solutions

HPC enabling of OpenFOAM R for CFD applications

Transcription:

Interactive Sound Rendering Dinesh Manocha Ming C. Lin Department of Computer Science University of North Carolina Chapel Hill, NC 27599-3175 {dm,lin}@cs.unc.edu http://gamma.cs.unc.edu/sound Abstract Extending the frontier of visual computing, sound rendering utilizes auditory display to communicate information to a user and offers an alternative means of visualization. By harnessing the sense of hearing, sound rendering can further enhance a user s experience in a multimodal virtual world. In addition to immersive environments, auditory displays can provide a natural and intuitive human-computer interface for many desktop applications. In this paper, we give a brief overview of recent work at UNC Chapel Hill on fast algorithms for sound synthesis and sound propagation. These include physically-based sound synthesis for rigid bodies and liquid sound generation, as well as numeric and geometric algorithms for sound propagation. We highlight their performance on different benchmarks and briefly discuss some problems for future research. 1. Introduction Traditionally, the focus in interactive applications has been on generating realistic images. These developments have been well supported by high growth rates and programmability of current graphics hardware. At the same time it is also important to develop interactive algorithms for other forms of rendering, including sound and haptics. Sound rendering utilizes auditory display to communicate information to a user and offers an alternative means of visualization. By harnessing the sense of hearing, audio rendering can further enhance a user s experience in a multimodal virtual world [17], [31]. Ultimately, the audio cues combined with visual rendering provide a far more immersive experience. In addition to immersive environments, auditory display can also provide a natural and intuitive human-computer interface for many desktop applications Furthermore, audio interfaces are increasingly used to develop assistive technologies for the visually impaired. Despite the significance of hearing as one of the dominant senses, auditory displays have not received as much attention as graphical rendering in the literature. The design of auditory display involves audio hardware technology, software rendering pipeline, modeling and simulation of acoustic spaces, signal processing, sonification techniques, perceptual evaluation, and application development. In this paper, we primarily focus on 978-1-4244-3701-6/09/$25.00 2009 IEEE two compute-intensive components of auditory displays: sound synthesis and sound propagation. Sound synthesis deals with how sound is generated using physical principles [16], [55], [35], [40], [3], [34]. Sound propagation deals with computational modeling of acoustic spaces that takes into account the knowledge of sound sources, listener locations, 3D models of the environments, and material absorption data to compute impulse response (IR) of the space [19], [48]. These IRs are convolved with recorded or synthetically produced sound signals to generate spatialized sound effects, i.e. auralization. The driving impetus of recent research in sound rendering has mainly come from interactive applications, including computer gaming, training systems, desktop interfaces, education, scientific visualization and computer-aided design. All these applications need capabilities for real-time sound synthesis and propagation techniques. Realistic sound quality can directly impact the perceived realism and full immersion of players and users of these interactive applications. Furthermore, interactive modeling and simulation of acoustic spaces can significantly enhance numerous scientific and engineering applications [11]. For examples, designers of man-made structures, such as urban layouts, habitats [54], and mechanical CAD structures [21], can benefit tremendously from interactive acoustic simulation technologies in terms of improved design and virtual prototyping. The current market for acoustic simulation software for engineering design alone is estimated to be around $300M a year. The audible sonic frequencies for humans range from 20 Hz to 20,000 Hz with wavelengths that fall exactly in the range of dimensions of common objects, i.e. a few centimeters to a few meters. Consequently sound diffracts (bends) appreciably, especially at low frequencies. Due to its low speed, sound reflections is also easily perceptible in a room in terms of delays and echos. The combination of sound diffraction and reflections makes interactive sound propagation a major computational challenge [48]. As a result, most interactive applications currently use sound effects based on fixed models of propagation or statically designed environment reverberation filers. In terms of synthesis, most interactive applications use pre-recorded sound clips for providing sounds corresponding to object interactions in a scene. Although this

approach has the advantages of sounds being realistic and the generation process is quite fast, there are many physical effects which cannot be captured by such a technique. The resulting sound often lacks variation in tones and timbers or typically appears repetitive. Physics-based sound synthesis can reproduce sound using physical principles, adding noticeable realism by introducing natural variations due to object interaction, but poses substantial computational challenges for interactive applications. Challenges in Sound Rendering: The current systems for sound synthesis and propagation can not offer interactive performance in complex, dynamic scenes or may not be able to synthesize different type of sounds. We need significant advances and also need to integrate the resulting algorithms with different applications. These include: Sound Synthesis: Most of existing synthesis approaches have focused on sound mixing or sound generation due to colliding solids in the air. We need to develop new algorithms for interactive sound synthesis techniques for complex environments consisting of hundreds of objects and sound generated in liquid medium and coupled fluid-solid interaction. Sound Propagation: The sensation of sound is due to small variations in air pressure. The variations are governed by the three-dimensional wave equation, a second-order linear partial differential equation, which relates the temporal and spatial derivatives of the pressure field [48]. We need clever domain decomposition techniques for low- and medium-frequency sources and novel geometric propagation algorithms for high-frequency sources. These algorithms can be further accelerated by using the capabilities of multi-core and many-core commodity processors for interactive computations. In the rest of this paper, we will describe some of our recent work on interactive sound synthesis and sound propagation. 2. Real-time Sound Synthesis Sound is produced by surface vibrations of an object under external forces. These vibrations travel through the surrounding medium. The pressure waves within the human audible range are sensed by the ear and render the perception of sound. One may consider the sound synthesis pipeline consists of two parts: the interaction model and the resonators. The forces from object interactions drive the resonators. The pipeline of our sound synthesis system is shown in Figure 1. The green boxes indicate the components that allow user input. The blue boxes refer to the preprocessing parts. The orange boxes show the processes that are computed at run-time. The interaction handling module detects different types of interactions among the objects and converts them into excitation forces for the sound synthesis module, which is composed of modal synthesis [35], [40], vibration calculation, and the real-time audio engine. The sound synthesis module approximates the oscillating responses of surfaces when external forces are applied and generates sound from the approximate vibration. Figure 1. System overview. The green boxes indicate the parts users can interact with. The blue boxes indicate the pre-processing parts. The orange modules are the runtime components. 2.1. Interactive Synthesis for Complex Scenes Typically, the number of modes of an object with a few thousand vertices is in the range of a few thousands and the basic synthesis procedure runs in real-time [40]. But as the number of objects increases beyond two or three, the performance can degrade. In this case, our goal is to decrease the number of modes mixed and utilize the listener s perception from noticing the difference. Here we briefly discuss a few techniques to improve the performance and ensure that the method works well for interactive applications. For more details, we refer to [40]. 2.1.1. Mode Truncation The sound of a typical object on being struck consists of a transient response composed of a blend of high frequencies, followed by a set of lower frequencies with low amplitude. The transient attack is essential to the quality of sound as it is perceived as the characteristic timbre of the object. The idea behind mode truncation is to stop mixing a mode as soon as its contribution to the total sound falls below a certain preset threshold, τ. Since mode truncation preserves the initial transient response of the object when τ is suitably set, the resulting degradation in quality is minimal. 2.1.2. Mode Compression A perceptual study described in [45] showed that humans have a limited capacity to discriminate between frequencies which are close to each other. That is, if two close enough frequencies are played in succession, the average human listener is unable to tell whether they were two different frequencies or the same frequency played out twice. [40] lists the frequency discrimination at different frequencies. For example, at 2 KHz the frequency discrimination is more than 1 Hz. That is, a human subject cannot tell apart 1999 Hz from 2000 Hz. Note that the frequency discrimination deteriorates

a lot as the frequencies are increased to higher values. Therefore, while synthesizing any sound consisting of many frequencies we can easily cheat the listener s perception by replacing multiple frequencies close to each other by a single one, which represents all of them. This saves computation because mixing one frequency is much cheaper than mixing many. This is the main idea behind mode compression and leads to large gains in performance. 2.1.3. Quality Scaling Mode compression, as described above, is aimed at increasing the efficiency of sound synthesis for a single object. However, when the number of sounding objects in a scene grows beyond a few tens, increasing the efficiency of individual objects is not sufficient. Additionally, it is critical for games that the sound system has a graceful way of varying quality in response to variable time constraints. This flexibility is achieved by scaling the sound quality for the objects. The sound quality of an object is changed by controlling the number of modes being mixed for synthesizing its sound. The main idea is that in most cases of scenes with many sounding objects, the listener s attention is on the objects in the foreground, that is, the objects which contribute the most to the total sound in terms of amplitude. Therefore, if it is ensured that the foreground sounds are mixed at high quality while the background sounds are mixed at a relatively lower quality, the resulting degradation in perceived aural quality should be reduced. Quality scaling achieves this by assigning time quotas to all objects prioritized on the loudness of the sound they re producing and then scaling their quality to force them to complete within the assigned time quota. This simple technique performs quite well in practice. We have integrated our sound system with publicly available game engines, such as Open Source 3D Graphics Engine (ORGE) and Crystal Space, and physics engines, such as NVIDIA PhysX and Open Dynamics Engine (ODE). We are able to achieve interactive sound synthesis for complex scenes consisting of hundreds of objects at a steady frame rates of at least 100 frames per second (FPS) and upto 500-700 FPS, with the sound system taking approximately 10% of the CPU time [40]. 2.2. Liquid Sound Generation Most of the prior sound synthesis research has focused on simulating sounds from interaction between rigid and deformable bodies [16], [55], [35], [40], [3]. The limited previous work on simulating sounds in a liquid medium has focused either on heuristic methods decoupled from actual fluid dynamics [55] or methods that provide no physical basis for the sounds generated from the liquids [22]. Although these methods eliminate the need for recording sounds generated by actual liquids, which can be difficult and expensive, they still require a great deal of user s time and energy to tweak the various parameters by a plethora of sliders to generate desired sound effects and provide little ability to generate sounds from real-time fluid simulations. One of our goals is to generalize and expand the physicsbased sound synthesis techniques to all media, including liquids, and all type of interaction, including fluid-solid interaction, and thereby make the paradigm applicable to a much broader class of scenarios. Sound in liquid is primarily due to bubble formation. We refer the reader to Leighton s excellent text on acoustics due to bubble resonance for more detail [30]. Minnearts formula, which derives the resonant frequency of a perfectly spherical bubble in an infinite volume of water from the radius, provides the physical basis for generating sounds in liquid. Recently, Moss et al. [34] have developed the first automatic sound synthesis algorithm. It is based on the fact that the sound generated by a bubble is dominated by the resonant frequency, since all other frequencies will rapidly die out. The resonant frequency is dependent on the restoring force, which is the result of the pressure in the bubble. The resulting approach takes into account both the spherical and non-spherical bubbles. The resulting simulator is coupled with a fluid simulator that computes liquid dynamics, which subsequently continues to drive the bubble resonance in liquid media. There are many challenging computational issues in this coupling. The first issue is what type of fluid simulation should be used. Three broad categories exist for fluid dynamics computation in visual simulation: grid-based methods, smoothed particle hydrodynamics (SPH), and shallow-water approximations and for different applications, all these three type of simulators are used. In practice, our overall approach can generate plausible liquid sounds for different scenarios, as described in [34]. 3. Physics-Based Sound Propagation The most common approach to acoustic simulation is a two-stage process: the computation of impulse responses (IR) representing an acoustic space, and the convolution of the impulse responses with dry (anechoically recorded or synthetically generated) source signals. The IR computation relies on an accurate calculation for modeling the sound field. The sensation of sound is due to small variations in air pressure. Physics-based approaches solve the wave equation numerically to obtain the exact behavior of wave propagation in a domain. We assume that the input to our acoustics simulation is a 3D geometric model, along with the boundary conditions and the locations of the sound sources and listener. The propagation of sound in the domain is governed by the Acoustic Wave Equation: 2 p t 2 c2 2 p = F (x,t), (1) This equation captures the complete wave nature of sound, which is treated as a time-varying pressure field p (x,t) in space. The speed of sound is c = 340ms 1 and F (x,t) is the forcing term corresponding to sound sources present in the scene. The operator 2 = 2 x + 2

2 y + 2 2 z is the Laplacian in 3D. 2 3.1. Related Work Numerical methods for solving the wave equation rely on discretizing space and time for solution on a computer. Depending on the underlying discretization method, numerical approaches for acoustics may be roughly classified into: Finite Element Method (FEM), Boundary Element Method (BEM), Digital Waveguide Mesh (DWM), Finite Difference Time Domain (FDTD), and Functional Transform Method (FTM) [36], [38]. The FEM and BEM have traditionally been employed mainly for the steady-state frequency domain response, as opposed to a full time domain solution of the wave equation. FEM is applied mainly to the interior and BEM to exterior scattering problems [26]. DWM approaches [56], on the other hand, are specific to the Wave Equation and use discrete waveguide elements, each of which is assumed to carry waves along its length along a single dimension [24], [43]. However, such approaches suffer from directional dispersion of sound, that is, sound doesn t travel with the same speed in different directions on the spatial grid. The most commonly used method for time-domain wave simulations is the Finite Difference Time Domain (FDTD) method, owing to its simplicity and efficiency. FDTD has been an active area of research for more than a decade [4], [5] and was first proposed for electromagnetic simulations [46]. FDTD works on a uniform Cartesian grid and solves for the field values at each simulation cell over time. Initial investigations into FDTD were hampered by the lack of computational power and memory, limiting its application to mostly small models in 2D. Over the last decade, the possibility of applying FDTD to medium sized models in 3D has been explored [42]. However, the computational and memory requirements for FDTD are beyond the capability of most desktop systems today [41], requiring days of computation on a small cluster even for medium-sized 3D models for simulating frequencies below 1 khz. Other methods are based on spectral techniques [6], which are a class of very high order numerical schemes in which the complete field is expressed in terms of global basis functions, which virtually eliminates spatial approximation errors. 3.2. Adaptive Spatial Decomposition Physics-based approaches for sound propagation attempt to directly solve the Acoustic Wave Equation numerically, which governs all linear sound propagation phenomena, and are thus capable of performing a full transient solution which correctly accounts for all wave phenomena, including diffraction, elegantly in one framework. However, as mentioned above, computational costs for acoustic simulation can be prohibitively high. Recently Raghuvanshi et al. [39] have developed a novel approach based on adaptive rectangular decomposition of the free space in the domain to address this issue. Fig. 2 gives an overview of this approach. Many problems of interest for the purpose of acoustic simulation necessarily have large empty spaces in their Figure 2. A schematic overview of of the adaptive rectangular propagation algorithm of [Raghuvanshi et al. 2009]. The scene is first discretized with a Cartesian grid at a prescribed cell size depending on the highest simulated frequency. An adaptive rectangular decomposition is then performed on the scene and the partitions distributed between different cores/processors. At each time-step, the field in each partition is independently updated. interior. Consider a simulation on a large scene like an auditorium with an impulse triggered near the floor. With FDTD, this impulse would travel upwards and would accumulate numerical dispersion error at each spatial cell it crosses. Typically, an impulse would cross thousands of cells just to travel from one end of the small scene to the other, accumulating a lot of error in the process. However, if we fit a rectangle in the scene extending from the bottom to the top, the impulse would have no propagation error, since we can utilize the analytical solution within a rectangle. This observation is the main motivation for computing a rectangular decomposition. In order to achieve high spectral accuracy, both in time and space, in the interior of the rectangular partitions, Raghuvanshi et al. [39] use sixth-order spatial accuracy on the interface between partitions. Fig. 3 shows the results of our acoustic simulations after 5, 20, 50, 100 msec respectively for sound propagation within a cathedral scene. From the perspective of parallelizing numerical solvers, Domain Decomposition Methods (DDM) have been widely studied for scientific computation [14], [37], [51]. The rectangular partitioning also provides a natural domain decomposition and therefore, is a good candidate for parallelization on GPU clusters or many-core processors. We plan to further investigate the parallelization of numerical acoustic simulations using commodity hardware. 4. Geometric Sound Propagation The exact solution of sound propagation reduces to solving the wave equation, as described in the previous discussion. However, numerical methods to solve wave equation are mainly limited to static scenes. Given that the underlying complexity of numerical methods increases as fourth power of sound frequency, they are mostly useful for low or medium frequency

Figure 4. Beam vs. frustum tracing: Our rayfrustum tracing approach compared to beam tracing for a simple example.(left): beam tracing. (Right): ray-frustum tracing. The discrete sampling in our frustum based approach underestimates the size of the exact reflection and transmission frustum for primitive 1 and overestimates the size for primitives 2 and 3. Figure 3. Sound simulation on a Cathedral. The dimensions of this scene are 35m 15m 26m. We are able to perform numerical sound simulation on this complex scene on a desktop computer and pre-compute a 1 second long impulse response in about 29 minutes, taking less than 1 GB of memory. A commonly used approach that we compare against, Finite Difference Time Domain (FDTD), would take 1 week of computation and 25GB of memory for this scene to achieve competitive accuracy. The auralization, or sound rendering at runtime consists of convolution of the calculated impulse responses with arbitrary source signals, that can be computed efficiently. sound sources. The alternate methods for handling highfrequency sound sources are based on sound field decomposition or geometric propagation [19], [47], [48]. These algorithms model the propagation of sound based on rectilinear propagation of waves and can accurately model the early reflections (up to 4 6). These equivalent sources conceptually radiate in free space and are, in principle, supposed to fulfill the boundary condition of the acoustic space. The various equivalent sources are used to compute a list of elementary waves that arrive at the receiver. This includes the source data, propagation data that consists of frequency dependent attenuation due to distance spreading, accumulated absorption loss at wall reflections, accumulated scattering loss at wall reflections, and air absorption. 4.1. Prior Work Some of the earliest methods for geometric sound propagation were based on tracing sampled-rays [27] or sound-particles (phonons) [23], [2], [15] from a source to the listener. However, these methods are susceptible to aliasing problems. As a result, accurate methods for geometric sound propagation are either based on image source methods or volumetric propagation. Image source methods are the easiest and most popular for computing specular reflections [1], [13]. However, they can only handle simple static scenes or very low order of reflections at interactive rates [28]. Many hybrid combinations [12] of ray-based and image source methods have been proposed and used in commercial room acoustics prediction software (e.g. ODEON). But they are limited to rather simple static scenes and are not fast enough for interactive applications. The volumetric geometric methods trace pyramidal or volumetric beams to compute an accurate geometric solution. These include beam tracing that has been used for specular reflection and edge diffraction for interactive sound propagation [20], [28], [52]. The results of beam tracing can be used to guide sampling for path tracing [18]. However, they are only limited to very simple CAD models and it is rather difficult to make them work on complex CAD models due to robustness problems. 4.2. Ray-Frustum Tracing Recently, we have introduced a new volumetric approach, called ray-frustum tracing [29], [10]. This formulation uses a volumetric frustum tracing and utilizes the recent developments in interactive ray tracing to perform fast volumetric tracing using commodity processors. The overall approach is built on fast raytracing algorithms and represent each frustum using four corner rays. Unlike, prior beam tracing algorithm performs discrete clipping at the intersections, as shown in Fig. 4. The resulting approach can easily exploit the current hardware features in terms of data-parallelism and thread-level parallelism and thereby makes it possible to perform specular reflections and direct contributions at interactive rates. Moreover, it uses bounding volume hierarchies to accelerate the intersection tests and can also handle dynamic scenes. Recently, Chandak et al. [9] have presented a practical and conservative

visibility culling algorithm and used that for exact sound propagation. It is almost 10-20 times faster than prior beam-tracing algorithms and can accurately compute the propagation paths from the source to the listener. The initial frustum-tracing algorithm been primarily used for handling specular reflections in complex, dynamic scenes. Recently, we combined the frustumtracing formulation with the UTD formulation to perform edge-diffraction in complex virtual environments [50]. This approach is primarily designed for scenes with very large (or infinite-sized) edges, e.g. outdoor scenes with buildings. In terms of future work, it may be useful to combine this approach with the Biot-Tolstoy-Medwin (BTM) method [7], [25]. The BTM method is more accurate than UTD and can be formulated for use with finite edge [49]. However, it involves solving the BTM line integral over the edge length exposed to the sound field. As a result, BTM-based methods has a higher computational cost and recent research has focused on noninteractive acceleration techniques [32] used in offline simulations [33], [44] and edge subdivision techniques for interactive simulation of relative simple scenes [49], [8], [7]. Recen 4.3. Acoustic Levels-of-detail The 3D models used in most desktop and CAD applications are highly detailed. Such detail is often needed for visually realistic rendering of the scene. Acoustic simulation, on the contrary, requires much simpler 3D models with lower level of detail. A detailed 3D model may increase the computation cost of the simulation without any perceptual differences in sound rendering [53]. In order to accelerate the computations, a separate acoustic geometric representation may be needed and used for interactive frustum tracing. 4.4. Audio Rendering and 3D Audio Audio rendering is the process of producing the final output audio signal and is an important building block of interactive sound rendering. In order to render audio from geometric sound propagation we first construct an impulse response (IR) for each source listener pair from the sound paths that reach the listener from the source. A source listener pair s IR is then convolved with the input audio signal of the source to produce the final output audio signal. In terms of extending this pipeline to complex, dynamic scenes, there are two main challenges; firstly, the IRs are constantly changing due to the dynamic nature of the application. Secondly, the incoming path to the listener has a direction relative to the listener and a path length associated with it. Therefore, we treat it as a 3D sound source for realistic audio rendering. The computation of 3D sound involves convolution of incoming sound with the head related impulse response (HRIR) for every path. This can become computationally expensive and can be challenging when sound propagation is performed for a large number of sound sources. Some exciting work has recently been done to perform 3D audio rendering of large number of sound sources using sampling-based techniques [57]. 5. Conclusions In this paper, we gave a brief overview of our recent work on sound synthesis and propagation. This includes fast algorithms to generate sound from rigid bodies and liquid sounds. Moreover, we also gave an overview of new algorithms for numeric and geometric propagation, which are considerably faster than prior approaches. We also highlighted a few areas for future research. Acknowledgements: The work described in this paper represents joint work of the authors with Anish Chandak, Christian Lauterbach, William Moss, Nikunj Raghuvanshi, Zhimin Ren, Micah Taylor and Hengchin Yeh at UNC Chapel Hill. This research was supported in part by ARO Contract W911NF-04-1-0088, NSF awards 0429583 and 0636208, DARPA/RDECOM Contracts N61340-015-C-0043 and WR91CRB-08-C-0137, Intel, and Microsoft. References [1] Jont B. Allen and David A. Berkley. Image method for efficiently simulating small-room acoustics. The Journal of the Acoustical Society of America, 65(4):943 950, April 1979. [2] M. Bertram, E. Deines, J. Mohring, J. Jegorovs, and H. Hagen. Phonon tracing for auralization and visualization of sound. In Proceedings of IEEE Visualization, pages 151 158, 2005. [3] Nicolas Bonneel, George Drettakis, Nicolas Tsingos, Isabelle V. Delmon, and Doug James. Fast modal sounds with scalable frequency-domain synthesis, August 2008. [4] D. Botteldooren. Acoustical finite-difference timedomain simulation in a quasi-cartesian grid. The Journal of the Acoustical Society of America, 95(5):2313 2319, 1994. [5] D. Botteldooren. Finite-difference time-domain simulation of low-frequency room acoustic problems. Acoustical Society of America Journal, 98:3302 3308, December 1995. [6] John P. Boyd. Chebyshev and Fourier Spectral Methods : Second Revised Edition. Dover Publications, December 2001. [7] P. Calamia and U. P. Svensson. Fast time-domain edgediffraction calculations for interactive acoustic simulations. EURASIP Journal on Advances in Signal Processing, 2007. [8] P. T. Calamia and U. P. Svensson. Edge subdivision for fast diffraction calculations. In IEEE Trans. Vehicular Technology, pages 931 946, 1999. [9] A. Chandak, L. Antani, M. Taylor, and D. Manocha. Fastv: From-point visibility culling on complex models. Eurographics Symposium on Rendering, 2009. [10] Anish Chandak, Christian Lauterbach, Micah Taylor, Zhimin Ren, and Dinesh Manocha. Ad-frustum: Adaptive frustum tracing for interactive sound propagation. In Proc. IEEE Visualization, 2008. [11] M. J. Crocker. Handbook of Acoustics. Wiley-IEEE, 1998.

Figure 5. The left image shows the frusta generated by our adaptive volume tracing algorithm inside the 2.5 million triangle model of Soda Hall. The simulation is run with up to four specular reflections, edge diffraction and transmission. The right hand images show the new frusta generated on the fly, as we vary the dimensions of the room. The AD- FRUSTA algorithm can compute the propagation paths at five frames per second in this dynamic scene by using seven cores on a high PC. [12] B. Dalenbäck. Room acoustic prediction based on a unified treatment of diffuse and specular reflection. The Journal of the Acoustical Society of America, 100(2):899 909, 1996. [13] Bengt-Inge Dalenbäck and M. Strömberg. Real time walkthrough auralization - the first year. Proceedings of the Institute of Acoustics, 28(2), 2006. [14] Carlos A. de Moura. Parallel numerical methods for differential equations - a survey. [15] E. Deines, M. Bertram, J. Mohring, J. Jegorovs, F. Michel, H. Hagen, and G.M. Nielson. Comparative visualization for wave-based and geometric acoustics. IEEE Transactions on Visualization and Computer Graphics, 12(5), 2006. [16] Yoshinori Dobashi, Tsuyoshi Yamamoto, and Tomoyuki Nishita. Real-time rendering of aerodynamic sound using sound textures based on computational fluid dynamics. ACM Trans. Graph., 22(3):732 740, July 2003. [17] N.I. Durlach and A.S. Mavor. Virtual Reality Scientific and Technological Challenges. National Academy Press, 1995. [18] T. Funkhouser, N. Tsingos, I. Carlbom, G. Elko, M. Sondhi, and J. West. Modeling sound reflection and diffraction in architectural environments with beam tracing. In Forum Acusticum, September 2002. [19] T. Funkhouser, N. Tsingos, and Jean-Marc Jot. Survey of Methods for Modeling Sound Propagation in Interactive Virtual Environment Systems. Presence and Teleoperation, 2003. [20] Thomas Funkhouser, Ingrid Carlbom, Gary Elko, Gopal Pingali, Mohan Sondhi, and Jim West. A beam tracing approach to acoustic modeling for interactive virtual environments. In Proc. of ACM SIGGRAPH, pages 21 32, 1998. [21] K. Genuit and W. Bray. A virtual car: Prediction of sound and vibration in an interactive simulation environment. SAE Transactions, 110:1748 1754, 2001. [22] Masataka Imura, Yoshinobu Nakano, Yoshihiro Yasumuro, Yoshitsugu Manabe, and Kunihiro Chihara. Real-time generation of CG and sound of liquid with bubble. In ACM SIGGRAPH posters, page 97, San Diego, California, 2007. [23] B. Kapralos, M. Jenkin, and E. Milios. Acoustic Modeling Utilizing an Acoustic Version of Phonon Mapping. In Proc. of IEEE Workshop on HAVE, 2004. [24] Matti Karjalainen and Cumhur Erkut. Digital waveguides versus finite difference structures: equivalence and mixed modeling. EURASIP J. Appl. Signal Process., 2004(1):978 989, January 2004. [25] Joseph B. Keller. Geometrical theory of diffraction. Journal of the Optical Society of America, 52(2):116 130, 1962. [26] Mendel Kleiner, Bengt-Inge Dalenbäck, and Peter Svensson. Auralization - an overview. JAES, 41:861 875, 1993. [27] A. Krokstad, S. Strom, and S. Sorsdal. Calculating the acoustical room response by the use of a ray tracing technique. Journal of Sound and Vibration, 8(1):118 125, July 1968. [28] S. Laine, S. Siltanen, T. Lokki, and L. Savioja. Accelerated beam tracing algorithm. Applied Acoustic, 70(1):172 181, 2009.

[29] Christian Lauterbach, Anish Chandak, and Dinesh Manocha. Interactive sound rendering in complex and dynamic scenes using frustum tracing. IEEE Transactions on Visualization and Computer Graphics, 13(6):1672 1679, 2007. [30] T. G. Leighton. The acoustic bubble. Academic Press, 1994. [31] R. Bowen Loftin. Multisensory perception: Beyond the visual in visualization. Computing in Science and Engineering, 05(4):56 58, 2003. [32] T. Lokki, L. Savioja, R. Vaananen, J. Huopaniemi, and T. Takala. Creating interactive virtual auditory environments. IEEE Computer Graphics and Applications, 22(4):49 57, 2002. [33] T. Lokki, L. Savioja, R. Väänänen, J. Huopaniemi, and T. Takala. Creating interactive virtual auditory environments. IEEE Comput. Graph. Appl., 22(4):49 57, 2002. [34] W. Moss, H. Yeh, J. Hong, M. Lin, and D. Manocha. Sounding liquids: Automatic sound synthesis from fluid simulations. Technical report, University of North Carolina-Chapel Hill, 2009. [35] James F. O Brien, Chen Shen, and Christine M. Gatchalian. Synthesizing sounds from rigid-body simulations. In SCA 02: Proceedings of the 2002 ACM SIGGRAPH/Eurographics symposium on Computer animation, pages 175 181, New York, NY, USA, 2002. ACM. [36] R. Petrausch and S. Rabenstein. Simulation of room acoustics via block-based physical modeling with the functional transformation method. Applications of Signal Processing to Audio and Acoustics, 2005. IEEE Workshop on, pages 195 198, 16-19 Oct. 2005. [37] Alfio Quarteroni and Alberto Valli. Domain Decomposition Methods for Partial Differential Equations. Oxford University Press, 1999. [38] R. Rabenstein, S. Petrausch, A. Sarti, G. De Sanctis, C. Erkut, and M. Karjalainen. Block-based physical modeling for digital sound synthesis. Signal Processing Magazine, IEEE, 24(2):42 54, 2007. [39] N. Raghuvanshi, R. Narain, and M. C. Lin. Efficient and accurate sound propagation using adaptive rectangular propagation. IEEE Transactions on Visualization and Computer Graphics, 2009. [40] Nikunj Raghuvanshi and Ming C. Lin. Interactive sound synthesis for large scale environments. In SI3D 06: Proceedings of the 2006 symposium on Interactive 3D graphics and games, pages 101 108, New York, NY, USA, 2006. ACM Press. [41] S. Sakamoto, T. Yokota, and H. Tachibana. Numerical sound field analysis in halls using the finite difference time domain method. In RADS 2004, Awaji, Japan, 2004. [42] Shinichi Sakamoto, Takuma Seimiya, and Hideki Tachibana. Visualization of sound reflection and diffraction using finite difference time domain method. Acoustical Science and Technology, 23(1):34 39, 2002. [43] L. Savioja. Modeling Techniques for Virtual Acoustics. Doctoral thesis, Helsinki University of Technology, Telecommunications Software and Multimedia Laboratory, Report TML-A3, 1999. [44] L. Savioja, T. Lokki, and J. Huopaniemi. Auralization applying the parametric room acoustic modeling technique - the diva auralization system. 2002. [45] A. Sek and B. C. Moore. Frequency discrimination as a function of frequency, measured in several ways. J. Acoust. Soc. Am., 97(4):2479 2486, April 1995. [46] K. L. Shlager and J. B. Schneider. A selective survey of the finite-difference time-domain literature. Antennas and Propagation Magazine, IEEE, 37(4):39 57, 1995. [47] Samuel Siltanen, Tapio Lokki, Sami Kiminki, and Lauri Savioja. The room acoustic rendering equation. The Journal of the Acoustical Society of America, 122(3):1624 1635, September 2007. [48] P. Svensson and R. Kristiansen. Computational modelling and simulation of acoustic spaces. In 22nd International Conference: Virtual, Synthetic, and Entertainment Audio, June 2002. [49] U. P. Svensson, R. I. Fred, and J. Vanderkooy. An analytic secondary source model of edge diffraction impulse responses. Acoustical Society of America Journal, 106:2331 2344, November 1999. [50] Micah Taylor, Anish Chandak, Zhimin Ren, Christian Lauterbach, and Dinesh Manocha. Fast Edge-Diffraction for Sound Propagation in Complex Virtual Environments. In EAA Auralization Symposium, Espoo, Finland, June 2009. [51] Andrea Toselli and Olof Widlund. Domain Decomposition Methods. Springer, 1 edition, November 2004. [52] N. Tsingos, T. Funkhouser, A. Ngan, and I. Carlbom. Modeling acoustics in virtual environments using the uniform theory of diffraction. In Proc. of ACM SIGGRAPH, pages 545 552, 2001. [53] Nicolas Tsingos, Carsten Dachsbacher, Sylvain Lefebvre, and Matteo Dellepiane. Extending Geometrical Acoustics to Highly Detailed Architectural Environments. In 19th Intl. Congress on Acoustics, sep 2007. [54] J. Tsou, B. Chow, and S. Lam. Performance-based simulation for the planning and design of hyper-dense urban habitation. Computer Aided Architectural Design Research, 12(5):521 526, 2003. [55] Kees van den Doel, Paul G. Kry, and Dinesh K. Pai. Foleyautomatic: physically-based sound effects for interactive simulation and animation. In SIGGRAPH 01: Proceedings of the 28th annual conference on Computer graphics and interactive techniques, pages 537 544, New York, NY, USA, 2001. ACM Press. [56] S. Van Duyne and J. O. Smith. The 2-d digital waveguide mesh. In Applications of Signal Processing to Audio and Acoustics, 1993. Final Program and Paper Summaries., 1993 IEEE Workshop on, pages 177 180, 1993. [57] Michael Wand and Wolfgang Straßer. Multi-resolution sound rendering. In SIGGRAPH 04: ACM SIGGRAPH 2004 Sketches, page 47, 2004.