HEARING IN TRUE 3-D
- Created: Thursday, 01 January 2004
In 1984, researchers from Ames Research Center came together to develop advanced human interfaces for NASA's teleoperations that would come to be known as virtual reality. The basis of the work theorized that if the sensory interfaces met a certain threshold and sufficiently supported each other, then the operator would feel present in the remote/synthetic environment, rather than present in their physical location. Twenty years later, this prolific research continues to pay dividends to society in the form of cutting-edge virtual reality products, such as an interactive audio simulation system.
Throughout the 1990s, virtual reality technology was applied to multiple areas, from video games to military equipment. William Chapin founded AuSIM, Inc., in 1998 to develop three-dimensional (3-D) audio products for mission-critical applications, such as those originally proposed by NASA. Prior to launching his Mountain View, California-based company, Chapin joined NASA partners and researchers to develop several iterations of increasingly more-detailed, physically-based acoustic room simulation models. Over a 4-year period, they would develop three successively more accurate models of acoustic simulation.
When AuSIM came to be in 1998, Chapin would further fortify his ties with NASA. Dr. Stephen Ellis, a member of Ames Human Information Processing Branch, was conducting research on perceptual issues relating to latency in visual displays, along with Dr. Dov Adelstein and Dr. Elizabeth Wenzel one of the NASA researchers who helped to develop the original virtual reality interfaces for Ames. AuSIM assisted Ellis, Adelstein, and Wenzel by integrating aural and visual displays so the three could study the inter-relationship of latency. Ames contracted with AuSIM to provide the synchronization control in the aural and visual display systems. This work would lead to a series of annually renewed contracts between Ames and AuSIM.
Meanwhile, across the country at NASA's Langley Research Center, Dr. Stephen Rizzi of the Structural Acoustics Branch needed an auralization architecture on which he could develop his own models. Rizzi and AuSIM collaborated to make a version of the company's technology in which sub-models could be replaced with a plug-in design. This open kernel architecture collaboration continues through 2004, with support from Phase I and Phase II Small Business Innovation Research (SBIR) contracts. Additionally, Rizzi and AuSIM produced joint research papers based on their studies of advanced propagation models and structural acoustics.
In noisy environments such as restaurants and lobbies, people are well-adapted to tuning into desired sound and tuning out noise, a perceptual phenomenon referred to as the cocktail effect. Humans perceive signatures in sound from the propagation and from the source to their ears, and hence create a mental image of the environment that allows them to discriminate independently originating sounds. AuSIM notes that traditional audio technologies do not simulate the propagation of sounds through a medium and therefore present false aural signatures.
AuSIM's solutions to help humans differentiate between simultaneous sounds are based on NASA-influenced audio simulation techniques that create and preserve the perceptual spatial clues in electronically transmitted sound. The solutions apply to military, industrial, voice telecommunications, and academic research projects.
As the company's core technology, AuSIM3D gathers dynamic acoustic properties, 3-D position, and 3-D orientation of all objects to drive complex models based on the physics of sound waves. Applied to real-world tasks, AuSIM3D reduces fatigue with naturally presented information, maintains more efficient and productive workers, increases accuracy and quality of listeners work, yields fewer critical and costly mistakes, and saves time, money, and even lives.
AuSIM has broadened its original mission-critical business plan to make room for human interest applications. On a general level, AuSIM's products can be utilized in teleconferences, where spatially consistent voices can sound more natural, and in driver's education schools, so that a student can learn to react to realistic sound events in a simulator, putting fewer people at risk during the learning and thereby creating a prepared driver for the real situation. For future applications, AuSIM has teamed with the Girvan Institute of Technology, with the intentions of developing and capitalizing on end-user products for more key markets. The institute selects the best and most promising small companies commercializing NASA technology to be incubated and capitalized.