“Augmented Audio Reality (AAR) ” has become another buzz word with the advent of Air Buds, apps like RjDj Here and innovations from audio companies like Doppler Labs and Harman as well as the popularity of applications of it’s big brother, “Augmented Reality” like Pokeman Go. Will consumers embrace AAR or will it be another personal intrusion only geeks will embrace?
I suspect it will be a mixed bag with broad adoption of AAR applications that enhance user experience and resistance to applications that intrude on the users brain. But considering AAR is a close cousin to Interactive Audio who better qualified than the BarBQ to sort the wheat from the chaff (You’all).
I pose these questions to the BarBQ Brain:
- What AAR applications will be embraced by consumers and why?
- What audio technology is needed to make advanced AAR compelling?
- Will all AAR solutions be proprietary or is there a need for any standards?
- What will AAR applications look like in 2021?
I’m sure other can add even more meaningful questions to this list.
What advances in audio can enhance the effectiveness of augmented reality solutions? Is state-of-the-art audio enough or are there sound barriers to be broken to make the augmented reality user experience more compelling? What AR applications will drive these needs?
While I considered inclusion of virtual reality I believe there are some many diverse applications for augmented reality it is a huge topic in itself.
Humans communicate with each other not just with pure speech but with speech augmented by a variety of cues including audio, facial, posture and gestures. Do any of these cues also add value in human to machine communications? Could they serve as another form of contextual awareness making ASR more accurate and machine responses more meaningful.
Some suggest these cues are too individual in nature to augment human to machine communication. Others point out that experts can identify and interpret these audio and visual cues by observing anyone which suggests computer intelligence could easily programmed to interpret these cues.
Let’s identify specific applications where a multifactor interface (speech plus other visual and audio cues) would add value in the new world where your voice becomes the primary interface to consumer products.
What wearable and IoT applications featuring audio will make a splash and stick? What applications will just splat? What sound voice, speech and audio technologies will make a difference in these emerging applications? What advances in sound technologies are needed in the next five years to create compelling and sustainable applications in this space? Let’s brainstorm potential applications with real consumer value propositions, debate their merit, prioritize them and define a sound technology roadmap to support them.