Meta wants the virtual landscape to sound like real life


Meta and a group of researchers from the University of Texas at Austin (UT Austin) are working on bringing realistic audio to the metaverse.
As Kristen Garuman, Research Director at Meta AI, explains (opens in new tab), there’s more to augmented and virtual reality (AR and VR, respectively) than just visuals. Audio plays a very important role in making a world feel alive. Garuman says “audio is shaped by the environment that [it’s] in.” There are various factors that influence how sound behaves like the geometry of a room, what’s in said room, and how far someone is from a source.
To achieve this, Meta’s plan is to use AR glasses to record both audio and video from one location, then using a set of three AI models, transform and clean the recording so it feels like it’s happening in front of you when you play it back at home. The AIs will take into account the room that you’re in so it can match the environment.
Looking at the projects, it appears Meta is focusing on AR glasses. Meta’s plan for VR headsets includes replicating the sights and sounds of an environment, like a concert, so it feels like you’re there in person.
We asked Meta how can people listen to the enhanced audio. Will people need a pair of headphones to listen or will it come from the headset? We didn’t get a response.
We also asked Meta how can developers get a hold of these AI models. They’ve been made open source so third-party developers can work on the tech, but Meta didn’t offer any further details.
Transformed by AI
The question is how can Meta record audio on a pair of AR glasses and have it reflect a new setting.
The first solution is known as AViTAR which is a ”Visual Acoustic Matching model.” (opens in new tab) This is the AI that transforms audio to match a new environment. Meta offers the example of a mother recording her child’s dance recital at an auditorium with a pair of AR glasses.
One of the researchers claims that the mother in question can take that recording and play it back at home where the AI will morph the audio. It’ll scan the environment, take into account any obstacles in a room, and have the recital sound like it’s happening right in front of her with the same glasses. The researcher states the audio will come from the glasses.
To help clean up audio, there is Visually-Informed Dereverberation (opens in new tab). Basically, it removes distracting reverb from the clip. The example given is recording a violin concert at a train station, taking it home, and having the AI clean up the clip so you hear nothing but music.
The last AI model is VisualVoice (opens in new tab), which uses a combination of visual and audio cues to separate voices from other noises. Imagine recording a video of two people arguing. This AI will isolate one voice so you can understand them while silencing everything else. Meta explains visual cues are important because AI needs to see who’s talking in order to understand certain nuances and know who’s speaking.
In relation to visuals, Meta states they plan on bringing in video and other cues to further enhance AI-driven audio. Since this technology is still early in development, it’s unknown if and when Meta will bring these AIs to a Quest headset near you.
Be sure to read our latest review on the Oculus Quest 2 if you’re thinking of buying one. Spoiler alert: we like it.
Audio player loading… Meta and a group of researchers from the University of Texas at Austin (UT Austin) are working on bringing realistic audio to the metaverse. As Kristen Garuman, Research Director at Meta AI, explains (opens in new tab), there’s more to augmented and virtual reality (AR and VR,…
Recent Posts
- Elon Musk’s AI said he and Trump deserve the death penalty
- The GSA is shutting down its EV chargers, calling them ‘not mission critical’
- Lenovo is going all out with yet another funky laptop design: this time, it’s a business notebook with a foldable OLED screen
- Elon Musk’s first month of destroying America will cost us decades
- The first iOS 18.4 developer beta is here, with support for Priority Notifications
Archives
- February 2025
- January 2025
- December 2024
- November 2024
- October 2024
- September 2024
- August 2024
- July 2024
- June 2024
- May 2024
- April 2024
- March 2024
- February 2024
- January 2024
- December 2023
- November 2023
- October 2023
- September 2023
- August 2023
- July 2023
- June 2023
- May 2023
- April 2023
- March 2023
- February 2023
- January 2023
- December 2022
- November 2022
- October 2022
- September 2022
- August 2022
- July 2022
- June 2022
- May 2022
- April 2022
- March 2022
- February 2022
- January 2022
- December 2021
- November 2021
- October 2021
- September 2021
- August 2021
- July 2021
- June 2021
- May 2021
- April 2021
- March 2021
- February 2021
- January 2021
- December 2020
- November 2020
- October 2020
- September 2020
- August 2020
- July 2020
- June 2020
- May 2020
- April 2020
- March 2020
- February 2020
- January 2020
- December 2019
- November 2019
- September 2018
- October 2017
- December 2011
- August 2010