首页  >>  来自播客: Meta AI 更新   反馈

Acoustic synthesis for AR/VR experiences

发布时间 2022-06-24 22:25:02    来源

摘要

Existing AI models do a good job of understanding images but require more work to understand the acoustics of environments in the related image. This is why researchers from Meta AI and the University of Texas are open-sourcing three new models for audio-visual understanding of human speech and sounds in video, helping us achieve immersive AR and VR experiences much more quickly. Using multimodal AI models that can take audio, video, and text signals at one time, AI will be able to deliver sound quality that realistically matches the settings people are immersed in. Learn more about this state-of-the-art work here (insert link).

GPT-4正在为你翻译摘要中......

中英文字稿