\"\"
<\/span><\/figcaption><\/figure>New Delhi: Meta<\/a> (formerly Facebook<\/a>) has built three new artificial intelligence (AI<\/a>) models designed to make sound more realistic in mixed and virtual reality experiences.

The three AI models --
Visual-Acoustic Matching<\/a>, Visually-Informed Dereverberation<\/a> and VisualVoice<\/a> -- focus on human speech and sounds in video and are designed to push \"us toward a more immersive reality at a faster rate,\" the company said in a statement.

\"Acoustics play a role in how sound will be experienced in the metaverse, and we believe AI will be core to delivering realistic sound quality,\" said Meta's AI researchers and audio specialists from its Reality Labs team.

They built the AI models in collaboration with researchers from the University of Texas at Austin, and are making these models for audio-visual understanding open to developers.

The self-supervised Visual-Acoustic Matching model, called
AViTAR<\/a>, adjusts audio to match the space of a target image.

The self-supervised training objective learns acoustic matching from in-the-wild web videos, despite their lack of acoustically mismatched audio and unlabelled data, informed Meta.

VisualVoice learns in a way that's similar to how people master new skills, by learning visual and auditory cues from unlabelled videos to achieve audio-visual speech separation.

For example, imagine being able to attend a group meeting in the metaverse with colleagues from around the world, but instead of people having fewer conversations and talking over one another, the reverberation and acoustics would adjust accordingly as they moved around the virtual space and joined smaller groups.

\"VisualVoice generalises well to challenging real-world videos of diverse scenarios,\" said
Meta AI<\/a> researchers.
<\/body>","next_sibling":[{"msid":92485076,"title":"Facebook Watch video app no longer works on Apple TV","entity_type":"ARTICLE","link":"\/news\/facebook-watch-video-app-no-longer-works-on-apple-tv\/92485076","category_name":null,"category_name_seo":"telecomnews"}],"related_content":[],"msid":92485252,"entity_type":"ARTICLE","title":"Meta builds AI models that provide realistic sounds in VR settings","synopsis":"The three AI models -- Visual-Acoustic Matching, Visually-Informed Dereverberation and VisualVoice -- focus on human speech and sounds in video and are designed to push \"us toward a more immersive reality at a faster rate,\" the company said in a statement.","titleseo":"telecomnews\/meta-builds-ai-models-that-provide-realistic-sounds-in-vr-settings","status":"ACTIVE","authors":[],"Alttitle":{"minfo":""},"artag":"IANS","artdate":"2022-06-27 11:47:04","lastupd":"2022-06-27 11:55:21","breadcrumbTags":["meta","internet","facebook","VisualVoice","Meta AI","avitar","ai","Visually-Informed Dereverberation","Visual-Acoustic Matching","vr"],"secinfo":{"seolocation":"telecomnews\/meta-builds-ai-models-that-provide-realistic-sounds-in-vr-settings"}}" data-authors="[" "]" data-category-name="" data-category_id="" data-date="2022-06-27" data-index="article_1">

元构建人工智能模型,提供了现实的声音在虚拟现实环境中

三种人工智能模型——Visual-Acoustic匹配,Visually-Informed反混响和VisualVoice——关注人类语言和声音的视频,是为了推动“我们走向一个更身临其境的真实速度,”该公司在一份声明中说。

  • 更新在2022年6月27日上午11时55坚持
新德里消息:(原脸谱网)建立了三个新的人工智能(人工智能)模型旨在使声音更现实的混合和虚拟现实的体验。

——三个人工智能模型Visual-Acoustic匹配,Visually-Informed反混响VisualVoice——关注人类语言和声音的视频,是为了推动“我们走向一个更身临其境的真实速度,”该公司在一份声明中说。

“声学扮演一个角色在metaverse如何经验丰富的声音,我们相信人工智能将核心提供现实的声音质量,“元的人工智能研究人员和音频专家说其现实实验室团队。

广告
他们建造的人工智能模型与德州大学奥斯丁分校的研究人员合作,并使视听理解这些模型对开发者开放。

的self-supervised Visual-Acoustic匹配模型,称为AViTAR调整音频匹配的空间目标图像。

self-supervised培训目标学习声学匹配从野生的网络视频,尽管他们缺乏听觉上不匹配的音频和未标记的数据,通知元。

VisualVoice学习的方式类似于人们如何掌握新技能,通过学习达到视觉和听觉线索从无标号的视频视听语音分离。

例如,假设能够参加一个小组会议与来自世界各地的同事metaverse,而是人有更少的对话和讨论,混响和声学将作出相应调整移动虚拟空间时,加入了较小的团体。

“VisualVoice难题挑战真实世界的不同场景的视频,”说元的人工智能研究人员。
  • 发表在2022年6月27日上午11点坚持
是第一个发表评论。
现在评论

加入2 m +行业专业人士的社区

订阅我们的通讯最新见解与分析。乐动扑克

下载ETTelec乐动娱乐招聘om应用

  • 得到实时更新
  • 保存您最喜爱的文章
扫描下载应用程序
\"\"
<\/span><\/figcaption><\/figure>New Delhi: Meta<\/a> (formerly Facebook<\/a>) has built three new artificial intelligence (AI<\/a>) models designed to make sound more realistic in mixed and virtual reality experiences.

The three AI models --
Visual-Acoustic Matching<\/a>, Visually-Informed Dereverberation<\/a> and VisualVoice<\/a> -- focus on human speech and sounds in video and are designed to push \"us toward a more immersive reality at a faster rate,\" the company said in a statement.

\"Acoustics play a role in how sound will be experienced in the metaverse, and we believe AI will be core to delivering realistic sound quality,\" said Meta's AI researchers and audio specialists from its Reality Labs team.

They built the AI models in collaboration with researchers from the University of Texas at Austin, and are making these models for audio-visual understanding open to developers.

The self-supervised Visual-Acoustic Matching model, called
AViTAR<\/a>, adjusts audio to match the space of a target image.

The self-supervised training objective learns acoustic matching from in-the-wild web videos, despite their lack of acoustically mismatched audio and unlabelled data, informed Meta.

VisualVoice learns in a way that's similar to how people master new skills, by learning visual and auditory cues from unlabelled videos to achieve audio-visual speech separation.

For example, imagine being able to attend a group meeting in the metaverse with colleagues from around the world, but instead of people having fewer conversations and talking over one another, the reverberation and acoustics would adjust accordingly as they moved around the virtual space and joined smaller groups.

\"VisualVoice generalises well to challenging real-world videos of diverse scenarios,\" said
Meta AI<\/a> researchers.
<\/body>","next_sibling":[{"msid":92485076,"title":"Facebook Watch video app no longer works on Apple TV","entity_type":"ARTICLE","link":"\/news\/facebook-watch-video-app-no-longer-works-on-apple-tv\/92485076","category_name":null,"category_name_seo":"telecomnews"}],"related_content":[],"msid":92485252,"entity_type":"ARTICLE","title":"Meta builds AI models that provide realistic sounds in VR settings","synopsis":"The three AI models -- Visual-Acoustic Matching, Visually-Informed Dereverberation and VisualVoice -- focus on human speech and sounds in video and are designed to push \"us toward a more immersive reality at a faster rate,\" the company said in a statement.","titleseo":"telecomnews\/meta-builds-ai-models-that-provide-realistic-sounds-in-vr-settings","status":"ACTIVE","authors":[],"Alttitle":{"minfo":""},"artag":"IANS","artdate":"2022-06-27 11:47:04","lastupd":"2022-06-27 11:55:21","breadcrumbTags":["meta","internet","facebook","VisualVoice","Meta AI","avitar","ai","Visually-Informed Dereverberation","Visual-Acoustic Matching","vr"],"secinfo":{"seolocation":"telecomnews\/meta-builds-ai-models-that-provide-realistic-sounds-in-vr-settings"}}" data-news_link="//www.iser-br.com/news/meta-builds-ai-models-that-provide-realistic-sounds-in-vr-settings/92485252">