HomeNewsTechnologyMeta to train ASR models by clustering speech at an 'utterance level'

Meta to train ASR models by clustering speech at an 'utterance level'

The goal is to meld similar utterances from a diverse group of speakers together in one data set, and then use that to train the ASR model.

July 14, 2023 / 16:30 IST
Story continues below Advertisement
Meta to train ASR models by clustering speech at an 'utterance level'
The most popular examples of ASR's are smartphone assistants like Apple's Siri, Amazon Echo or Google Assistant.(Representative image)

Meta has developed a new way to train Automatic Speech Recognition (ASR) models by clustering speech at an "utterance level".

ASR models, as the name implies, are used in systems that aim to transcribe spoken language into text, that can be used to carry out various functions. The most popular examples of ASR's are smartphone assistants like Apple's Siri, Amazon Echo or Google Assistant.

Story continues below Advertisement

Also read | WhatsApp testing logins using phone numbers on WhatsApp Web

Despite the advancement of AI technology, you may find these assistants will sometimes have a hard time understanding you speech. Meta aims to improve this clustering various speakers from different ethnicities together, rather than traditional data sets that train ASR models based on metrics such as age group or gender.