Meta to train speech recognition engines on 'clusters' of speakers using new dataset


San Francisco, Jul 14 (IANS): Meta (formerly Facebook) has developed a new dataset which the company will use to improve the performance of automatic speech recognition (ASR) tools by clustering speech at the "utterance level".

As part of Meta's continued commitment to improving ASR performance, the company has taught ASRs to train without transcripts, recognise over 4,000 spoken languages, and even read lips more accurately than humans.

However, many of the datasets used to train ASR models are organised by demographic such as age group, gender, nationality, and English accent, which limits the variation of pronunciations that models are trained on, ultimately hampering their function in understanding a wide range of users.

In order to overcome this, Meta AI developed a dataset that relies instead on utterance clustering.

"Instead of dividing a dataset based on speakers' demographic information -- such as their age group or gender -- our proposed algorithm clusters speech at the utterance level," Meta said in a blogpost on Thursday.

"A single cluster will contain similar utterances from a diverse group of speakers. We can then train our model using the various clusters and use fairness datasets to measure how the model impacts outcomes across different demographic groups," it added.

The company's resulting dataset includes about 27,055 utterances in a recorded speech by 595 people in the US who were paid to record and submit audio of themselves saying commands.

Their utterances are organised around seven main themes -- music, capture, utilities, notification control, messaging, calling, and dictation, which other researchers can use to train their own models and digital assistants on.

The speakers were asked how they would voice search for a song, make plans with friends, and decide where to meet up.

To evaluate this new system, Meta trained their model on de-identified, publicly available Facebook videos in English which were evaluated on two datasets.

The first was a de-identified dataset collected from a data supplier for ASR that includes 48,000 utterances from 867 speakers, and the second dataset is Casual Conversations v1, a dataset of transcribed speech that Meta built and made publicly available in 2021.

"During testing, we observed that a model trained in this manner improved speech recognition accuracy for all measured demographic groups, and in particular for different accents, which are identified in sociolinguistics as a way of pronouncing a language that is distinctive to a country, area, social class, or individual," Meta said.

"While our proposed algorithm was built using English-language data, we hope these approaches can be extended to work for other languages as well," it added.

 

  

Top Stories


Leave a Comment

Title: Meta to train speech recognition engines on 'clusters' of speakers using new dataset



You have 2000 characters left.

Disclaimer:

Please write your correct name and email address. Kindly do not post any personal, abusive, defamatory, infringing, obscene, indecent, discriminatory or unlawful or similar comments. Daijiworld.com will not be responsible for any defamatory message posted under this article.

Please note that sending false messages to insult, defame, intimidate, mislead or deceive people or to intentionally cause public disorder is punishable under law. It is obligatory on Daijiworld to provide the IP address and other details of senders of such comments, to the authority concerned upon request.

Hence, sending offensive comments using daijiworld will be purely at your own risk, and in no way will Daijiworld.com be held responsible.