Meta’s new AI models can recognize and produce speech for more than 1,000 languages

10 months ago 141

Meta has built AI models that tin admit and nutrient code for much than 1,000 languages—a tenfold summation connected what’s presently available. It’s a important measurement toward preserving languages that are astatine hazard of disappearing, the institution says.

Meta is releasing its models to the nationalist via the codification hosting work GitHub. It claims that making them unfastened source volition assistance developers moving successful antithetic languages to physique caller code applications—like messaging services that recognize everyone, oregon virtual-reality systems that tin beryllium utilized successful immoderate language.

There are astir 7,000 languages successful the world, but existing code designation models screen lone astir 100 of them comprehensively. This is due to the fact that these kinds of models thin to necessitate immense amounts of labeled grooming data, which is disposable for lone a tiny fig of languages, including English, Spanish, and Chinese.

Meta researchers got astir this occupation by retraining an existing AI exemplary developed by the institution successful 2020 that is capable to larn code patterns from audio without requiring ample amounts of labeled data, specified arsenic transcripts. 

They trained it connected 2 caller information sets: 1 that contains audio recordings of the New Testament Bible and its corresponding substance taken from the net successful 1,107 languages, and different containing unlabeled New Testament audio recordings successful 3,809 languages. The squad processed the code audio and the substance information to amended its prime earlier moving an algorithm designed to align audio recordings with accompanying text. They past repeated this process with a 2nd algorithm trained connected the recently aligned data. With this method, the researchers were capable to thatch the algorithm to larn a caller connection much easily, adjacent without the accompanying text.

“We tin usage what that exemplary learned to past rapidly physique code systems with very, precise small data,” says Michael Auli, a probe idiosyncratic astatine Meta who worked connected the project.

“For English, we person tons and tons of bully information sets, and we person that for a fewer much languages, but we conscionable don’t person that for languages that are spoken by, say, 1,000 people.” 

The researchers accidental their models tin converse successful implicit 1,000 languages but admit much than 4,000. 

They compared the models with those from rival companies, including OpenAI Whisper, and assertion theirs had fractional the mistake rate, contempt covering 11 times much languages.

However, the squad warns the exemplary is inactive astatine hazard of mistranscribing definite words oregon phrases, which could effect successful inaccurate oregon perchance violative labels. They besides admit that their code designation models yielded much biased words than different models, albeit lone 0.7% more. 

While the scope of the probe is impressive, the usage of spiritual texts to bid AI models tin beryllium controversial, says Chris Emezue, a researcher astatine Masakhane, an enactment moving connected natural-language processing for African languages, who was not progressive successful the project.

“The Bible has a batch of bias and misrepresentations,” helium says.

Read Entire Article