Monday, December 23, 2024
Advertisement
  1. You Are At:
  2. News
  3. Technology
  4. Facebook introduces first AI model to translate 100 languages without dependency on English

Facebook introduces first AI model to translate 100 languages without dependency on English

For years, AI researchers have been working toward building a single universal model that can understand all languages across different tasks.

Reported by: IANS New Delhi Published : Oct 20, 2020 15:03 IST, Updated : Oct 20, 2020 15:03 IST
facebook, artificial intelligence, AI, facebook multilingual Machine Translation model, facebook tra
Image Source : FACEBOOK

Facebook's AI translation model

Facebook on Monday introduced the first-ever, open-source multilingual machine translation (MMT) model that can translate between any pair of 100 languages without relying on English data. Called "M2M-100," it is trained on a total of 2,200 language directions — or 10 times more than previous best, English-centric multilingual models.

"Deploying M2M-100 will improve the quality of translations for billions of people, especially those that speak low-resource languages," Facebook AI said in a statement.

When translating, say, Chinese to French, most English-centric multilingual models train on Chinese to English and English to French, because English training data is the most widely available.

The new Facebook ML model directly trains on Chinese to French data to better preserve meaning. It outperforms English-centric systems by 10 points on the widely used BLEU metric for evaluating machine translations.

"We're also releasing the model, training, and evaluation setup to help other researchers reproduce and further advance multilingual models," the social network announced.

Using novel mining strategies to create translation data, Facebook built the first truly "many-to-many" data set with 7.5 billion sentences for 100 languages.

"We used several scaling techniques to build a universal model with 15 billion parameters, which captures information from related languages and reflects a more diverse script of languages and morphology," the company said.

One challenge in multilingual translation is that a singular model must capture information in many different languages and diverse scripts. To address this, Facebook saw a clear benefit of scaling the capacity of its model and adding language-specific parameters.

"The combination of dense scaling and language-specific sparse parameters (3.2 billion) enabled us to create an even better model, with 15 billion parameters".

For years, AI researchers have been working toward building a single universal model that can understand all languages across different tasks.

"A single model that supports all languages, dialects, and modalities will help us better serve more people, keep translations up to date, and create new experiences for billions of people equally," Facebook said.

Latest technology reviews, news and more

Advertisement

Read all the Breaking News Live on indiatvnews.com and Get Latest English News & Updates from Technology

Advertisement
Advertisement
Advertisement
Advertisement