Home > Technology peripherals > AI > Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

WBOY
Release: 2023-04-25 12:04:06
forward
1489 people have browsed it

Last week, OpenAI released the ChatGPT API and Whisper API, which just triggered a carnival among developers.

#On March 6, Google launched a benchmark model-USM. Not only can it support more than 100 languages, but the number of parameters has also reached 2 billion.

# Of course, the model is still not open to the public, "This is very Google"!

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

## Simply put, the USM model covers 12 million hours of speech and 28 billion sentences. It is pre-trained on an unlabeled dataset of 300 different languages ​​and fine-tuned on a smaller labeled training set.

Google researchers said that although the annotation training set used for fine-tuning is only 1/7 of Whisper, USM has equivalent or even better results. performance, and also the ability to efficiently adapt to new languages ​​and data.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

##Paper address: https://arxiv.org/abs/2303.01037

The results show that USM not only achieves SOTA in multilingual automatic speech recognition and speech-text translation task evaluation, but can also be actually used in YouTube subtitle generation.

#Currently, the languages ​​that support automatic detection and translation include mainstream English, Chinese, and small languages ​​such as Assamese.

#The most important thing is that it can also be used for real-time translation of future AR glasses demonstrated by Google at last year’s IO conference.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages##Jeff Dean personally announced: Let AI support 1,000 languages

While Microsoft and Google argue over who has the better AI chatbot, know that large language models can be used for more than just that.

In November last year, Google first announced a new project to "develop an artificial intelligence language model that supports the 1,000 most commonly used languages ​​in the world."

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

In the same year, Meta also released a model called “No Language Left Behind”. It also claims that it can translate more than 200 languages ​​and aims to create a "universal translator".

#The release of the latest model is described by Google as a "critical step" towards its goal.

# When it comes to building language models, it can be said that there are many heroes competing.

According to rumors, Google plans to showcase more than 20 products powered by artificial intelligence at this year’s annual I/O conference.

Currently, automatic speech recognition faces many challenges:

    Traditional Supervised learning methods lack scalability
#In traditional methods, audio data needs to be manually labeled in a time-consuming and expensive manner, or it is not pre-existing Collected from sources of transcriptions, which can be difficult to find for languages ​​that lack broad representation.

  • While expanding language coverage and quality, models must be improved in computationally efficient ways

This requires algorithms that can use large amounts of data from different sources, enable model updates without the need for complete retraining, and be able to generalize to new languages ​​and use cases .

Fine-tuned self-supervised learning

According to the paper, USM training uses three databases: unpaired audio data set, unpaired Text data set, paired ASR corpus.

  • Unpaired Audio Dataset

Includes YT-NTL-U (over 12 million hours of YouTube untagged audio data) and Pub-U (over 429,000 hours of speech content in 51 languages)

  • Unpaired text data set

Web-NTL (28 billion sentences in over 1140 different languages)

  • Paired ASR corpus

YT-SUP and Pub-S corpora (over 10,000 hours of audio content and matching text)

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

USM uses a standard encoder-decoder structure, where the decoder can be CTC, RNN -T or LAS.

#For the encoder, USM uses Conformor, or convolution enhanced Transformer.

#The training process is divided into three stages.

#In the initial stage, unsupervised pre-training is performed using BEST-RQ (BERT-based random projection quantizer for speech pre-training). The goal is to optimize RQ.

#In the next stage, the speech representation learning model is further trained.

Use MOST (Multi-Object Supervised Pre-training) to integrate information from other text data.

The model introduces an additional encoder module that takes text as input and introduces additional layers to combine the speech encoder and text encoder output, and jointly train the model on unlabeled speech, labeled speech, and text data.

The last step is to fine-tune the ASR (automatic speech recognition) and AST (automatic speech translation) tasks. The pre-trained USM model only requires a small amount of Supervisory data can achieve good performance.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

##USM overall training process

How does USM perform? Google tested it on YouTube subtitles, promotion of downstream ASR tasks, and automatic speech translation.

Performance on YouTube multi-language subtitles

Supervised YouTube data includes 73 languages, with an average of just under 3,000 hours of data per language. Despite limited supervision data, the model achieved an average word error rate (WER) of less than 30% across 73 languages, which is lower than state-of-the-art models within the United States.

In addition, Google compared it with the Whisper model (big-v2) trained with more than 400,000 hours of annotated data.

Among the 18 languages ​​that Whisper can decode, its decoding error rate is less than 40%, while the average USM error rate is only 32.7%.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

##Promotion of downstream ASR tasks

On publicly available datasets, USM shows lower performance on CORAAL (African American Dialect English), SpeechStew (English - United States), and FLEURS (102 languages) compared to Whisper WER, regardless of whether there is in-domain training data.

#The difference between the two models in FLEURS is particularly obvious.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

##Performance on the AST task

Fine-tuning USM on the CoVoST dataset.

Divide the languages ​​​​in the data set into three categories: high, medium, and low according to resource availability, and calculate the BLEU score on each category (the higher the better) , USM performs better than Whisper in every category.

Research has found that BEST-RQ pre-training is an effective way to extend speech representation learning to large data sets.

When combined with text injection in MOST, it improves the quality of downstream speech tasks, achieving state-of-the-art results on the FLEURS and CoVoST 2 benchmarks performance.

By training lightweight residual adapter modules, MOST represents the ability to quickly adapt to new domains. These remaining adapter modules only increase parameters by 2%.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

Google said that currently, USM supports more than 100 languages ​​and will expand to more than 100 languages ​​in the future. More than 1000 languages. With this technology, it may be safe for everyone to travel around the world.

#Even, real-time translation of Google AR glasses products will attract many fans in the future.

#However, the application of this technology still has a long way to go.

#After all, in its speech at the IO conference facing the world, Google also wrote the Arabic text backwards, attracting many netizens to watch.

Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages

The above is the detailed content of Beat OpenAI again! Google releases 2 billion parameter universal model to automatically recognize and translate more than 100 languages. For more information, please follow other related articles on the PHP Chinese website!

Related labels:
source:51cto.com
Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
Popular Tutorials
More>
Latest Downloads
More>
Web Effects
Website Source Code
Website Materials
Front End Template