Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
×
Jun 20, 2021 · Based on our cost-effective pipeline, we pre-train two models: an encoder-decoder bilingual model with 11 billion parameters (CPM-2) and its ...
People also ask
In this work, we propose a cost-effective pipeline for large-scale pre-trained language models, including pre-training with knowledge inheritance, fine-tuning ...
CPM-2 is a 11 billion parameters pre-trained language model based on a standard Transformer architecture consisting of a bidirectional encoder and a ...
CPM is an open-source program on large-scale pre-trained models, which is conducted by Beijing Academy of Artificial Intelligence and Tsinghua University, ...
A unified framework named ERNIE 3.0 is proposed for pre-training large-scale knowledge enhanced models that fuses auto-regressive network and auto-encoding ...
CPM is a Transformer-based autoregressive language model, with 2.6 billion parameters and 100 GB Chinese training data.
We keep training and releasing large-scale PLMs in recent years, which are listed as follows. Welcome to try them. CPM-2. Cost-Effective Pre-trained Language ...
Apr 9, 2024 · 3. CPM-2 CPM-2 : Cost-efficient Pre-trained language Models (CPM-2) pre-trains bilingual (English and Chinese) 11B and 198B mixture-of-experts ...
Large-scale pre-trained models (PTMs) such as BERT and GPT have recently achieved great success and become a milestone in the field of artificial intelligence ( ...