1-Transformer_models-5-Decoder_models

中英文对照学习,效果更佳!
原课程链接:https://huggingface.co/course/chapter1/6?fw=pt

Decoder models

解码器型号

Ask a Question

问一个问题

Decoder models use only the decoder of a Transformer model. At each stage, for a given word the attention layers can only access the words positioned before it in the sentence. These models are often called auto-regressive models.

解码器模型仅使用Transformer模型的解码器。在每个阶段,对于一个给定的单词,注意力层只能访问句子中位于该单词之前的单词。这些模型通常被称为自回归模型。

The pretraining of decoder models usually revolves around predicting the next word in the sentence.

译码模型的预训练通常围绕着预测句子中的下一个单词。

These models are best suited for tasks involving text generation.

这些模型最适合于涉及文本生成的任务。

Representatives of this family of models include:

这一系列模型的代表包括: