transformer画图
时间: 2023-08-21 10:07:08 浏览: 52
在Transformer模型中,有多个Encoder和多个Decoder组成。每个Encoder和Decoder内部都由多个层堆叠而成,通常是6层。每一层都包括Self-Attention、Multi-Headed Attention、Feed-Forward Network等模块。Encoder和Decoder之间没有权重共享。以下是Transformer的一个示意图:
<em>1</em><em>2</em><em>3</em>
#### 引用[.reference_title]
- *1* [图解transformer | The Illustrated Transformer](https://blog.csdn.net/qq_36667170/article/details/124359818)[target="_blank" data-report-click={"spm":"1018.2226.3001.9630","extra":{"utm_source":"vip_chatgpt_common_search_pc_result","utm_medium":"distribute.pc_search_result.none-task-cask-2~all~insert_cask~default-1-null.142^v92^chatsearchT3_1"}} ] [.reference_item]
- *2* *3* [Transformer完全图解](https://blog.csdn.net/weixin_44305115/article/details/101622645)[target="_blank" data-report-click={"spm":"1018.2226.3001.9630","extra":{"utm_source":"vip_chatgpt_common_search_pc_result","utm_medium":"distribute.pc_search_result.none-task-cask-2~all~insert_cask~default-1-null.142^v92^chatsearchT3_1"}} ] [.reference_item]
[ .reference_list ]