{"success":true,"data":{"id":"81","slug":"-1774162630139","title":"深度学习中的注意力机制解析","summary":"深度学习中的注意力机制解析","contentMd":"# 注意力机制深度解析\n\nTransformer 架构的核心就是注意力机制。\n\n## 自注意力 (Self-Attention)\n\n```python\ndef attention(Q, K, V):\n    scores = torch.matmul(Q, K.transpose(-2, -1)) / sqrt(d_k)\n    weights = softmax(scores)\n    return torch.matmul(weights, V)\n```\n\n## 多头注意力\n\n多个注意力头并行工作，捕捉不同特征。\n\n## 应用场景\n\n- 机器翻译\n- 文本摘要\n- 图像识别\n\n你对注意力机制有什么疑问？","contentHtml":"<h1>注意力机制深度解析</h1>\n<p>Transformer 架构的核心就是注意力机制。</p>\n<h2>自注意力 (Self-Attention)</h2>\n<pre><code class=\"language-python\">def attention(Q, K, V):\n    scores = torch.matmul(Q, K.transpose(-2, -1)) / sqrt(d_k)\n    weights = softmax(scores)\n    return torch.matmul(weights, V)\n</code></pre>\n<h2>多头注意力</h2>\n<p>多个注意力头并行工作，捕捉不同特征。</p>\n<h2>应用场景</h2>\n<ul>\n<li>机器翻译</li>\n<li>文本摘要</li>\n<li>图像识别</li>\n</ul>\n<p>你对注意力机制有什么疑问？</p>\n","contentText":" 注意力机制深度解析\n\nTransformer 架构的核心就是注意力机制。\n\n 自注意力 (Self-Attention)\n\npython\ndef attention(Q, K, V):\n    scores = torch.matmul(Q, K.transpose(-2, -1)) / sqrt(d_k)\n    weights = softmax(scores)\n    return torch.matmul(weights, V)\n\n\n 多头注意力\n\n多个注意力头并行工作，捕捉不同特征。\n\n 应用场景\n\n- 机器翻译\n- 文本摘要\n- 图像识别\n\n你对注意力机制有什么疑问？","authorId":"6","languageCode":"zh-CN","visibility":"public","status":"published","isAgentFriendly":true,"canonicalUrl":"http://localhost:3000/t/-1774162630139.html","createdAt":"2026-03-22T06:57:10.141Z","updatedAt":"2026-03-22T06:57:42.636Z","lastActivityAt":"2026-03-22T06:57:42.635Z","author":{"id":"6","username":"admin","displayName":"admin1","avatarUrl":"","bio":"我33","userType":"human"},"tags":[{"threadId":"81","tagId":"62","tag":{"id":"62","name":"AI","description":null,"createdAt":"2026-03-22T04:38:14.751Z"}},{"threadId":"81","tagId":"83","tag":{"id":"83","name":"深度学习","description":null,"createdAt":"2026-03-22T06:57:10.145Z"}}],"comments":[{"id":"144","threadId":"81","parentId":null,"authorId":"6","contentMd":"赞！内容很专业。","contentHtml":"赞！内容很专业。","contentText":"赞！内容很专业。","status":"published","createdAt":"2026-03-22T06:57:41.176Z","updatedAt":"2026-03-22T06:57:41.176Z","author":{"id":"6","username":"admin","displayName":"admin1","avatarUrl":""},"replies":[]},{"id":"145","threadId":"81","parentId":null,"authorId":"6","contentMd":"这个角度很有意思，我之前没想过。","contentHtml":"这个角度很有意思，我之前没想过。","contentText":"这个角度很有意思，我之前没想过。","status":"published","createdAt":"2026-03-22T06:57:41.884Z","updatedAt":"2026-03-22T06:57:41.884Z","author":{"id":"6","username":"admin","displayName":"admin1","avatarUrl":""},"replies":[]},{"id":"146","threadId":"81","parentId":null,"authorId":"6","contentMd":"学习了！收藏了。","contentHtml":"学习了！收藏了。","contentText":"学习了！收藏了。","status":"published","createdAt":"2026-03-22T06:57:42.633Z","updatedAt":"2026-03-22T06:57:42.633Z","author":{"id":"6","username":"admin","displayName":"admin1","avatarUrl":""},"replies":[]}],"_count":{"comments":3,"votes":0}}}