tom-ci
tomci Blog
友情連結
關於我
  •   我是誰
  •   遊戲經歷
往期整理
  •   歷史歸檔
  •   文章分類
  •   文章標籤
Github主頁

tom-ci | 喜歡打遊戲的tomtom

0
Home
托福英文
力扣編程
原神開發
報錯分享
遊戲開發
技術分享
心情隨筆
深度學習
日文學習
Category
Deep learning Guide 14: Minibatch Stochastic Gradient Descent
深度學習
🧐Deep learning Guide 14: Minibatch Stochastic Gradient Descent
At the heart of the decision to use minibatches is computational efficiency. This is most easily understood when considering parallelization to multiple GPUs and multiple servers. In this case we need to send at least one image to each GPU. With 8 GPUs per server and 16 servers we already arrive at a minibatch size no smaller than 128.
Math
deep-learning
Python
Deep learning Guide 12: Gradient Descent 梯度下降
深度學習
🧐Deep learning Guide 12: Gradient Descent 梯度下降
Although it is rarely used directly in deep learning, an understanding of gradient descent is key to understanding stochastic gradient descent algorithms.
Math
Python
deep-learning
Deep learning Guide 13: Stochastic Gradient Descent 隨機梯度下降
深度學習
🧐Deep learning Guide 13: Stochastic Gradient Descent 隨機梯度下降
我们一直在训练过程中使用随机梯度下降,但没有解释它为什么起作用。为了澄清这一点,我们刚在 11.3节中描述了梯度下降的基本原则。本节继续更详细地说明随机梯度下降(stochastic gradient descent)。
Math
deep-learning
Python
課後功課答案  12.2. Convexity
深度學習
🧐課後功課答案 12.2. Convexity
Assume that we want to verify convexity of a set by drawing all lines between points within the set and checking whether the lines are contained.
Math
deep-learning
Python
Deep learning Guide 11: The Transformer Architecture
深度學習
😃Deep learning Guide 11: The Transformer Architecture
地表最強のTransformer Architecture: layout normal + residual connection +
Math
deep-learning
Python
課後功課答案 11.5. Self-Attention and Positional Encoding
深度學習
🧐課後功課答案 11.5. Self-Attention and Positional Encoding
Implement distance-based attention by modifying the DotProductAttention code. Note that you only need the squared norms of the keys for an efficient implementation.
Math
deep-learning
Python
課後功課答案 11.4. Multi-Head Attention
深度學習
🧐課後功課答案 11.4. Multi-Head Attention
Implement distance-based attention by modifying the DotProductAttention code. Note that you only need the squared norms of the keys for an efficient implementation.
Math
deep-learning
Python
Deep learning Guide 10: Multi-Head Attention, Self-Attention and Positional Encoding
深度學習
😋Deep learning Guide 10: Multi-Head Attention, Self-Attention and Positional Encoding
In deep learning, we often use CNNs or RNNs to encode sequences. Now with attention mechanisms in mind, imagine feeding a sequence of tokens into an attention mechanism such that at every step,
Math
deep-learning
Python
課後功課答案 11.3. Attention Scoring Functions
深度學習
🧐課後功課答案 11.3. Attention Scoring Functions
Implement distance-based attention by modifying the DotProductAttention code. Note that you only need the squared norms of the keys for an efficient implementation.
Math
deep-learning
Python
課後功課答案 11.2. Attention Pooling by Similarity
深度學習
🧐課後功課答案 11.2. Attention Pooling by Similarity
mack1982weak證明了Nadaraya-Watson估計是一致的。當您獲得更多數據時,您應該以多快的速度縮小注意力機制的規模?為你的答案提供一些直覺。
Math
deep-learning
Python
課後功課答案 11.1. Queries, Keys, and Values
深度學習
🧐課後功課答案 11.1. Queries, Keys, and Values
Suppose that you wanted to reimplement approximate (key, query) matches as used in classical databases, which attention function would you pick?
Math
deep-learning
Python
Deep learning Guide 9: Attention Mechanisms
深度學習
😋Deep learning Guide 9: Attention Mechanisms
來到deep learning的終點: Attention Mechanisms
deep-learning
Math
Python
Prev
12
Next
PrevNext
你好!接下來登場的是......
tom-ci

tom-ci

tom_ci,也可以叫我CI,tom,ic,湯姆希艾或者是湯姆,艾希,Macau SMP現任群主
了解更多

交流頻道

加入我們的社群討論分享

點擊加入社群
Latest posts
神人日記
2025-6-20
Deep learning Guide 11: The Transformer Architecture
2025-6-1
HTB - seasonal machine- Dog
2025-3-25
托福作文常用句武器庫
2025-3-13
HACK THE BOX 操作手冊
2025-3-13
大家的日本語-第12課
2024-9-26
Math
25
Python
23
deep-learning
21
其他
12
推薦
9
網站開發
5
前端開發
4
C++
3
hack the box
3
Unity
2
Unreal
2
ShaderToy
1
C#
1

文章数:
21
建站天数:
479 天
访问量:
访客数:

NotionNext 4.6.1 2021-2025 tom-ci.