Skip to content

Commit

Permalink
Deploying to gh-pages from @ 3e63fcc 🚀
Browse files Browse the repository at this point in the history
  • Loading branch information
vince62s committed Oct 20, 2023
1 parent ee441f0 commit ab67486
Show file tree
Hide file tree
Showing 22 changed files with 1,020 additions and 834 deletions.
66 changes: 46 additions & 20 deletions _modules/onmt/decoders/transformer.html

Large diffs are not rendered by default.

16 changes: 10 additions & 6 deletions _modules/onmt/encoders/transformer.html
Original file line number Diff line number Diff line change
Expand Up @@ -212,7 +212,11 @@ <h1>Source code for onmt.encoders.transformer</h1><div class="highlight"><pre>
<span class="kn">from</span> <span class="nn">onmt.modules.position_ffn</span> <span class="kn">import</span> <span class="n">PositionwiseFeedForward</span>
<span class="kn">from</span> <span class="nn">onmt.modules.position_ffn</span> <span class="kn">import</span> <span class="n">ActivationFunction</span>
<span class="kn">from</span> <span class="nn">onmt.utils.misc</span> <span class="kn">import</span> <span class="n">sequence_mask</span>
<span class="kn">from</span> <span class="nn">onmt.modules.rmsnorm</span> <span class="kn">import</span> <span class="n">RMSNorm</span>

<span class="k">try</span><span class="p">:</span>
<span class="kn">from</span> <span class="nn">apex.normalization</span> <span class="kn">import</span> <span class="n">FusedRMSNorm</span> <span class="k">as</span> <span class="n">RMSNorm</span>
<span class="k">except</span> <span class="ne">ImportError</span><span class="p">:</span>
<span class="kn">from</span> <span class="nn">onmt.modules.rmsnorm</span> <span class="kn">import</span> <span class="n">RMSNorm</span>


<span class="k">class</span> <span class="nc">TransformerEncoderLayer</span><span class="p">(</span><span class="n">nn</span><span class="o">.</span><span class="n">Module</span><span class="p">):</span>
Expand Down Expand Up @@ -283,6 +287,7 @@ <h1>Source code for onmt.encoders.transformer</h1><div class="highlight"><pre>
<span class="bp">self</span><span class="o">.</span><span class="n">layer_norm</span> <span class="o">=</span> <span class="n">RMSNorm</span><span class="p">(</span><span class="n">d_model</span><span class="p">,</span> <span class="n">eps</span><span class="o">=</span><span class="n">norm_eps</span><span class="p">)</span>
<span class="k">else</span><span class="p">:</span>
<span class="k">raise</span> <span class="ne">ValueError</span><span class="p">(</span><span class="sa">f</span><span class="s2">&quot;</span><span class="si">{</span><span class="n">layer_norm</span><span class="si">}</span><span class="s2"> layer norm type is not supported&quot;</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dropout_p</span> <span class="o">=</span> <span class="n">dropout</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dropout</span> <span class="o">=</span> <span class="n">nn</span><span class="o">.</span><span class="n">Dropout</span><span class="p">(</span><span class="n">dropout</span><span class="p">)</span>

<span class="k">def</span> <span class="nf">forward</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">layer_in</span><span class="p">,</span> <span class="n">mask</span><span class="p">):</span>
Expand All @@ -299,16 +304,15 @@ <h1>Source code for onmt.encoders.transformer</h1><div class="highlight"><pre>
<span class="n">context</span><span class="p">,</span> <span class="n">_</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">self_attn</span><span class="p">(</span>
<span class="n">norm_layer_in</span><span class="p">,</span> <span class="n">norm_layer_in</span><span class="p">,</span> <span class="n">norm_layer_in</span><span class="p">,</span> <span class="n">mask</span><span class="o">=</span><span class="n">mask</span>
<span class="p">)</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">dropout_p</span> <span class="o">&gt;</span> <span class="mi">0</span><span class="p">:</span>
<span class="n">context</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">dropout</span><span class="p">(</span><span class="n">context</span><span class="p">)</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">parallel_residual</span><span class="p">:</span>
<span class="c1"># feed_forward applies residual, so we remove and apply residual with un-normed</span>
<span class="n">layer_out</span> <span class="o">=</span> <span class="p">(</span>
<span class="bp">self</span><span class="o">.</span><span class="n">feed_forward</span><span class="p">(</span><span class="n">norm_layer_in</span><span class="p">)</span>
<span class="o">-</span> <span class="n">norm_layer_in</span>
<span class="o">+</span> <span class="n">layer_in</span>
<span class="o">+</span> <span class="bp">self</span><span class="o">.</span><span class="n">dropout</span><span class="p">(</span><span class="n">context</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">feed_forward</span><span class="p">(</span><span class="n">norm_layer_in</span><span class="p">)</span> <span class="o">-</span> <span class="n">norm_layer_in</span> <span class="o">+</span> <span class="n">layer_in</span> <span class="o">+</span> <span class="n">context</span>
<span class="p">)</span>
<span class="k">else</span><span class="p">:</span>
<span class="n">layer_out</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">dropout</span><span class="p">(</span><span class="n">context</span><span class="p">)</span> <span class="o">+</span> <span class="n">layer_in</span>
<span class="n">layer_out</span> <span class="o">=</span> <span class="n">context</span> <span class="o">+</span> <span class="n">layer_in</span>
<span class="n">layer_out</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">feed_forward</span><span class="p">(</span><span class="n">layer_out</span><span class="p">)</span>

<span class="k">return</span> <span class="n">layer_out</span>
Expand Down
Loading

0 comments on commit ab67486

Please sign in to comment.