Skip to content

Commit c5cf232

Browse files
committed
auto-generating sphinx docs
1 parent f8f700d commit c5cf232

File tree

5 files changed

+19
-35
lines changed

5 files changed

+19
-35
lines changed

main/_modules/torchtune/modules/peft/_utils.html

-15
Original file line numberDiff line numberDiff line change
@@ -457,7 +457,6 @@ <h1>Source code for torchtune.modules.peft._utils</h1><div class="highlight"><pr
457457

458458
<span class="kn">import</span><span class="w"> </span><span class="nn">torch</span>
459459
<span class="kn">from</span><span class="w"> </span><span class="nn">torch</span><span class="w"> </span><span class="kn">import</span> <span class="n">nn</span>
460-
<span class="kn">from</span><span class="w"> </span><span class="nn">torchtune.utils._logging</span><span class="w"> </span><span class="kn">import</span> <span class="n">deprecated</span>
461460

462461
<span class="c1"># Modules from MultiHeadAttention that LoRA can be applied to</span>
463462
<span class="n">LORA_ATTN_MODULES</span> <span class="o">=</span> <span class="n">Literal</span><span class="p">[</span><span class="s2">&quot;q_proj&quot;</span><span class="p">,</span> <span class="s2">&quot;k_proj&quot;</span><span class="p">,</span> <span class="s2">&quot;v_proj&quot;</span><span class="p">,</span> <span class="s2">&quot;output_proj&quot;</span><span class="p">]</span>
@@ -761,20 +760,6 @@ <h1>Source code for torchtune.modules.peft._utils</h1><div class="highlight"><pr
761760
<span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="sa">f</span><span class="s2">&quot;Missing LoRA key </span><span class="si">{</span><span class="n">k</span><span class="si">}</span><span class="s2"> from adapter state dict&quot;</span><span class="p">)</span>
762761
<span class="k">if</span> <span class="n">lora_unexpected</span><span class="p">:</span>
763762
<span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">&quot;Unexpected key loading adapter&quot;</span><span class="p">)</span></div>
764-
765-
766-
<span class="nd">@deprecated</span><span class="p">(</span>
767-
<span class="n">msg</span><span class="o">=</span><span class="s2">&quot;load_dora_magnitudes will be deprecated in 0.6.0. Please use DoRALinear.initialize_dora_magnitude instead.&quot;</span>
768-
<span class="p">)</span>
769-
<span class="k">def</span><span class="w"> </span><span class="nf">load_dora_magnitudes</span><span class="p">(</span><span class="n">model</span><span class="p">:</span> <span class="n">nn</span><span class="o">.</span><span class="n">Module</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="kc">None</span><span class="p">:</span>
770-
<span class="w"> </span><span class="sd">&quot;&quot;&quot;</span>
771-
<span class="sd"> For DoRA magnitude we use setattr to move from meta device</span>
772-
<span class="sd"> &quot;&quot;&quot;</span>
773-
<span class="n">dora_parents</span> <span class="o">=</span> <span class="p">{</span>
774-
<span class="n">n</span><span class="p">:</span> <span class="n">p</span> <span class="k">for</span> <span class="n">n</span><span class="p">,</span> <span class="n">p</span> <span class="ow">in</span> <span class="n">model</span><span class="o">.</span><span class="n">named_modules</span><span class="p">()</span> <span class="k">if</span> <span class="nb">hasattr</span><span class="p">(</span><span class="n">p</span><span class="p">,</span> <span class="s2">&quot;adapter_params&quot;</span><span class="p">)</span>
775-
<span class="p">}</span>
776-
<span class="n">sd</span> <span class="o">=</span> <span class="p">{</span><span class="sa">f</span><span class="s2">&quot;</span><span class="si">{</span><span class="n">n</span><span class="si">}</span><span class="s2">.magnitude&quot;</span><span class="p">:</span> <span class="n">p</span><span class="o">.</span><span class="n">magnitude</span> <span class="k">for</span> <span class="n">n</span><span class="p">,</span> <span class="n">p</span> <span class="ow">in</span> <span class="n">dora_parents</span><span class="o">.</span><span class="n">items</span><span class="p">()}</span>
777-
<span class="n">model</span><span class="o">.</span><span class="n">load_state_dict</span><span class="p">(</span><span class="n">sd</span><span class="p">,</span> <span class="n">strict</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="n">assign</span><span class="o">=</span><span class="kc">True</span><span class="p">)</span>
778763
</pre></div>
779764

780765
</article>

main/_modules/torchtune/training/quantization.html

+2-7
Original file line numberDiff line numberDiff line change
@@ -453,7 +453,6 @@ <h1>Source code for torchtune.training.quantization</h1><div class="highlight"><
453453
<span class="c1"># LICENSE file in the root directory of this source tree.</span>
454454

455455
<span class="kn">from</span><span class="w"> </span><span class="nn">typing</span><span class="w"> </span><span class="kn">import</span> <span class="n">Callable</span><span class="p">,</span> <span class="n">Optional</span>
456-
<span class="kn">from</span><span class="w"> </span><span class="nn">warnings</span><span class="w"> </span><span class="kn">import</span> <span class="n">warn</span>
457456

458457
<span class="kn">from</span><span class="w"> </span><span class="nn">torch</span><span class="w"> </span><span class="kn">import</span> <span class="n">nn</span>
459458
<span class="kn">from</span><span class="w"> </span><span class="nn">torchtune.modules.peft.lora</span><span class="w"> </span><span class="kn">import</span> <span class="n">LoRALinear</span><span class="p">,</span> <span class="n">QATLoRALinear</span>
@@ -592,6 +591,7 @@ <h1>Source code for torchtune.training.quantization</h1><div class="highlight"><
592591
<span class="s2">&quot;4w-qat-module-swap&quot;</span>
593592
<span class="p">]</span> <span class="o">=</span> <span class="n">enable_4w_fake_quant_module_swap</span>
594593

594+
595595
<span class="c1"># int8 dynamic activations + int4 weight</span>
596596
<span class="k">class</span><span class="w"> </span><span class="nc">Int8DynActInt4WeightQATQuantizerModuleSwap</span><span class="p">(</span><span class="n">Int8DynActInt4WeightQATQuantizer</span><span class="p">):</span>
597597
<span class="k">pass</span>
@@ -627,12 +627,7 @@ <h1>Source code for torchtune.training.quantization</h1><div class="highlight"><
627627
<span class="sd"> Returns:</span>
628628
<span class="sd"> Optional[str]: The quantization mode.</span>
629629
<span class="sd"> &quot;&quot;&quot;</span>
630-
<span class="n">mode</span> <span class="o">=</span> <span class="n">_quantizer_to_mode</span><span class="o">.</span><span class="n">get</span><span class="p">(</span><span class="nb">type</span><span class="p">(</span><span class="n">quantizer</span><span class="p">),</span> <span class="kc">None</span><span class="p">)</span>
631-
<span class="k">if</span> <span class="n">mode</span> <span class="ow">is</span> <span class="ow">not</span> <span class="kc">None</span> <span class="ow">and</span> <span class="s2">&quot;module-swap&quot;</span> <span class="ow">in</span> <span class="n">mode</span><span class="p">:</span>
632-
<span class="n">warn</span><span class="p">(</span>
633-
<span class="s2">&quot;*QuantizerModuleSwap is deprecated. Please use the version without &#39;ModuleSwap&#39; instead&quot;</span>
634-
<span class="p">)</span>
635-
<span class="k">return</span> <span class="n">mode</span></div>
630+
<span class="k">return</span> <span class="n">_quantizer_to_mode</span><span class="o">.</span><span class="n">get</span><span class="p">(</span><span class="nb">type</span><span class="p">(</span><span class="n">quantizer</span><span class="p">),</span> <span class="kc">None</span><span class="p">)</span></div>
636631

637632

638633
<span class="k">def</span><span class="w"> </span><span class="nf">_get_disable_fake_quant</span><span class="p">(</span><span class="n">quantizer_mode</span><span class="p">:</span> <span class="nb">str</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="n">Callable</span><span class="p">:</span>

main/_sources/tutorials/e2e_flow.rst.txt

+8-6
Original file line numberDiff line numberDiff line change
@@ -275,18 +275,20 @@ Let's first copy over the config to our local working directory so we can make c
275275
276276
$ tune cp generation ./custom_generation_config.yaml
277277
Copied file to custom_generation_config.yaml
278+
$ mkdir /tmp/torchtune/llama3_2_3B/lora_single_device/out
278279
279280
Let's modify ``custom_generation_config.yaml`` to include the following changes. Again, you only need
280281
to replace two fields: ``output_dir`` and ``checkpoint_files``
281282

282283
.. code-block:: yaml
283284
284-
output_dir: /tmp/torchtune/llama3_2_3B/lora_single_device/epoch_0
285+
checkpoint_dir: /tmp/torchtune/llama3_2_3B/lora_single_device/epoch_0
286+
output_dir: /tmp/torchtune/llama3_2_3B/lora_single_device/out
285287
286288
# Tokenizer
287289
tokenizer:
288290
_component_: torchtune.models.llama3.llama3_tokenizer
289-
path: ${output_dir}/original/tokenizer.model
291+
path: ${checkpoint_dir}/original/tokenizer.model
290292
prompt_template: null
291293
292294
model:
@@ -295,7 +297,7 @@ Let's modify ``custom_generation_config.yaml`` to include the following changes.
295297
296298
checkpointer:
297299
_component_: torchtune.training.FullModelHFCheckpointer
298-
checkpoint_dir: ${output_dir}
300+
checkpoint_dir: ${checkpoint_dir}
299301
checkpoint_files: [
300302
ft-model-00001-of-00002.safetensors,
301303
ft-model-00002-of-00002.safetensors,
@@ -312,8 +314,8 @@ Let's modify ``custom_generation_config.yaml`` to include the following changes.
312314
313315
# Generation arguments; defaults taken from gpt-fast
314316
prompt:
315-
system: null
316-
user: "Tell me a joke. "
317+
system: null
318+
user: "Tell me a joke. "
317319
max_new_tokens: 300
318320
temperature: 0.6 # 0.8 and 0.6 are popular values to try
319321
top_k: 300
@@ -330,7 +332,7 @@ these parameters.
330332

331333
.. code-block:: text
332334
333-
$ tune run generate --config ./custom_generation_config.yaml prompt="tell me a joke. "
335+
$ tune run generate --config ./custom_generation_config.yaml prompt.user="Tell me a joke. "
334336
Tell me a joke. Here's a joke for you:
335337
336338
What do you call a fake noodle?

main/searchindex.js

+1-1
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

main/tutorials/e2e_flow.html

+8-6
Original file line numberDiff line numberDiff line change
@@ -707,18 +707,20 @@ <h3>Generate some output<a class="headerlink" href="#generate-some-output" title
707707
<p>Let’s first copy over the config to our local working directory so we can make changes.</p>
708708
<div class="highlight-text notranslate"><div class="highlight"><pre><span></span>$ tune cp generation ./custom_generation_config.yaml
709709
Copied file to custom_generation_config.yaml
710+
$ mkdir /tmp/torchtune/llama3_2_3B/lora_single_device/out
710711
</pre></div>
711712
</div>
712713
<dl class="simple">
713714
<dt>Let’s modify <code class="docutils literal notranslate"><span class="pre">custom_generation_config.yaml</span></code> to include the following changes. Again, you only need</dt><dd><p>to replace two fields: <code class="docutils literal notranslate"><span class="pre">output_dir</span></code> and <code class="docutils literal notranslate"><span class="pre">checkpoint_files</span></code></p>
714715
</dd>
715716
</dl>
716-
<div class="highlight-yaml notranslate"><div class="highlight"><pre><span></span><span class="nt">output_dir</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">/tmp/torchtune/llama3_2_3B/lora_single_device/epoch_0</span>
717+
<div class="highlight-yaml notranslate"><div class="highlight"><pre><span></span><span class="nt">checkpoint_dir</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">/tmp/torchtune/llama3_2_3B/lora_single_device/epoch_0</span>
718+
<span class="nt">output_dir</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">/tmp/torchtune/llama3_2_3B/lora_single_device/out</span>
717719

718720
<span class="c1"># Tokenizer</span>
719721
<span class="nt">tokenizer</span><span class="p">:</span>
720722
<span class="w"> </span><span class="nt">_component_</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">torchtune.models.llama3.llama3_tokenizer</span>
721-
<span class="w"> </span><span class="nt">path</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">${output_dir}/original/tokenizer.model</span>
723+
<span class="w"> </span><span class="nt">path</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">${checkpoint_dir}/original/tokenizer.model</span>
722724
<span class="w"> </span><span class="nt">prompt_template</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">null</span>
723725

724726
<span class="nt">model</span><span class="p">:</span>
@@ -727,7 +729,7 @@ <h3>Generate some output<a class="headerlink" href="#generate-some-output" title
727729

728730
<span class="nt">checkpointer</span><span class="p">:</span>
729731
<span class="w"> </span><span class="nt">_component_</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">torchtune.training.FullModelHFCheckpointer</span>
730-
<span class="w"> </span><span class="nt">checkpoint_dir</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">${output_dir}</span>
732+
<span class="w"> </span><span class="nt">checkpoint_dir</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">${checkpoint_dir}</span>
731733
<span class="w"> </span><span class="nt">checkpoint_files</span><span class="p">:</span><span class="w"> </span><span class="p p-Indicator">[</span>
732734
<span class="w"> </span><span class="nv">ft-model-00001-of-00002.safetensors</span><span class="p p-Indicator">,</span>
733735
<span class="w"> </span><span class="nv">ft-model-00002-of-00002.safetensors</span><span class="p p-Indicator">,</span>
@@ -744,8 +746,8 @@ <h3>Generate some output<a class="headerlink" href="#generate-some-output" title
744746

745747
<span class="c1"># Generation arguments; defaults taken from gpt-fast</span>
746748
<span class="nt">prompt</span><span class="p">:</span>
747-
<span class="nt">system</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">null</span>
748-
<span class="nt">user</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;Tell</span><span class="nv"> </span><span class="s">me</span><span class="nv"> </span><span class="s">a</span><span class="nv"> </span><span class="s">joke.</span><span class="nv"> </span><span class="s">&quot;</span>
749+
<span class="w"> </span><span class="nt">system</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">null</span>
750+
<span class="w"> </span><span class="nt">user</span><span class="p">:</span><span class="w"> </span><span class="s">&quot;Tell</span><span class="nv"> </span><span class="s">me</span><span class="nv"> </span><span class="s">a</span><span class="nv"> </span><span class="s">joke.</span><span class="nv"> </span><span class="s">&quot;</span>
749751
<span class="nt">max_new_tokens</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">300</span>
750752
<span class="nt">temperature</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">0.6</span><span class="w"> </span><span class="c1"># 0.8 and 0.6 are popular values to try</span>
751753
<span class="nt">top_k</span><span class="p">:</span><span class="w"> </span><span class="l l-Scalar l-Scalar-Plain">300</span>
@@ -760,7 +762,7 @@ <h3>Generate some output<a class="headerlink" href="#generate-some-output" title
760762
<code class="docutils literal notranslate"><span class="pre">temperature=0.8</span></code>. These parameters control how the probabilities for
761763
sampling are computed. We recommend inspecting the model with these before playing around with
762764
these parameters.</p>
763-
<div class="highlight-text notranslate"><div class="highlight"><pre><span></span>$ tune run generate --config ./custom_generation_config.yaml prompt=&quot;tell me a joke. &quot;
765+
<div class="highlight-text notranslate"><div class="highlight"><pre><span></span>$ tune run generate --config ./custom_generation_config.yaml prompt.user=&quot;Tell me a joke. &quot;
764766
Tell me a joke. Here&#39;s a joke for you:
765767

766768
What do you call a fake noodle?

0 commit comments

Comments
 (0)