| 
235 | 235 |           <div class="pytorch-left-menu-search">  | 
236 | 236 | 
 
  | 
237 | 237 |     <div class="version">  | 
238 |  | -      <a href='https://pytorch.org/docs/versions.html'>master (2.0.0a0+git0157e2e ) ▼</a>  | 
 | 238 | +      <a href='https://pytorch.org/docs/versions.html'>master (2.0.0a0+gitbb77907 ) ▼</a>  | 
239 | 239 |     </div>  | 
240 | 240 | 
 
  | 
241 | 241 | 
 
  | 
@@ -514,7 +514,8 @@ <h1>Source code for torch</h1><div class="highlight"><pre>  | 
514 | 514 |     <span class="s1">'set_deterministic_debug_mode'</span><span class="p">,</span> <span class="s1">'get_deterministic_debug_mode'</span><span class="p">,</span>  | 
515 | 515 |     <span class="s1">'set_float32_matmul_precision'</span><span class="p">,</span> <span class="s1">'get_float32_matmul_precision'</span><span class="p">,</span>  | 
516 | 516 |     <span class="s1">'set_warn_always'</span><span class="p">,</span> <span class="s1">'is_warn_always_enabled'</span><span class="p">,</span> <span class="s1">'SymInt'</span><span class="p">,</span> <span class="s1">'SymFloat'</span><span class="p">,</span>  | 
517 |  | -    <span class="s1">'sym_int'</span><span class="p">,</span> <span class="s1">'sym_float'</span><span class="p">,</span> <span class="s1">'compile'</span><span class="p">,</span> <span class="s1">'vmap'</span><span class="p">]</span>  | 
 | 517 | +    <span class="s1">'sym_int'</span><span class="p">,</span> <span class="s1">'sym_float'</span><span class="p">,</span> <span class="s1">'sym_max'</span><span class="p">,</span> <span class="s1">'sym_min'</span><span class="p">,</span> <span class="s1">'compile'</span><span class="p">,</span> <span class="s1">'vmap'</span>  | 
 | 518 | +<span class="p">]</span>  | 
518 | 519 | 
 
  | 
519 | 520 | <span class="c1">################################################################################</span>  | 
520 | 521 | <span class="c1"># Load the extension module</span>  | 
@@ -723,6 +724,12 @@ <h1>Source code for torch</h1><div class="highlight"><pre>  | 
723 | 724 |     <span class="k">def</span> <span class="fm">__ge__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">other</span><span class="p">)</span> <span class="o">-></span> <span class="n">builtins</span><span class="o">.</span><span class="n">bool</span><span class="p">:</span>  | 
724 | 725 |         <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
725 | 726 | 
 
  | 
 | 727 | +    <span class="k">def</span> <span class="nf">__sym_max__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">other</span><span class="p">):</span>  | 
 | 728 | +        <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
 | 729 | + | 
 | 730 | +    <span class="k">def</span> <span class="nf">__sym_min__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">other</span><span class="p">):</span>  | 
 | 731 | +        <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
 | 732 | + | 
726 | 733 |     <span class="k">def</span> <span class="nf">__sym_float__</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>  | 
727 | 734 |         <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
728 | 735 | 
 
  | 
@@ -767,6 +774,12 @@ <h1>Source code for torch</h1><div class="highlight"><pre>  | 
767 | 774 |     <span class="k">def</span> <span class="fm">__ge__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">other</span><span class="p">)</span> <span class="o">-></span> <span class="n">builtins</span><span class="o">.</span><span class="n">bool</span><span class="p">:</span>  | 
768 | 775 |         <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
769 | 776 | 
 
  | 
 | 777 | +    <span class="k">def</span> <span class="nf">__sym_max__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">other</span><span class="p">):</span>  | 
 | 778 | +        <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
 | 779 | + | 
 | 780 | +    <span class="k">def</span> <span class="nf">__sym_min__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">other</span><span class="p">):</span>  | 
 | 781 | +        <span class="k">raise</span> <span class="ne">AssertionError</span><span class="p">(</span><span class="s2">"type stub not overridden"</span><span class="p">)</span>  | 
 | 782 | + | 
770 | 783 |     <span class="k">def</span> <span class="fm">__repr__</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>  | 
771 | 784 |         <span class="k">return</span> <span class="bp">self</span><span class="o">.</span><span class="n">node</span><span class="o">.</span><span class="n">str</span><span class="p">()</span>  | 
772 | 785 | 
 
  | 
@@ -807,6 +820,22 @@ <h1>Source code for torch</h1><div class="highlight"><pre>  | 
807 | 820 |         <span class="k">return</span> <span class="n">_sym_floor</span><span class="p">(</span><span class="n">a</span><span class="p">)</span> <span class="k">if</span> <span class="n">a</span> <span class="o">></span> <span class="mi">0</span> <span class="k">else</span> <span class="n">_sym_ceil</span><span class="p">(</span><span class="n">a</span><span class="p">)</span>  | 
808 | 821 |     <span class="k">return</span> <span class="n">py_int</span><span class="p">(</span><span class="n">a</span><span class="p">)</span>  <span class="c1"># type: ignore[operator]</span></div>  | 
809 | 822 | 
 
  | 
 | 823 | +<div class="viewcode-block" id="sym_max"><a class="viewcode-back" href="../generated/torch.sym_max.html#torch.sym_max">[docs]</a><span class="k">def</span> <span class="nf">sym_max</span><span class="p">(</span><span class="n">a</span><span class="p">,</span> <span class="n">b</span><span class="p">):</span>  | 
 | 824 | +    <span class="sd">""" SymInt-aware utility for max()."""</span>  | 
 | 825 | +    <span class="k">if</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">a</span><span class="p">,</span> <span class="p">(</span><span class="n">SymInt</span><span class="p">,</span> <span class="n">SymFloat</span><span class="p">)):</span>  | 
 | 826 | +        <span class="k">return</span> <span class="n">a</span><span class="o">.</span><span class="n">__sym_max__</span><span class="p">(</span><span class="n">b</span><span class="p">)</span>  | 
 | 827 | +    <span class="k">elif</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">b</span><span class="p">,</span> <span class="p">(</span><span class="n">SymInt</span><span class="p">,</span> <span class="n">SymFloat</span><span class="p">)):</span>  | 
 | 828 | +        <span class="k">return</span> <span class="n">b</span><span class="o">.</span><span class="n">__sym_max__</span><span class="p">(</span><span class="n">a</span><span class="p">)</span>  | 
 | 829 | +    <span class="k">return</span> <span class="n">builtins</span><span class="o">.</span><span class="n">max</span><span class="p">(</span><span class="n">a</span><span class="p">,</span> <span class="n">b</span><span class="p">)</span>  <span class="c1"># type: ignore[operator]</span></div>  | 
 | 830 | + | 
 | 831 | +<div class="viewcode-block" id="sym_min"><a class="viewcode-back" href="../generated/torch.sym_min.html#torch.sym_min">[docs]</a><span class="k">def</span> <span class="nf">sym_min</span><span class="p">(</span><span class="n">a</span><span class="p">,</span> <span class="n">b</span><span class="p">):</span>  | 
 | 832 | +    <span class="sd">""" SymInt-aware utility for max()."""</span>  | 
 | 833 | +    <span class="k">if</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">a</span><span class="p">,</span> <span class="p">(</span><span class="n">SymInt</span><span class="p">,</span> <span class="n">SymFloat</span><span class="p">)):</span>  | 
 | 834 | +        <span class="k">return</span> <span class="n">a</span><span class="o">.</span><span class="n">__sym_min__</span><span class="p">(</span><span class="n">b</span><span class="p">)</span>  | 
 | 835 | +    <span class="k">elif</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">b</span><span class="p">,</span> <span class="p">(</span><span class="n">SymInt</span><span class="p">,</span> <span class="n">SymFloat</span><span class="p">)):</span>  | 
 | 836 | +        <span class="k">return</span> <span class="n">b</span><span class="o">.</span><span class="n">__sym_min__</span><span class="p">(</span><span class="n">a</span><span class="p">)</span>  | 
 | 837 | +    <span class="k">return</span> <span class="n">builtins</span><span class="o">.</span><span class="n">min</span><span class="p">(</span><span class="n">a</span><span class="p">,</span> <span class="n">b</span><span class="p">)</span>  <span class="c1"># type: ignore[operator]</span></div>  | 
 | 838 | + | 
810 | 839 | <span class="c1"># Check to see if we can load C extensions, and if not provide some guidance</span>  | 
811 | 840 | <span class="c1"># on what the problem might be.</span>  | 
812 | 841 | <span class="k">try</span><span class="p">:</span>  | 
@@ -1792,6 +1821,17 @@ <h1>Source code for torch</h1><div class="highlight"><pre>  | 
1792 | 1821 | 
 
  | 
1793 | 1822 | <span class="kn">from</span> <span class="nn">torch</span> <span class="kn">import</span> <span class="n">func</span> <span class="k">as</span> <span class="n">func</span>  | 
1794 | 1823 | <span class="kn">from</span> <span class="nn">torch.func</span> <span class="kn">import</span> <span class="n">vmap</span>  | 
 | 1824 | + | 
 | 1825 | +<span class="c1"># The function _sparse_coo_tensor_unsafe is removed from PyTorch</span>  | 
 | 1826 | +<span class="c1"># Python API (v. 1.13), here we temporarily provide its replacement</span>  | 
 | 1827 | +<span class="c1"># with a deprecation warning.</span>  | 
 | 1828 | +<span class="c1"># TODO: remove the function for PyTorch v 1.15.</span>  | 
 | 1829 | +<span class="k">def</span> <span class="nf">_sparse_coo_tensor_unsafe</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>  | 
 | 1830 | +    <span class="kn">import</span> <span class="nn">warnings</span>  | 
 | 1831 | +    <span class="n">warnings</span><span class="o">.</span><span class="n">warn</span><span class="p">(</span><span class="s1">'torch._sparse_coo_tensor_unsafe is deprecated, '</span>  | 
 | 1832 | +                  <span class="s1">'use torch.sparse_coo_tensor(..., check_invariants=False) instead.'</span><span class="p">)</span>  | 
 | 1833 | +    <span class="n">kwargs</span><span class="p">[</span><span class="s1">'check_invariants'</span><span class="p">]</span> <span class="o">=</span> <span class="kc">False</span>  | 
 | 1834 | +    <span class="k">return</span> <span class="n">torch</span><span class="o">.</span><span class="n">sparse_coo_tensor</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>  | 
1795 | 1835 | </pre></div>  | 
1796 | 1836 | 
 
  | 
1797 | 1837 |              </article>  | 
 | 
0 commit comments