Skip to content

Commit

Permalink
Deployed 727c787 with MkDocs version: 1.5.3
Browse files Browse the repository at this point in the history
  • Loading branch information
jndiogo committed Feb 10, 2024
1 parent 5f47159 commit 1219e22
Show file tree
Hide file tree
Showing 3 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion search/search_index.json

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion setup-local-models/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -691,7 +691,7 @@ <h2 id="use-the-model-with-modeldir">Use the model with ModelDir<a class="header
</span></code></pre></div>
<p>Note that after "llamacpp:", instead of the model name we're directly passing the filename. If you plan to use a model for a while, creating an entry in ModelDir is more flexible.</p>
<h2 id="out-of-memory-running-local-models">Out of memory running local models<a class="headerlink" href="#out-of-memory-running-local-models" title="Permanent link">#</a></h2>
<p>An important thing to know if you'll be using local models is about Out of memory errors.</p>
<p>An important thing to know if you'll be using local models is about "Out of memory" errors.</p>
<p>A 7B model like OpenChat-3.5, when quantized to 4 bits will occupy about 6.8 Gb of memory, in either GPU's VRAM or common RAM. If you try to run a second model at the same time, you might get an out of memory error and/or llama.cpp may crash.</p>
<p>This is less of a problem when running scripts from the command line, but in environments like Jupyter where you can have multiple open notebooks, you may get python kernel errors like:</p>
<div class="language-text highlight"><pre><span></span><code><span id="__span-10-1"><a id="__codelineno-10-1" name="__codelineno-10-1" href="#__codelineno-10-1"></a>Kernel Restarting
Expand Down
Binary file modified sitemap.xml.gz
Binary file not shown.

0 comments on commit 1219e22

Please sign in to comment.