Skip to content

Questions related to llama.cpp options #3111

Answered by staviq
zastroyshchik asked this question in Q&A

You must be logged in to vote
  1. Context is what a model can see and work with, basically the only thing that can change dynamically, the rest is pretty much "static" and it cannot be changed at runtime. Models trained for particular size of context, are trained on examples not exceeding that size (not exceeding the "text length" limit). So artificially increasing context ("models working memory"), won't do much, because model is somewhat "hardcoded" to only fill the context up to the predefined size. To overcome that, rope scaling may be used to force the model to use bigger context, but that kind of makes is selectively ignore certain words or parts of the text, it's sort of like JPEG, when you decrease data quality t…

Replies: 2 comments 9 replies

You must be logged in to vote
1 reply
@zastroyshchik

Answer selected by zastroyshchik

You must be logged in to vote
8 replies
@Green-Sky

@zastroyshchik

@staviq

@zastroyshchik

@Green-Sky

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants