Skip to content

Commit

Permalink
bump to llamafile v0.7
Browse files Browse the repository at this point in the history
  • Loading branch information
amakropoulos committed Apr 1, 2024
1 parent 5b6bdbe commit 23b90cf
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 4 deletions.
6 changes: 3 additions & 3 deletions Runtime/LLM.cs
Original file line number Diff line number Diff line change
Expand Up @@ -82,8 +82,8 @@ public class LLM : LLMClient
("Phi 2 (small, decent)", "https://huggingface.co/TheBloke/phi-2-GGUF/resolve/main/phi-2.Q4_K_M.gguf?download=true"),
};
public int SelectedModel = 0;
private static readonly string serverZipUrl = "https://github.com/Mozilla-Ocho/llamafile/releases/download/0.6.2/llamafile-0.6.2.zip";
private static readonly string server = LLMUnitySetup.GetAssetPath("llamafile-0.6.2.exe");
private static readonly string serverZipUrl = "https://github.com/Mozilla-Ocho/llamafile/releases/download/0.7/llamafile-0.7.zip";
private static readonly string server = LLMUnitySetup.GetAssetPath("llamafile-0.7.exe");
private static readonly string apeARMUrl = "https://cosmo.zip/pub/cosmos/bin/ape-arm64.elf";
private static readonly string apeARM = LLMUnitySetup.GetAssetPath("ape-arm64.elf");
private static readonly string apeX86_64Url = "https://cosmo.zip/pub/cosmos/bin/ape-x86_64.elf";
Expand Down Expand Up @@ -294,7 +294,7 @@ private void CheckIfListening(string message)
try
{
ServerStatus status = JsonUtility.FromJson<ServerStatus>(message);
if (status.message == "model loaded")
if (status.msg == "HTTP server listening")
{
Debug.Log("LLM Server started!");
serverListening = true;
Expand Down
2 changes: 1 addition & 1 deletion Runtime/LLMInterface.cs
Original file line number Diff line number Diff line change
Expand Up @@ -128,7 +128,7 @@ public struct ServerStatus
public string level;
public string function;
public int line;
public string message;
public string msg;
public string hostname;
public int port;
}
Expand Down

0 comments on commit 23b90cf

Please sign in to comment.