Skip to content

Conversation

@davidpissarra
Copy link
Member

This PR refines the code completion support for MLC LLM. Previously, these code completion models used inappropriate templates, that were harming the completion task. This template was changed to execute the LM without memory, and also to ignore roles and separators. Code instruct models still use the normal conversational templates, since were trained for this purpose (e.g., CodeGeeX2, StableCode-Instruct). I was able to test it with the vscode hf code completion extension (solves #748, concludes #761). For this extension, models trained using the Fill-in-the-Middle objective are the most suitable (e.g., StarCoder). If you try others than these you may get worse results.

Additionaly this PR adds support for StableCode-Instruct and StableCode-Completion.

image

@tqchen tqchen merged commit fbce5a3 into mlc-ai:main Aug 16, 2023
jimscard added a commit to jimscard/mlc-llm that referenced this pull request Sep 15, 2023
* 'main' of https://github.com/jimscard/mlc-llm:
  [Doc] Minor update to `Build Android Package from Source` section (mlc-ai#785)
  added cors to fast api (mlc-ai#757)
  Update Llama2 cached sin/cos to use max_sequence_length (mlc-ai#780)
  Update gpu.rst to add sudo apt update before first install (mlc-ai#784)
  [Doc] Update doc for prebuilt models (mlc-ai#767)
  Improve code completion experience (mlc-ai#772)
  Automatically set 'offset' parameter if 'messages' parameter is set (mlc-ai#754)
  Update tokenizers-cpp to latest and fix rust build error (mlc-ai#762)
  [Utils] Skip generating benchmark scripts in cases (mlc-ai#759)
  [Android] Add libOpenCL-pixel for supporintg Pixel phones. (mlc-ai#723)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants