{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":724855841,"defaultBranch":"main","name":"mlx-examples","ownerLogin":"ml-explore","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-11-28T23:37:49.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/102832242?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1715359902.0","currentOid":""},"activityList":{"items":[{"before":"6f0a69e682e575458b2966fae2fdba6a20aa5f8e","after":"10853b57d9a5a8b696ac3d0ad9bc71e1b0d15c6f","ref":"refs/heads/main","pushedAt":"2024-05-10T17:13:34.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Add `model_config` parameter to `load()` and `load_model()` (#770)\n\n* Add `model_config` parameter to `load()` and `load_model()`\r\n\r\nFor easy editing of the loaded model configuration (e.g., for changing RoPE theta or scaling of Phi-3 model)\r\n\r\nExample:\r\n\r\n```python\r\nfrom mlx_lm import load, generate\r\nmodel, tokenizer = load(\"mlx-community/Phi-3-mini-4k-instruct-4bit-no-q-embed\", model_config={\"rope_theta\":50000.0})\r\nresponse = generate(model, tokenizer, prompt, max_tokens=MAX_TOKENS)\r\n```\r\n\r\n* Possible bug (default_loss)\r\n\r\n* Revert \"Possible bug (default_loss)\"\r\n\r\nThis reverts commit 70a55ace1847f545300d9e62df835f65386f4fc0.\r\n\r\n* Fix default_loss for lora\r\n\r\n* 1. move load_model's new optional `model_config` arg to the end (fetch_from_hub()'s `model = load_model(model_path, lazy)`) 2. fix indentations (`black` hook)","shortMessageHtmlLink":"Add model_config parameter to load() and load_model() (#770)"}},{"before":"fad959837232b0b4deeaa29b147dc69cbc9c5d19","after":"6f0a69e682e575458b2966fae2fdba6a20aa5f8e","ref":"refs/heads/main","pushedAt":"2024-05-10T16:51:42.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"fix lora for openelm (#773)","shortMessageHtmlLink":"fix lora for openelm (#773)"}},{"before":"fd31f853ba0770d7ef43cd816c31ce5edc38e29e","after":null,"ref":"refs/heads/fix_openelm","pushedAt":"2024-05-10T16:51:42.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"}},{"before":null,"after":"fd31f853ba0770d7ef43cd816c31ce5edc38e29e","ref":"refs/heads/fix_openelm","pushedAt":"2024-05-10T15:59:07.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"fix lora for openelm","shortMessageHtmlLink":"fix lora for openelm"}},{"before":null,"after":"dc1c60b20b9244733e91366cbc7957fdd5a83bfd","ref":"refs/heads/non_incremental_kv","pushedAt":"2024-05-08T17:33:05.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Support non incremental kv cache growth","shortMessageHtmlLink":"Support non incremental kv cache growth"}},{"before":"ee60e2a9d585788da30efa90326be9d2f1bceb97","after":"fad959837232b0b4deeaa29b147dc69cbc9c5d19","ref":"refs/heads/main","pushedAt":"2024-05-08T15:35:54.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Fix llama cache check (#763)\n\n* fix llama cache check\r\n\r\n* add test","shortMessageHtmlLink":"Fix llama cache check (#763)"}},{"before":"2ba984d1495a5de344f997f979dfc88541d56e33","after":null,"ref":"refs/heads/llama_fix","pushedAt":"2024-05-08T15:35:54.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"}},{"before":"65fc0176a63e736d50f88d8d122ba0d268d3bd13","after":"2ba984d1495a5de344f997f979dfc88541d56e33","ref":"refs/heads/llama_fix","pushedAt":"2024-05-08T15:31:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"add test","shortMessageHtmlLink":"add test"}},{"before":null,"after":"65fc0176a63e736d50f88d8d122ba0d268d3bd13","ref":"refs/heads/llama_fix","pushedAt":"2024-05-08T15:20:58.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"fix llama cache check","shortMessageHtmlLink":"fix llama cache check"}},{"before":"539d6e7927f43c1442f5e2121a02b7ac14b09947","after":null,"ref":"refs/heads/kv_cache","pushedAt":"2024-05-08T15:18:14.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"}},{"before":"bfbc0e434a02e96a822ccf7bbcc7ae49ca951db6","after":"ee60e2a9d585788da30efa90326be9d2f1bceb97","ref":"refs/heads/main","pushedAt":"2024-05-08T15:18:13.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Kv cache (#643)\n\n* in place kv_cache\r\n\r\n* fix\r\n\r\n* fix kv cache size\r\n\r\n* partially fix kv cache dtype\r\n\r\n* step kv cache\r\n\r\n* multiple of step size\r\n\r\n* more teests + kv cache\r\n\r\n* more kv cache\r\n\r\n* udpate all models to use kv cache","shortMessageHtmlLink":"Kv cache (#643)"}},{"before":"6711a913b6d53047db9020f27323377d8f9aeaee","after":"539d6e7927f43c1442f5e2121a02b7ac14b09947","ref":"refs/heads/kv_cache","pushedAt":"2024-05-08T14:57:54.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"udpate all models to use kv cache","shortMessageHtmlLink":"udpate all models to use kv cache"}},{"before":"6b6586ed5bd0743c3c35de4d68d6e441883e00c9","after":"6711a913b6d53047db9020f27323377d8f9aeaee","ref":"refs/heads/kv_cache","pushedAt":"2024-05-08T14:57:34.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"udpate all models to use kv cache","shortMessageHtmlLink":"udpate all models to use kv cache"}},{"before":"c0019c4908068ac9a3a62d8f2e376575b93dbd26","after":"bfbc0e434a02e96a822ccf7bbcc7ae49ca951db6","ref":"refs/heads/main","pushedAt":"2024-05-08T13:04:36.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Add optional EOS token for llava example (#753)\n\n* add optional EOS token\r\n\r\n* add tokenizer config to align with MLX LM example\r\n\r\n* formtatting fixes","shortMessageHtmlLink":"Add optional EOS token for llava example (#753)"}},{"before":"f9a16b60853d29625d2e5244f77b2117c87b85e3","after":"6b6586ed5bd0743c3c35de4d68d6e441883e00c9","ref":"refs/heads/kv_cache","pushedAt":"2024-05-06T15:45:03.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"more kv cache","shortMessageHtmlLink":"more kv cache"}},{"before":"f30413b63c37de73751ac93b0dec305d26b87b20","after":"c0019c4908068ac9a3a62d8f2e376575b93dbd26","ref":"refs/heads/main","pushedAt":"2024-05-04T23:32:25.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Pad mask with zeros for non-square attention matrices (#715)\n\n* Pad mask with zeros for non-square attention matrices\r\n\r\nThe current implementation of the mask assumes the attention matrix is square, which is true if there is no cache. However, if one wishes to produce multiple tokens at a time, such as in speculative decoding implementations, a rectangular mask is necessary.\r\n\r\nThis change pads the bottom of the mask with zeros so multi-token decoding with a cache works correctly.\r\n\r\n* Directly create mask instead of padding\r\n\r\n* Update llama.py","shortMessageHtmlLink":"Pad mask with zeros for non-square attention matrices (#715)"}},{"before":"2bf11c463318f83f785a67bf7b7d573cba7ac57c","after":"f30413b63c37de73751ac93b0dec305d26b87b20","ref":"refs/heads/main","pushedAt":"2024-05-04T13:52:42.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"chore(mlx-lm): fix the number of validation batches configuration. (#752)\n\n* chore: fix number of validation batches\r\n\r\n* clean up\r\n\r\n* address comment","shortMessageHtmlLink":"chore(mlx-lm): fix the number of validation batches configuration. (#752"}},{"before":"4a0e04b236dfb887248b20c3cab401c2ce34f9bc","after":null,"ref":"refs/heads/mnist","pushedAt":"2024-05-04T00:13:06.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"}},{"before":"d1c35fa684ca070ad9e79d99885a2cd4261f8b8f","after":"2bf11c463318f83f785a67bf7b7d573cba7ac57c","ref":"refs/heads/main","pushedAt":"2024-05-04T00:13:05.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Use stable url for MNIST (#749)\n\n* use stable url\r\n\r\n* remove deprecated flag","shortMessageHtmlLink":"Use stable url for MNIST (#749)"}},{"before":"ee7a328116103e46e285680ff65e688644da1c36","after":"4a0e04b236dfb887248b20c3cab401c2ce34f9bc","ref":"refs/heads/mnist","pushedAt":"2024-05-03T23:39:22.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"remove deprecated flag","shortMessageHtmlLink":"remove deprecated flag"}},{"before":null,"after":"ee7a328116103e46e285680ff65e688644da1c36","ref":"refs/heads/mnist","pushedAt":"2024-05-03T20:21:15.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"use stable url","shortMessageHtmlLink":"use stable url"}},{"before":"b468091f7ff0b289352f773b9793c5613437465d","after":"d1c35fa684ca070ad9e79d99885a2cd4261f8b8f","ref":"refs/heads/main","pushedAt":"2024-05-03T19:42:49.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Add MLX Cache Limit setting for mlx_lm.generate and mlx_lm.server CLI (#744)\n\n* Add support for setting MLX cache limit in GB\r\n\r\n* Add support for setting MLX cache limit in GB in mlx_lm.server\r\n\r\n* format\r\n\r\n---------\r\n\r\nCo-authored-by: Awni Hannun ","shortMessageHtmlLink":"Add MLX Cache Limit setting for mlx_lm.generate and mlx_lm.server CLI ("}},{"before":"92430df0a03355ef2bc07f4a442bd0381780b4b9","after":"b468091f7ff0b289352f773b9793c5613437465d","ref":"refs/heads/main","pushedAt":"2024-05-03T19:20:13.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Add model management functionality for local caches (#736)\n\n* Add model management functionality for local caches\r\n\r\nThis commit introduces a set of command-line utilities for managing MLX models downloaded and saved locally in Hugging Face cache. The functionalities include scanning existing models, retrieving detailed information about a specific model, and deleting a model by its name.\r\n\r\n* Added mlx_lm.model to setup.py\r\n\r\n* nits\r\n\r\n---------\r\n\r\nCo-authored-by: Awni Hannun ","shortMessageHtmlLink":"Add model management functionality for local caches (#736)"}},{"before":"5079af62db260ac2ec838e8788c085e2cb8023d5","after":"92430df0a03355ef2bc07f4a442bd0381780b4b9","ref":"refs/heads/main","pushedAt":"2024-05-03T04:55:09.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Fix lora for qwen moe (#743)\n\n* fix lora for qwen moe\r\n\r\n* use max seq length in test as well","shortMessageHtmlLink":"Fix lora for qwen moe (#743)"}},{"before":"ffd3af167518522acf584b673331618d269de050","after":null,"ref":"refs/heads/fix_qwen_lora","pushedAt":"2024-05-03T04:55:09.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"}},{"before":"6775d6cb3f9426c13219ccd0dd854bbe899630a9","after":"5079af62db260ac2ec838e8788c085e2cb8023d5","ref":"refs/heads/main","pushedAt":"2024-05-03T04:22:04.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Update model card describe (#654)\n\n* Update model card describe\r\n\r\n- Add full link jump\r\n- Add the address of the model uploader's Hugging Face homepage\r\n\r\n* Add user_info to reduce whoami calls\r\n\r\n* Remove the -U argument\r\n\r\n* remove HF user info\r\n\r\n* run pre-commit","shortMessageHtmlLink":"Update model card describe (#654)"}},{"before":"bba9883bd6a2962f931e7a13a50b0e61af4a17f8","after":"ffd3af167518522acf584b673331618d269de050","ref":"refs/heads/fix_qwen_lora","pushedAt":"2024-05-01T23:23:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"use max seq length in test as well","shortMessageHtmlLink":"use max seq length in test as well"}},{"before":null,"after":"bba9883bd6a2962f931e7a13a50b0e61af4a17f8","ref":"refs/heads/fix_qwen_lora","pushedAt":"2024-05-01T20:44:27.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"fix lora for qwen moe","shortMessageHtmlLink":"fix lora for qwen moe"}},{"before":"7b6cedb53a4ae6032a033378d9a60746ce92a2c4","after":"c228a9b706f3bd090980f16e78dc24883a13d585","ref":"refs/heads/block_mask_moe","pushedAt":"2024-05-01T20:31:58.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"use new API","shortMessageHtmlLink":"use new API"}},{"before":"4bf2eb17f2f68e107219e853e51d921f263aabae","after":"6775d6cb3f9426c13219ccd0dd854bbe899630a9","ref":"refs/heads/main","pushedAt":"2024-05-01T16:00:02.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"awni","name":"Awni Hannun","path":"/awni","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/1542805?s=80&v=4"},"commit":{"message":"Whisper: Add pip distribution configuration to support pip installations. (#739)\n\n* Whisper: rename whisper to mlx_whisper\r\n\r\n* Whisper: add setup.py config for publish\r\n\r\n* Whisper: add assets data to setup config\r\n\r\n* Whisper: pre-commit for setup.py\r\n\r\n* Whisper: Update README.md\r\n\r\n* Whisper: Update README.md\r\n\r\n* nits\r\n\r\n* fix package data\r\n\r\n* nit in readme\r\n\r\n---------\r\n\r\nCo-authored-by: Awni Hannun ","shortMessageHtmlLink":"Whisper: Add pip distribution configuration to support pip installati…"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAERsjaswA","startCursor":null,"endCursor":null}},"title":"Activity · ml-explore/mlx-examples"}