{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":744753322,"defaultBranch":"main","name":"BentoVLLM","ownerLogin":"bentoml","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2024-01-17T23:53:26.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/49176046?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1717180824.0","currentOid":""},"activityList":{"items":[{"before":"da3bcf0f3299d40b11ce10192440e52a6403345f","after":"df583958b9201f848dd22e7055f25ea60ab338c4","ref":"refs/heads/main","pushedAt":"2024-05-31T18:41:10.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"feat: openai endpoints decorator with default parameters (#47)","shortMessageHtmlLink":"feat: openai endpoints decorator with default parameters (#47)"}},{"before":null,"after":"230f2c0a2a65528aa93176014b0720d8c2159dd0","ref":"refs/heads/feat/openai-endpoint-default-parameters","pushedAt":"2024-05-31T18:40:24.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"feat: openai endpoints decorator with default parameters","shortMessageHtmlLink":"feat: openai endpoints decorator with default parameters"}},{"before":"65befe879fdb77bc5505480cd31ce7020ee91c6c","after":"da3bcf0f3299d40b11ce10192440e52a6403345f","ref":"refs/heads/main","pushedAt":"2024-05-31T02:32:07.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"ssheng","name":"Sean Sheng","path":"/ssheng","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/861225?s=80&v=4"},"commit":{"message":"Merge pull request #46 from bentoml/pr-concurrency\n\nchore: set service API concurrency to match vLLM default max_num_seqs","shortMessageHtmlLink":"Merge pull request #46 from bentoml/pr-concurrency"}},{"before":null,"after":"77c08b998f143e49383717fc92ab68aa5a924c75","ref":"refs/heads/pr-concurrency","pushedAt":"2024-05-29T20:43:12.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"rickzx","name":"Rick Zhou","path":"/rickzx","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/22135348?s=80&v=4"},"commit":{"message":"chore: set service API concurrency to match vLLM default max_num_seqs","shortMessageHtmlLink":"chore: set service API concurrency to match vLLM default max_num_seqs"}},{"before":"8361e5a1d927a6591404ccce18df29640ce90f4e","after":"65befe879fdb77bc5505480cd31ce7020ee91c6c","ref":"refs/heads/main","pushedAt":"2024-05-29T18:18:04.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"chore: serving llama3-8b without importing model first (#45)","shortMessageHtmlLink":"chore: serving llama3-8b without importing model first (#45)"}},{"before":null,"after":"8257caa62afa87c90abf9b39b7530e990f7527c3","ref":"refs/heads/chore/llama3-8b-instruct-download","pushedAt":"2024-05-29T18:17:31.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"chore: serving llama3-8b without importing model first","shortMessageHtmlLink":"chore: serving llama3-8b without importing model first"}},{"before":"19c02ae769f040b990f4ae998c30925a115236fd","after":"8361e5a1d927a6591404ccce18df29640ce90f4e","ref":"refs/heads/main","pushedAt":"2024-05-29T17:42:33.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: openai endpoints for vllm 0.4.2 (#44)","shortMessageHtmlLink":"fix: openai endpoints for vllm 0.4.2 (#44)"}},{"before":null,"after":"5646007c087a27564abe8e27b73bda32cd3a7607","ref":"refs/heads/fix/vllm-openai","pushedAt":"2024-05-29T17:41:28.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: openai endpoints for vllm 0.4.2","shortMessageHtmlLink":"fix: openai endpoints for vllm 0.4.2"}},{"before":"03a3ff219b52002fb9f0ec48bca07b01370df127","after":"19c02ae769f040b990f4ae998c30925a115236fd","ref":"refs/heads/main","pushedAt":"2024-05-24T07:40:13.000Z","pushType":"pr_merge","commitsCount":3,"pusher":{"login":"rickzx","name":"Rick Zhou","path":"/rickzx","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/22135348?s=80&v=4"},"commit":{"message":"Fix compatibility issue with vllm 0.4.2. Enable prefix caching","shortMessageHtmlLink":"Fix compatibility issue with vllm 0.4.2. Enable prefix caching"}},{"before":"20e0471990f2d15b58222b98a2baaa3abb72e534","after":"d8583f72a65b058f699b4007cdfc49fcf5475e7b","ref":"refs/heads/pr-vllm-comp","pushedAt":"2024-05-24T07:37:51.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"rickzx","name":"Rick Zhou","path":"/rickzx","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/22135348?s=80&v=4"},"commit":{"message":"clean up manual gpu_memory_utilization configuration","shortMessageHtmlLink":"clean up manual gpu_memory_utilization configuration"}},{"before":null,"after":"20e0471990f2d15b58222b98a2baaa3abb72e534","ref":"refs/heads/pr-vllm-comp","pushedAt":"2024-05-24T07:23:53.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"rickzx","name":"Rick Zhou","path":"/rickzx","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/22135348?s=80&v=4"},"commit":{"message":"Fix compatibility issue withh vllm 0.4.2. Enable prefix caching","shortMessageHtmlLink":"Fix compatibility issue withh vllm 0.4.2. Enable prefix caching"}},{"before":"cd79cc3d2c8b5af9c480718e572ff58d852bb51e","after":"03a3ff219b52002fb9f0ec48bca07b01370df127","ref":"refs/heads/main","pushedAt":"2024-05-22T20:54:42.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"aarnphm","name":"Aaron Pham","path":"/aarnphm","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/29749331?s=80&v=4"},"commit":{"message":"chore: update vllm dependencies and synchronize bentos\n\nSigned-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>","shortMessageHtmlLink":"chore: update vllm dependencies and synchronize bentos"}},{"before":"e133531c095b54dbfa961725248f66af79535ea4","after":"cd79cc3d2c8b5af9c480718e572ff58d852bb51e","ref":"refs/heads/main","pushedAt":"2024-05-20T06:31:00.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"ssheng","name":"Sean Sheng","path":"/ssheng","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/861225?s=80&v=4"},"commit":{"message":"Merge pull request #41 from bentoml/jiang-dev5\n\nrequires HF_TOKEN","shortMessageHtmlLink":"Merge pull request #41 from bentoml/jiang-dev5"}},{"before":null,"after":"9fc684d34a5691c2998b72ced74d5b7587f451f7","ref":"refs/heads/jiang-dev5","pushedAt":"2024-05-20T05:59:17.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"bojiang","name":null,"path":"/bojiang","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/5886138?s=80&v=4"},"commit":{"message":"requires HF_TOKEN","shortMessageHtmlLink":"requires HF_TOKEN"}},{"before":null,"after":"e133531c095b54dbfa961725248f66af79535ea4","ref":"refs/heads/benchmark","pushedAt":"2024-05-09T17:07:01.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"Model/llama3 70b instruct awq (#40)\n\n* add llama3-70b-instruct-awq\r\n\r\n* model: add llama3-70b-instruct-awq","shortMessageHtmlLink":"Model/llama3 70b instruct awq (#40)"}},{"before":"a95f7c0719c164a694c85c70c10b21eb7c9bbfe1","after":"e133531c095b54dbfa961725248f66af79535ea4","ref":"refs/heads/main","pushedAt":"2024-04-29T20:26:26.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"Model/llama3 70b instruct awq (#40)\n\n* add llama3-70b-instruct-awq\r\n\r\n* model: add llama3-70b-instruct-awq","shortMessageHtmlLink":"Model/llama3 70b instruct awq (#40)"}},{"before":"5d7f9fe8a10985f1e11e4a95e5e23e6bd49f9eb8","after":"a83dae944df285f886f4216eefa3c8ae1197fae3","ref":"refs/heads/model/llama3-70b-instruct-awq","pushedAt":"2024-04-29T20:24:18.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"model: add llama3-70b-instruct-awq","shortMessageHtmlLink":"model: add llama3-70b-instruct-awq"}},{"before":null,"after":"5d7f9fe8a10985f1e11e4a95e5e23e6bd49f9eb8","ref":"refs/heads/model/llama3-70b-instruct-awq","pushedAt":"2024-04-28T07:43:34.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"add llama3-70b-instruct-awq","shortMessageHtmlLink":"add llama3-70b-instruct-awq"}},{"before":"91421f3b1a2d2fad8cbc29b649bee5215e729bca","after":"a95f7c0719c164a694c85c70c10b21eb7c9bbfe1","ref":"refs/heads/main","pushedAt":"2024-04-25T01:36:31.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"update readme (#37)","shortMessageHtmlLink":"update readme (#37)"}},{"before":null,"after":"0d9b802ac72ead66a92efe88556102d180a22d15","ref":"refs/heads/doc/complete-readme-index","pushedAt":"2024-04-25T01:35:48.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"update readme","shortMessageHtmlLink":"update readme"}},{"before":"5d78e3a4a5f8845e6eb3f60b078253bcd500064d","after":null,"ref":"refs/heads/doc/requirements-installation","pushedAt":"2024-04-24T09:02:02.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"Sherlock113","name":"Sherlock Xu","path":"/Sherlock113","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/65327072?s=80&v=4"}},{"before":"2ac494cdc91a6a87ab57e2c22c4856d8dbba2892","after":"91421f3b1a2d2fad8cbc29b649bee5215e729bca","ref":"refs/heads/main","pushedAt":"2024-04-24T09:01:57.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"Add Llama 3 requirements installation command (#33)","shortMessageHtmlLink":"Add Llama 3 requirements installation command (#33)"}},{"before":"406c3c5ec1080964588c827cfff79dc1a84b1978","after":"2ac494cdc91a6a87ab57e2c22c4856d8dbba2892","ref":"refs/heads/main","pushedAt":"2024-04-19T21:13:24.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: docker image building (#35)","shortMessageHtmlLink":"fix: docker image building (#35)"}},{"before":null,"after":"f1f8501740b3700efe814b866b0194206b92a117","ref":"refs/heads/fix/docker-image-building","pushedAt":"2024-04-19T21:12:22.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: docker image building","shortMessageHtmlLink":"fix: docker image building"}},{"before":"9104bfe978f2109f454514f65aae65a579c9403b","after":"406c3c5ec1080964588c827cfff79dc1a84b1978","ref":"refs/heads/main","pushedAt":"2024-04-19T20:14:44.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: use local tokenzier (#34)","shortMessageHtmlLink":"fix: use local tokenzier (#34)"}},{"before":null,"after":"3597dc8f8e22222dcfcbea56cfec9a5b046807ba","ref":"refs/heads/fix/gated-tokenizer","pushedAt":"2024-04-19T20:14:21.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: use local tokenzier","shortMessageHtmlLink":"fix: use local tokenzier"}},{"before":null,"after":"5d78e3a4a5f8845e6eb3f60b078253bcd500064d","ref":"refs/heads/doc/requirements-installation","pushedAt":"2024-04-19T13:14:02.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"Sherlock113","name":"Sherlock Xu","path":"/Sherlock113","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/65327072?s=80&v=4"},"commit":{"message":"Add Llama 3 requirements installation command","shortMessageHtmlLink":"Add Llama 3 requirements installation command"}},{"before":"5d1950837c5526e05f97b476f8b0e1caa2678993","after":"9104bfe978f2109f454514f65aae65a579c9403b","ref":"refs/heads/main","pushedAt":"2024-04-19T09:14:23.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: minor llama3 readme typo (#32)","shortMessageHtmlLink":"fix: minor llama3 readme typo (#32)"}},{"before":null,"after":"d3ac572c38a2dcae5fd062fae3f1e76915e50ec1","ref":"refs/heads/fix/llama3-readme","pushedAt":"2024-04-19T09:13:56.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"fix: minor llama3 readme typo","shortMessageHtmlLink":"fix: minor llama3 readme typo"}},{"before":"01ad738216d6913491e1330a75541c529af2b49b","after":"5d1950837c5526e05f97b476f8b0e1caa2678993","ref":"refs/heads/main","pushedAt":"2024-04-19T09:06:28.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"larme","name":"Zhao Shenyang","path":"/larme","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/66847?s=80&v=4"},"commit":{"message":"add llama3-8b-instruct example (#31)","shortMessageHtmlLink":"add llama3-8b-instruct example (#31)"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEWU7aBwA","startCursor":null,"endCursor":null}},"title":"Activity ยท bentoml/BentoVLLM"}