{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":672934950,"defaultBranch":"main","name":"models","ownerLogin":"tracel-ai","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-07-31T13:50:14.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/111992358?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1715084519.0","currentOid":""},"activityList":{"items":[{"before":"553f3dd549fc778fcee8b38af447c1d7f8efc3a8","after":"46f91eeab675dbd155126cae96a1f78496eb98b9","ref":"refs/heads/llama","pushedAt":"2024-05-29T14:54:39.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Update readme and change rev to fix zip version","shortMessageHtmlLink":"Update readme and change rev to fix zip version"}},{"before":"93f9eccf7cdfb0ea15276620e5c46c0bc896d1e8","after":"553f3dd549fc778fcee8b38af447c1d7f8efc3a8","ref":"refs/heads/llama","pushedAt":"2024-05-29T14:03:17.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Pin burn git rev and add tch/wgpu feature flags","shortMessageHtmlLink":"Pin burn git rev and add tch/wgpu feature flags"}},{"before":"7946acd513b58f564b7a6352c1e51b6aead6f80f","after":"93f9eccf7cdfb0ea15276620e5c46c0bc896d1e8","ref":"refs/heads/llama","pushedAt":"2024-05-28T18:29:01.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Fix typo","shortMessageHtmlLink":"Fix typo"}},{"before":"7fa84c0ec0d6ed9718ae4b9e0f1c8a44a93a44f0","after":"7946acd513b58f564b7a6352c1e51b6aead6f80f","ref":"refs/heads/llama","pushedAt":"2024-05-28T18:24:30.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Link to models root not readme file","shortMessageHtmlLink":"Link to models root not readme file"}},{"before":"8c9d9e22328249cdbb9a58cbc874debba65d2d75","after":"7fa84c0ec0d6ed9718ae4b9e0f1c8a44a93a44f0","ref":"refs/heads/llama","pushedAt":"2024-05-28T18:22:11.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add llama to repo README","shortMessageHtmlLink":"Add llama to repo README"}},{"before":"59061d1098e2eaee37d61a3700a95aacc36d6065","after":"8c9d9e22328249cdbb9a58cbc874debba65d2d75","ref":"refs/heads/llama","pushedAt":"2024-05-28T17:50:41.000Z","pushType":"push","commitsCount":6,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add llama burn generated image","shortMessageHtmlLink":"Add llama burn generated image"}},{"before":"9492e202323f3e8632cef54abbfdf3c7b37c400a","after":"59061d1098e2eaee37d61a3700a95aacc36d6065","ref":"refs/heads/llama","pushedAt":"2024-05-27T15:00:05.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Switch to f16","shortMessageHtmlLink":"Switch to f16"}},{"before":"e2f060fa816ba544db91251714eace59e5fa9f9c","after":"be4855640b262866838eba333b8a258d4f98855a","ref":"refs/heads/main","pushedAt":"2024-05-27T12:50:30.000Z","pushType":"pr_merge","commitsCount":3,"pusher":{"login":"nathanielsimard","name":"Nathaniel Simard","path":"/nathanielsimard","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14095719?s=80&v=4"},"commit":{"message":"Merge pull request #34 from seurimas/main\n\nAdd LM head and masked fill_mask for bert-burn.","shortMessageHtmlLink":"Merge pull request #34 from seurimas/main"}},{"before":"67740e390a196087f2f4904059d25cca705fe830","after":"9492e202323f3e8632cef54abbfdf3c7b37c400a","ref":"refs/heads/llama","pushedAt":"2024-05-24T19:12:31.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Update README and features deps","shortMessageHtmlLink":"Update README and features deps"}},{"before":"edbf3a62801f648067b50d413f546d4e266565c0","after":"67740e390a196087f2f4904059d25cca705fe830","ref":"refs/heads/llama","pushedAt":"2024-05-24T18:40:53.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Fix TinyLlama query/key weights and add chat mode","shortMessageHtmlLink":"Fix TinyLlama query/key weights and add chat mode"}},{"before":"fb573887a29e6135bba726e34b0ae443a513e855","after":"edbf3a62801f648067b50d413f546d4e266565c0","ref":"refs/heads/llama","pushedAt":"2024-05-23T12:53:28.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Use sentencepiece from hf tokenizers","shortMessageHtmlLink":"Use sentencepiece from hf tokenizers"}},{"before":"ed1277c5e3d749db6ffb2fdfc29787d0d308e8a1","after":"fb573887a29e6135bba726e34b0ae443a513e855","ref":"refs/heads/llama","pushedAt":"2024-05-22T19:46:23.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add pretrained model/tokenizer download","shortMessageHtmlLink":"Add pretrained model/tokenizer download"}},{"before":"dff88b4f5f3c52a0b2494eb37368ae42bc70f65d","after":"ed1277c5e3d749db6ffb2fdfc29787d0d308e8a1","ref":"refs/heads/llama","pushedAt":"2024-05-21T15:36:09.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add default prompt and replace sentencepiece newline tokens","shortMessageHtmlLink":"Add default prompt and replace sentencepiece newline tokens"}},{"before":"7623cf47e242fdfb8f128c0db4113ab6e85f3537","after":"dff88b4f5f3c52a0b2494eb37368ae42bc70f65d","ref":"refs/heads/llama","pushedAt":"2024-05-15T20:11:37.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add tiny llama and llama3 feature flags","shortMessageHtmlLink":"Add tiny llama and llama3 feature flags"}},{"before":"14ae73796afa72a49b710b4b72a3e3274034d092","after":"e2f060fa816ba544db91251714eace59e5fa9f9c","ref":"refs/heads/main","pushedAt":"2024-05-15T13:10:41.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"[Squeezenet] Upgrade to burn 0.13.0 (#33)\n\n* [Squeezenet] Upgrade to burn 0.13.0\r\n\r\n* Upgrade to burn 0.13.0\r\n* Add feature flag to disable copy the weights file\r\n\r\nSigned-off-by: vincent \r\n\r\n* [CI] Update the minimum Rust version to 1.75\r\n\r\nSigned-off-by: vincent \r\n\r\n---------\r\n\r\nSigned-off-by: vincent ","shortMessageHtmlLink":"[Squeezenet] Upgrade to burn 0.13.0 (#33)"}},{"before":"f9586036c3c8ebbe12c3baf733190ef6b39226bd","after":"7623cf47e242fdfb8f128c0db4113ab6e85f3537","ref":"refs/heads/llama","pushedAt":"2024-05-14T20:06:40.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Fix top-p sampling index","shortMessageHtmlLink":"Fix top-p sampling index"}},{"before":"f73caf34600316ab2d0a4aeca4641a43e50c2267","after":"f9586036c3c8ebbe12c3baf733190ef6b39226bd","ref":"refs/heads/llama","pushedAt":"2024-05-10T19:14:43.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add load/save model","shortMessageHtmlLink":"Add load/save model"}},{"before":"d514e3ecc3f9ef83f788eb3ac50a3854d782fdbb","after":"f73caf34600316ab2d0a4aeca4641a43e50c2267","ref":"refs/heads/llama","pushedAt":"2024-05-10T19:08:29.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Use args.model to load record","shortMessageHtmlLink":"Use args.model to load record"}},{"before":"1c7a01f147cd6f0cf2b3ab70d14a7d1a63dbcc8c","after":"d514e3ecc3f9ef83f788eb3ac50a3854d782fdbb","ref":"refs/heads/llama","pushedAt":"2024-05-10T19:05:50.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Whoops, generic recorder","shortMessageHtmlLink":"Whoops, generic recorder"}},{"before":"ac0d3b8f1b310362244fcb639c4ccd0cd93abfb1","after":"1c7a01f147cd6f0cf2b3ab70d14a7d1a63dbcc8c","ref":"refs/heads/llama","pushedAt":"2024-05-10T19:03:47.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Missing elapsed","shortMessageHtmlLink":"Missing elapsed"}},{"before":"78fbc2eeb3231db9ed3561331f4751c96d13c9f0","after":"ac0d3b8f1b310362244fcb639c4ccd0cd93abfb1","ref":"refs/heads/llama","pushedAt":"2024-05-10T19:01:02.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add load/save model","shortMessageHtmlLink":"Add load/save model"}},{"before":"2500ef0ca4ec2bd0af5192937978b2f9ae7ae2a4","after":"78fbc2eeb3231db9ed3561331f4751c96d13c9f0","ref":"refs/heads/llama","pushedAt":"2024-05-10T13:34:47.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Fix sampling sort dim","shortMessageHtmlLink":"Fix sampling sort dim"}},{"before":"3fa0fbcf9a40aee8081ea54c0bf335bb3a98bd81","after":"2500ef0ca4ec2bd0af5192937978b2f9ae7ae2a4","ref":"refs/heads/llama","pushedAt":"2024-05-10T13:09:36.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Move RotaryEncoding outside of Transformer block so pytorch checkpoints can be loaded w/o issue","shortMessageHtmlLink":"Move RotaryEncoding outside of Transformer block so pytorch checkpoin…"}},{"before":"00d5dfab14a2fea2b81dfece1226d8ea63b6aaf2","after":"14ae73796afa72a49b710b4b72a3e3274034d092","ref":"refs/heads/main","pushedAt":"2024-05-08T11:49:50.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"[Bert] Feature: Custom Model Outputs (#31)\n\n* [Bert] Feature: Custom Model Outputs\r\n\r\n* Make the with_pooling_layer flag optional, since it isn't present in the original model","shortMessageHtmlLink":"[Bert] Feature: Custom Model Outputs (#31)"}},{"before":"ac9a0320fa4191de1003e745858359c1a4e3888a","after":"3fa0fbcf9a40aee8081ea54c0bf335bb3a98bd81","ref":"refs/heads/llama","pushedAt":"2024-05-07T19:37:46.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Fix RoPE dim and mask cache_seq_len","shortMessageHtmlLink":"Fix RoPE dim and mask cache_seq_len"}},{"before":"f689d8eb23836a140aeaf5e351774fcb6d679499","after":"00d5dfab14a2fea2b81dfece1226d8ea63b6aaf2","ref":"refs/heads/main","pushedAt":"2024-05-07T17:40:39.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"antimora","name":"Dilshod Tadjibaev","path":"/antimora","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/939125?s=80&v=4"},"commit":{"message":"update readme for whisper-live (#32)","shortMessageHtmlLink":"update readme for whisper-live (#32)"}},{"before":"fda9ccf1677c1670884970999587ac72c8bd5549","after":null,"ref":"refs/heads/resnet/fine-tune","pushedAt":"2024-05-07T12:21:59.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"nathanielsimard","name":"Nathaniel Simard","path":"/nathanielsimard","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14095719?s=80&v=4"}},{"before":"a00bee052c7a6c50f8aed6ea7ee3e7cb24dd82f5","after":"f689d8eb23836a140aeaf5e351774fcb6d679499","ref":"refs/heads/main","pushedAt":"2024-05-07T12:21:56.000Z","pushType":"pr_merge","commitsCount":4,"pusher":{"login":"nathanielsimard","name":"Nathaniel Simard","path":"/nathanielsimard","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/14095719?s=80&v=4"},"commit":{"message":"Merge pull request #29 from tracel-ai/resnet/fine-tune\n\n[ResNet] Add fine-tuning example","shortMessageHtmlLink":"Merge pull request #29 from tracel-ai/resnet/fine-tune"}},{"before":null,"after":"ac9a0320fa4191de1003e745858359c1a4e3888a","ref":"refs/heads/llama","pushedAt":"2024-05-03T18:43:20.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Add Llama-3-8b","shortMessageHtmlLink":"Add Llama-3-8b"}},{"before":"5a49503c82d0ef17f6502de69f6c62b4a06d7154","after":"fda9ccf1677c1670884970999587ac72c8bd5549","ref":"refs/heads/resnet/fine-tune","pushedAt":"2024-04-29T13:18:13.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"laggui","name":"Guillaume Lagrange","path":"/laggui","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/7225623?s=80&v=4"},"commit":{"message":"Fix src link and finetune performance statement","shortMessageHtmlLink":"Fix src link and finetune performance statement"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEVxRuEAA","startCursor":null,"endCursor":null}},"title":"Activity · tracel-ai/models"}