{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":670205761,"defaultBranch":"main","name":"localGPT","ownerLogin":"TDL77","currentUserCanPush":false,"isFork":true,"isEmpty":false,"createdAt":"2023-07-24T14:20:49.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/69913933?v=4","public":true,"private":false,"isOrgOwned":false},"refInfo":{"name":"","listCacheKey":"v0:1690777982.986785","currentOid":""},"activityList":{"items":[{"before":"d4df6d06dfffdb84cad9eea97ee0a0b4ede99ae8","after":"a1dea3becb8b1ae28a87369b1636c4c4a4501c27","ref":"refs/heads/main","pushedAt":"2024-07-20T11:17:20.000Z","pushType":"push","commitsCount":22,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"Update README.md","shortMessageHtmlLink":"Update README.md"}},{"before":"0b4fa33afea2c864a5af99c1fe5c790c421553a1","after":"d4df6d06dfffdb84cad9eea97ee0a0b4ede99ae8","ref":"refs/heads/main","pushedAt":"2024-01-08T10:52:00.000Z","pushType":"push","commitsCount":4,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"Merge pull request #704 from SanelD/patch-1\n\nFix typo in README","shortMessageHtmlLink":"Merge pull request PromtEngineer#704 from SanelD/patch-1"}},{"before":"d30aaeffa79296182a1787da817b0ddcce67ec5c","after":"0b4fa33afea2c864a5af99c1fe5c790c421553a1","ref":"refs/heads/main","pushedAt":"2023-12-23T12:34:11.000Z","pushType":"push","commitsCount":4,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"Merge pull request #687 from FrothBite/issue-662-No-matching-distribution-found-for-autoawq\n\nadd Mac checks for `autoawq` and `auto-gptq`","shortMessageHtmlLink":"Merge pull request PromtEngineer#687 from FrothBite/issue-662-No-matc…"}},{"before":"15109efffcf7284aa1947a5ce9ac33021cdae423","after":"d30aaeffa79296182a1787da817b0ddcce67ec5c","ref":"refs/heads/main","pushedAt":"2023-12-03T16:09:30.000Z","pushType":"push","commitsCount":21,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"Merge pull request #646 from ptanov/645-support-awq-models\n\n#645 Add support for AWQ models","shortMessageHtmlLink":"Merge pull request PromtEngineer#646 from ptanov/645-support-awq-models"}},{"before":"15e96488b67eb4145d743f5ef4f3cc9e7102bbb7","after":"15109efffcf7284aa1947a5ce9ac33021cdae423","ref":"refs/heads/main","pushedAt":"2023-10-08T10:13:07.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"Merge pull request #549 from marook/add-llama-cpp-to-docker\n\nadd llama-cpp-python to Dockerfile","shortMessageHtmlLink":"Merge pull request PromtEngineer#549 from marook/add-llama-cpp-to-docker"}},{"before":"0d2054473c320a9c05f53e503bb55add4ea48271","after":"15e96488b67eb4145d743f5ef4f3cc9e7102bbb7","ref":"refs/heads/main","pushedAt":"2023-10-01T08:20:07.000Z","pushType":"push","commitsCount":42,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"API Update\n\n- Updated the API code to use the prompt template.\n- Removed unused code from run_local_API.py\n- Standardized the API endpoint in the localGPTUI","shortMessageHtmlLink":"API Update"}},{"before":"6f894706d9c371115d691f0d9e3b117fb42ba216","after":"0d2054473c320a9c05f53e503bb55add4ea48271","ref":"refs/heads/main","pushedAt":"2023-08-19T11:45:12.000Z","pushType":"push","commitsCount":44,"pusher":{"login":"TDL77","name":null,"path":"/TDL77","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/69913933?s=80&v=4"},"commit":{"message":"Update ingest.py\n\nChanged the chunk size to 880 from 1000, the original size seems to be causing some issues for larger files. The best way is to around with it.","shortMessageHtmlLink":"Update ingest.py"}}],"hasNextPage":false,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEhI76pAA","startCursor":null,"endCursor":null}},"title":"Activity · TDL77/localGPT"}