{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"dust","owner":"dust-tt","isFork":false,"description":"Amplify your team's potential with customizable and secure AI assistants.","allTopics":["rust","large-language-models"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":23,"issueCount":96,"starsCount":902,"forksCount":98,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-11T16:16:04.356Z"}},{"type":"Public","name":"rust-eventsource-client","owner":"dust-tt","isFork":true,"description":"Server-sent events (SSE) client implementation for Rust","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":21,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,3,0,5,0,0,0,2,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,11,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-10T12:26:26.782Z"}},{"type":"Public","name":"llama-ssp","owner":"dust-tt","isFork":false,"description":"Experiments on speculative sampling with Llama models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":112,"forksCount":6,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-08T15:13:21.115Z"}}],"repositoryCount":3,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"dust-tt repositories"}