{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"transformers_zamba","owner":"Zyphra","isFork":true,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":25107,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T21:08:01.777Z"}},{"type":"Public","name":"Megatron-LM","owner":"Zyphra","isFork":true,"description":"Ongoing research training transformer models at scale","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":10,"starsCount":0,"forksCount":2001,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:45:17.646Z"}},{"type":"Public","name":"Zamba-torch","owner":"Zyphra","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T18:36:01.768Z"}},{"type":"Public","name":"alignment-handbook","owner":"Zyphra","isFork":true,"description":"Robust recipes to align language models with human and AI preferences","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":336,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T23:16:03.798Z"}},{"type":"Public","name":"Megatron-DeepSpeed","owner":"Zyphra","isFork":true,"description":"Ongoing research training transformer language models at scale, including: BERT & GPT-2","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":0,"forksCount":2001,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T09:56:00.199Z"}},{"type":"Public","name":"mamba","owner":"Zyphra","isFork":true,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":796,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,13,0,12,0,2,2,0,7,3,0,0,3,7,1,2,2,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-20T23:57:04.431Z"}},{"type":"Public","name":"BlackMamba","owner":"Zyphra","isFork":false,"description":"Code repository for Black Mamba","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":202,"forksCount":12,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-08T19:50:41.435Z"}},{"type":"Public","name":"tokenizers","owner":"Zyphra","isFork":true,"description":"💥 Fast State-of-the-Art Tokenizers optimized for Research and Production","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":1,"issueCount":0,"starsCount":1,"forksCount":736,"license":"Apache License 2.0","participation":[1,5,0,1,0,2,3,3,2,3,4,2,5,31,3,0,1,1,3,0,1,1,1,2,2,0,1,1,1,0,0,2,2,4,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-03T23:55:41.140Z"}},{"type":"Public","name":"transformers","owner":"Zyphra","isFork":true,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":25107,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-25T18:47:55.691Z"}},{"type":"Public","name":"DeepSpeed","owner":"Zyphra","isFork":true,"description":"DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":3904,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-02T21:18:22.926Z"}},{"type":"Public","name":"apex","owner":"Zyphra","isFork":true,"description":"A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1342,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-01T19:27:42.105Z"}}],"repositoryCount":11,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}