{"payload":{"pageCount":8,"repositories":[{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":241,"issueCount":853,"starsCount":126992,"forksCount":25143,"license":"Apache License 2.0","participation":[73,72,67,68,29,48,71,79,48,67,66,71,62,62,50,48,41,68,57,52,64,82,61,69,45,38,74,75,50,3,20,54,53,49,55,41,52,48,55,65,67,56,50,44,61,73,79,63,47,61,73,55],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T04:46:12.883Z"}},{"type":"Public","name":"trl","owner":"huggingface","isFork":false,"description":"Train transformer language models with reinforcement learning.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":58,"starsCount":8412,"forksCount":1013,"license":"Apache License 2.0","participation":[19,6,14,2,11,7,9,13,6,6,7,10,14,8,21,10,5,7,10,8,7,14,13,5,6,8,9,9,17,8,4,23,12,9,15,1,8,5,11,4,7,19,8,7,16,12,9,11,3,3,11,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T02:13:33.971Z"}},{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":113,"issueCount":341,"starsCount":23168,"forksCount":4773,"license":"Apache License 2.0","participation":[37,25,16,10,31,42,31,51,42,34,26,50,42,24,49,35,46,31,26,27,38,38,39,45,25,54,36,16,48,45,40,30,39,28,30,42,32,31,43,52,29,33,37,20,25,18,23,33,22,25,21,32],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T01:15:59.262Z"}},{"type":"Public","name":"huggingface.js","owner":"huggingface","isFork":false,"description":"Utilities to use the Hugging Face Hub API","allTopics":["machine-learning","inference","hub","api-client","huggingface"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":25,"issueCount":88,"starsCount":1244,"forksCount":145,"license":"MIT License","participation":[7,0,2,5,1,0,8,6,0,0,0,1,1,0,0,0,0,9,12,0,11,6,11,6,28,24,17,14,17,0,1,2,12,7,2,8,7,22,13,23,23,13,16,13,25,11,6,18,13,27,21,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T23:10:46.597Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":19,"issueCount":12,"starsCount":3298,"forksCount":239,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,1,3,12,37,56,58,57,64,41,73,56,10,22,20,10,10,15],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T23:10:33.189Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":37,"starsCount":14420,"forksCount":1344,"license":"Apache License 2.0","participation":[5,14,8,16,5,16,12,8,4,7,8,8,9,7,7,7,10,9,9,3,6,9,17,17,10,11,24,16,10,3,4,20,3,5,16,16,12,19,6,9,11,8,5,5,11,5,8,11,5,12,7,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T20:45:58.359Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":34,"forksCount":6,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,2,0,2,1,2,3,10,2,3,8,0,2,2,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T19:54:07.559Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":19,"starsCount":1445,"forksCount":89,"license":"Apache License 2.0","participation":[0,13,19,13,41,11,5,11,8,9,0,3,4,4,2,0,0,3,1,19,0,11,3,15,30,7,10,3,11,1,0,1,8,6,10,10,2,11,6,8,1,9,0,3,3,10,6,4,8,4,10,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T16:10:54.531Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":24,"issueCount":30,"starsCount":834,"forksCount":73,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,1,0,0,0,0,0,0,0,0,0,0,31,8,4,35,97,67,78,53,113,36,25,61,32,20,14,36,45,51,47,2,17,14,3,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T15:49:22.370Z"}},{"type":"Public","name":"blog","owner":"huggingface","isFork":false,"description":"Public repo for HF blog posts","allTopics":["hacktoberfest"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":67,"issueCount":134,"starsCount":2085,"forksCount":643,"license":null,"participation":[28,33,20,9,11,7,30,19,11,13,4,31,21,19,33,16,21,11,9,10,16,10,16,3,3,7,20,12,7,4,8,4,14,11,24,7,8,31,24,6,13,23,6,18,18,29,13,8,12,17,22,16],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T15:16:44.730Z"}},{"type":"Public","name":"chat-ui","owner":"huggingface","isFork":false,"description":"Open source codebase powering the HuggingChat app","allTopics":["typescript","svelte","hacktoberfest","tailwindcss","huggingface","svelte-kit","sveltekit","llm","chatgpt"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":20,"issueCount":194,"starsCount":6519,"forksCount":911,"license":"Apache License 2.0","participation":[6,2,7,1,1,3,10,4,5,0,4,12,0,0,9,16,6,8,18,13,10,1,0,8,11,6,13,8,5,11,0,9,7,21,16,25,19,16,8,18,10,10,8,11,20,19,28,14,13,6,15,46],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T14:52:06.633Z"}},{"type":"Public","name":"optimum-amd","owner":"huggingface","isFork":false,"description":"AMD related optimizations for transformer models","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":12,"issueCount":18,"starsCount":31,"forksCount":11,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,8,0,5,0,5,1,5,4,9,0,0,0,0,1,4,3,1,2,5,3,6,4,5,7,1,0,2,1,0,0,0,0,1,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T14:40:11.304Z"}},{"type":"Public","name":"huggingface_hub","owner":"huggingface","isFork":false,"description":"The official Python client for the Huggingface Hub.","allTopics":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":14,"issueCount":129,"starsCount":1762,"forksCount":454,"license":"Apache License 2.0","participation":[4,8,3,8,8,2,2,4,1,4,9,1,12,31,7,13,7,10,13,8,7,13,11,10,15,17,4,9,15,0,7,9,12,4,0,7,4,7,16,7,5,12,11,5,15,18,15,13,0,1,7,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T12:59:34.531Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":60,"issueCount":16,"starsCount":114,"forksCount":130,"license":"Apache License 2.0","participation":[11,7,7,5,6,4,8,13,2,6,18,11,15,3,19,10,21,7,13,6,13,16,10,12,20,15,14,9,10,5,5,7,10,16,11,2,14,12,12,9,19,14,12,10,10,11,21,18,16,5,12,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T12:16:45.929Z"}},{"type":"Public","name":"open-muse","owner":"huggingface","isFork":false,"description":"Open reproduction of MUSE for fast text2image generation. ","allTopics":["nlp","deep-learning","cv","transformer","generative-art","diffusion","text2image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":24,"starsCount":294,"forksCount":22,"license":"Apache License 2.0","participation":[0,9,4,19,5,21,28,12,0,6,1,0,14,9,18,11,2,0,0,11,4,1,0,0,2,1,0,11,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T11:47:50.030Z"}},{"type":"Public","name":"optimum-neuron","owner":"huggingface","isFork":false,"description":"Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":16,"issueCount":83,"starsCount":170,"forksCount":46,"license":"Apache License 2.0","participation":[9,7,4,6,12,4,11,2,5,4,5,12,6,8,8,7,2,4,5,2,8,3,8,12,14,1,1,4,7,0,0,3,9,10,6,6,7,5,3,4,1,4,13,8,11,4,5,3,3,6,3,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T10:09:02.568Z"}},{"type":"Public","name":"candle","owner":"huggingface","isFork":false,"description":"Minimalist ML framework for Rust","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":58,"issueCount":240,"starsCount":13984,"forksCount":770,"license":"Apache License 2.0","participation":[0,0,140,186,151,77,51,81,58,89,79,91,81,54,59,59,46,38,23,40,34,41,44,9,16,8,11,32,26,32,11,41,17,2,9,25,18,16,25,21,7,38,30,38,16,27,15,15,6,7,11,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T09:47:15.762Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":135,"starsCount":8137,"forksCount":897,"license":"Apache License 2.0","participation":[4,4,6,8,10,16,18,13,13,13,15,3,8,13,3,2,20,14,3,5,4,0,4,2,8,5,2,18,7,0,0,6,2,20,13,7,12,8,13,0,4,12,4,5,18,7,18,20,2,22,13,14],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T08:47:21.847Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":18,"issueCount":105,"starsCount":7166,"forksCount":838,"license":"Apache License 2.0","participation":[32,17,9,25,4,29,14,10,10,11,9,13,15,7,17,3,1,8,10,4,11,11,10,14,13,8,12,7,6,5,5,17,8,8,8,13,15,1,24,13,9,10,9,13,11,11,11,13,11,9,7,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-01T04:38:20.330Z"}},{"type":"Public","name":"hub-docs","owner":"huggingface","isFork":false,"description":"Docs of the Hugging Face Hub","allTopics":["machine-learning","hacktoberfest"],"primaryLanguage":null,"pullRequestCount":19,"issueCount":70,"starsCount":242,"forksCount":204,"license":"Apache License 2.0","participation":[11,9,18,8,13,8,4,3,4,4,2,4,8,8,9,9,13,19,9,15,10,23,14,11,17,21,5,5,2,0,3,7,4,7,6,7,3,2,6,2,12,4,3,4,9,1,1,3,1,5,3,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T23:01:42.169Z"}},{"type":"Public","name":"pytorch-image-models","owner":"huggingface","isFork":false,"description":"The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more","allTopics":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2","mobile-deep-learning","mobilenetv3","efficientnet","augmix","randaugment","nfnets","normalization-free-training","vision-transformer-models","convnext","maxvit"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":30,"issueCount":61,"starsCount":30247,"forksCount":4604,"license":"Apache License 2.0","participation":[2,7,1,0,2,0,1,10,17,10,15,26,20,0,2,3,8,4,4,19,3,15,2,5,23,1,4,7,1,5,2,5,3,9,2,7,16,4,0,2,4,11,0,2,18,1,3,18,8,22,6,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T22:09:58.223Z"}},{"type":"Public","name":"hf-endpoints-documentation","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":3,"issueCount":2,"starsCount":8,"forksCount":5,"license":null,"participation":[5,7,0,0,0,0,0,2,1,2,0,3,0,0,0,0,0,0,0,1,0,0,0,0,0,4,8,3,0,0,0,0,0,0,0,0,0,0,3,0,0,0,12,2,23,0,0,0,8,3,4,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T21:00:51.468Z"}},{"type":"Public","name":"datasets","owner":"huggingface","isFork":false,"description":"🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools","allTopics":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest","nlp","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":79,"issueCount":630,"starsCount":18605,"forksCount":2546,"license":"Apache License 2.0","participation":[6,11,6,6,5,12,7,19,6,3,7,7,6,12,5,3,6,1,6,9,10,7,2,14,7,11,6,10,16,4,1,6,1,8,10,13,1,5,19,8,5,6,2,3,9,11,2,4,5,9,3,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T19:04:05.483Z"}},{"type":"Public","name":"lm-evaluation-harness","owner":"huggingface","isFork":true,"description":"A framework for few-shot evaluation of language models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":0,"starsCount":2,"forksCount":1411,"license":"MIT License","participation":[56,104,106,56,97,100,70,30,73,96,145,65,60,54,66,67,13,37,21,46,7,48,23,33,40,102,50,50,30,8,7,11,17,15,14,9,8,15,16,17,9,13,6,5,2,4,5,8,14,3,12,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T17:58:53.376Z"}},{"type":"Public","name":"optimum-intel","owner":"huggingface","isFork":false,"description":"🤗 Optimum Intel: Accelerate inference with Intel optimization tools","allTopics":["optimization","intel","transformers","pruning","distillation","onnx","openvino","diffusers","inference","quantization"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":21,"issueCount":11,"starsCount":339,"forksCount":94,"license":"Apache License 2.0","participation":[11,7,2,5,11,1,6,15,18,2,2,6,1,6,2,5,6,3,4,3,2,4,10,0,0,3,3,6,1,5,4,9,10,11,13,9,9,4,5,12,14,16,15,2,12,11,25,15,10,17,4,10],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T17:01:35.144Z"}},{"type":"Public","name":"ratchet","owner":"huggingface","isFork":false,"description":"A cross-platform browser ML framework.","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":5,"issueCount":24,"starsCount":382,"forksCount":18,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,69,86,83,46,50,76,42,37,38,57,27,37,78,36,42,16,6,6,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T17:01:25.357Z"}},{"type":"Public","name":"lighteval","owner":"huggingface","isFork":false,"description":"LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.","allTopics":["evaluation","evaluation-metrics","evaluation-framework","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":42,"starsCount":391,"forksCount":48,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,20,7,16,6,7,12,15,0,12,7,3,4,4,1,8,1,2,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T16:07:18.552Z"}},{"type":"Public","name":"optimum-quanto","owner":"huggingface","isFork":false,"description":"A pytorch quantization backend for optimum","allTopics":["pytorch","quantization","optimum"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":11,"starsCount":615,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,26,13,16,37,0,0,4,7,19,56,39,25,0,10,2,11,5,13,6,6,40,11,13,25,4,11,8,10,13,1,4,3,9,7,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T15:24:50.198Z"}},{"type":"Public","name":"data-is-better-together","owner":"huggingface","isFork":false,"description":"Let's build better datasets, together!","allTopics":["community","machine-learning","datasets","human-feedback"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":1,"issueCount":2,"starsCount":160,"forksCount":26,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,41,18,6,4,5,0,0,11,1,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T15:03:01.548Z"}},{"type":"Public","name":"tgi-gaudi","owner":"huggingface","isFork":true,"description":"Large Language Model Text Generation Inference on Habana Gaudi","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":2,"starsCount":17,"forksCount":897,"license":"Apache License 2.0","participation":[4,4,6,8,10,16,18,13,13,13,15,3,8,13,3,2,20,14,3,5,4,0,4,2,8,5,5,20,9,0,1,9,10,21,13,7,12,20,29,1,8,16,9,9,23,0,4,1,6,3,2,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T13:44:19.319Z"}}],"repositoryCount":226,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}