{"payload":{"pageCount":4,"repositories":[{"type":"Public","name":"pytorch-image-models","owner":"huggingface","isFork":false,"description":"The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNetV4, MobileNet-V3 & V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more","allTopics":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2","mobile-deep-learning","mobilenetv3","efficientnet","augmix","randaugment","nfnets","normalization-free-training","vision-transformer-models","convnext","maxvit"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":25,"issueCount":54,"starsCount":30323,"forksCount":4608,"license":"Apache License 2.0","participation":[7,1,0,2,0,1,10,17,10,15,26,20,0,2,3,8,4,4,19,3,15,2,5,23,1,4,7,1,5,2,5,3,9,2,7,16,4,0,2,4,12,0,2,18,1,3,18,9,26,16,6,19],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-09T05:01:53.102Z"}},{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":243,"issueCount":863,"starsCount":127299,"forksCount":25227,"license":"Apache License 2.0","participation":[72,67,68,29,48,71,79,48,67,66,71,62,62,50,48,41,68,57,52,64,82,61,69,45,38,74,75,50,3,20,54,53,49,55,41,52,48,55,65,67,56,50,44,61,73,79,63,47,61,73,55,74],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-09T02:12:13.617Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":14,"issueCount":144,"starsCount":8193,"forksCount":910,"license":"Apache License 2.0","participation":[4,6,8,10,16,18,13,13,13,15,3,8,13,3,2,20,14,3,5,4,0,4,2,8,5,2,18,7,0,0,6,2,20,13,7,12,8,13,0,4,12,4,5,18,7,18,20,2,22,13,15,24],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T20:16:50.806Z"}},{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":122,"issueCount":348,"starsCount":23274,"forksCount":4794,"license":"Apache License 2.0","participation":[25,16,10,31,42,31,51,42,34,26,50,42,24,49,35,46,31,26,27,38,38,39,45,25,54,36,16,48,45,40,30,39,28,30,42,32,31,43,52,29,33,37,20,25,18,23,33,22,25,21,32,26],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T20:08:24.012Z"}},{"type":"Public","name":"optimum","owner":"huggingface","isFork":false,"description":"🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools","allTopics":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch","quantization","onnx"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":90,"issueCount":281,"starsCount":2244,"forksCount":390,"license":"Apache License 2.0","participation":[11,4,12,10,6,11,16,12,10,3,12,10,19,4,6,2,5,4,14,6,13,7,6,1,0,13,14,1,1,0,13,5,10,1,8,8,9,11,0,1,9,5,3,10,2,6,2,2,1,3,10,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T16:15:52.759Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: End-to-end Learning for Real-World Robotics in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":25,"issueCount":14,"starsCount":3465,"forksCount":256,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,1,3,12,37,56,58,57,64,41,73,56,10,22,20,10,10,15,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:54:18.023Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":36,"starsCount":14535,"forksCount":1363,"license":"Apache License 2.0","participation":[14,8,16,5,16,12,8,4,7,8,8,9,7,7,7,10,9,9,3,6,9,17,17,10,11,24,16,10,3,4,20,3,5,16,16,12,19,6,9,11,8,5,5,11,5,8,11,5,12,7,7,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:42:07.005Z"}},{"type":"Public","name":"datasets","owner":"huggingface","isFork":false,"description":"🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools","allTopics":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest","nlp","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":75,"issueCount":637,"starsCount":18629,"forksCount":2553,"license":"Apache License 2.0","participation":[11,6,6,5,12,7,19,6,3,7,7,6,12,5,3,6,1,6,9,10,7,2,14,7,11,6,10,16,4,1,6,1,8,10,13,1,5,19,8,5,6,2,3,9,11,2,4,5,9,3,9,14],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T14:52:19.562Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":46,"issueCount":18,"starsCount":114,"forksCount":132,"license":"Apache License 2.0","participation":[7,7,5,6,4,8,13,2,6,18,11,15,3,19,10,21,7,13,6,13,16,10,12,20,15,14,9,10,5,5,7,10,16,11,2,14,12,12,9,19,14,12,10,10,11,21,18,16,5,12,9,31],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-09T04:10:51.374Z"}},{"type":"Public","name":"huggingface_hub","owner":"huggingface","isFork":false,"description":"The official Python client for the Huggingface Hub.","allTopics":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":13,"issueCount":127,"starsCount":1778,"forksCount":456,"license":"Apache License 2.0","participation":[8,3,8,8,2,2,4,1,4,9,1,12,31,7,13,7,10,13,8,7,13,11,10,15,17,4,9,15,0,7,9,12,4,0,7,4,7,16,7,5,12,11,5,15,18,15,13,0,1,7,13,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T22:24:45.402Z"}},{"type":"Public","name":"optimum-nvidia","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":39,"starsCount":825,"forksCount":79,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,6,1,11,10,6,18,32,24,10,14,6,0,0,2,1,0,0,0,1,0,9,2,2,16,1,22,8,7,0,1,2,0,2,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T21:45:52.966Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":108,"starsCount":7207,"forksCount":841,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T16:33:14.875Z"}},{"type":"Public","name":"lm-evaluation-harness","owner":"huggingface","isFork":true,"description":"A framework for few-shot evaluation of language models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":0,"starsCount":2,"forksCount":1428,"license":"MIT License","participation":[104,106,56,97,100,70,30,73,96,145,65,60,54,66,67,13,37,21,46,7,48,23,33,40,102,50,50,30,8,7,11,17,15,14,9,8,15,16,17,9,13,6,5,2,4,5,8,14,3,12,7,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T20:09:32.797Z"}},{"type":"Public","name":"dataset-viewer","owner":"huggingface","isFork":false,"description":"Lightweight web API for visualizing and exploring any dataset - computer vision, speech, text, and tabular - stored on the Hugging Face Hub","allTopics":["nlp","data","machine-learning","api-rest","datasets","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":149,"starsCount":630,"forksCount":63,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T11:17:45.319Z"}},{"type":"Public","name":"trl","owner":"huggingface","isFork":false,"description":"Train transformer language models with reinforcement learning.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":50,"starsCount":8467,"forksCount":1023,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T09:46:23.473Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":26,"issueCount":31,"starsCount":897,"forksCount":77,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,1,0,0,0,0,0,0,0,0,0,0,31,8,4,35,97,67,78,53,113,36,25,61,32,20,14,36,45,51,47,2,17,14,3,5,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T06:00:54.062Z"}},{"type":"Public","name":"parler-tts","owner":"huggingface","isFork":false,"description":"Inference and training library for high-quality TTS models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":30,"starsCount":2733,"forksCount":281,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,29,17,11,10,4,1,2,5,59,0,8,4,4,8,11,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T08:01:47.785Z"}},{"type":"Public","name":"tgi-gaudi","owner":"huggingface","isFork":true,"description":"Large Language Model Text Generation Inference on Habana Gaudi","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":3,"starsCount":17,"forksCount":910,"license":"Apache License 2.0","participation":[4,6,8,10,16,18,13,13,13,15,3,8,13,3,2,20,14,3,5,4,0,4,2,8,5,5,20,9,0,1,9,10,21,13,7,12,20,29,1,8,16,9,9,23,7,4,1,6,3,2,3,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T15:19:05.520Z"}},{"type":"Public","name":"optimum-quanto","owner":"huggingface","isFork":false,"description":"A pytorch quantization backend for optimum","allTopics":["pytorch","quantization","optimum"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":8,"starsCount":616,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,26,13,16,37,0,0,4,7,19,56,39,25,0,10,2,11,5,13,6,6,40,11,13,25,4,11,8,10,13,1,4,3,9,7,13,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T11:19:13.740Z"}},{"type":"Public","name":"jat","owner":"huggingface","isFork":false,"description":"Distributed online training of a general multi-task Deep RL Agent","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":134,"forksCount":7,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T08:30:31.513Z"}},{"type":"Public","name":"evaluate","owner":"huggingface","isFork":false,"description":"🤗 Evaluate: A library for easily evaluating machine learning models and datasets.","allTopics":["evaluation","machine-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":52,"issueCount":140,"starsCount":1855,"forksCount":231,"license":"Apache License 2.0","participation":[0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,2,1,0,1,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,0,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:17:27.597Z"}},{"type":"Public","name":"distil-whisper","owner":"huggingface","isFork":false,"description":"Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.","allTopics":["audio","speech-recognition","whisper"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":54,"starsCount":3286,"forksCount":231,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T14:41:38.521Z"}},{"type":"Public","name":"autotrain-advanced","owner":"huggingface","isFork":false,"description":"🤗 AutoTrain Advanced","allTopics":["natural-language-processing","natural-language-understanding","huggingface","autotrain","python","machine-learning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":20,"starsCount":3529,"forksCount":421,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T13:57:01.916Z"}},{"type":"Public","name":"doc-builder","owner":"huggingface","isFork":false,"description":"The package used to build the documentation of our Hugging Face repos","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":45,"starsCount":72,"forksCount":30,"license":"Apache License 2.0","participation":[0,0,0,2,0,0,0,0,2,0,2,1,0,2,3,22,1,1,3,4,5,0,6,6,4,1,0,0,0,0,0,1,2,11,0,0,0,0,1,1,0,1,1,4,0,1,2,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T12:58:19.068Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":1,"starsCount":35,"forksCount":7,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,2,0,2,1,2,3,10,2,3,8,0,2,2,2,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T12:34:49.943Z"}},{"type":"Public","name":"controlnet_aux","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":37,"starsCount":326,"forksCount":74,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T18:58:42.923Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":20,"starsCount":1627,"forksCount":95,"license":"Apache License 2.0","participation":[13,19,13,41,11,5,11,8,9,0,3,4,4,2,0,0,3,1,19,0,11,3,15,30,7,10,3,11,1,0,1,8,6,10,10,2,11,6,8,1,9,0,3,3,10,6,4,8,4,10,6,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T08:30:57.989Z"}},{"type":"Public","name":"lighteval","owner":"huggingface","isFork":false,"description":"LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.","allTopics":["evaluation","evaluation-metrics","evaluation-framework","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":44,"starsCount":438,"forksCount":50,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,20,7,16,6,7,12,15,0,12,7,3,4,4,1,8,1,2,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T12:34:46.280Z"}},{"type":"Public","name":"cosmopedia","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":5,"starsCount":280,"forksCount":24,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T12:34:34.673Z"}},{"type":"Public","name":"dataspeech","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":8,"starsCount":207,"forksCount":23,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T10:09:25.640Z"}}],"repositoryCount":111,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}