{"payload":{"pageCount":4,"repositories":[{"type":"Public","name":"pytorch-image-models","owner":"huggingface","isFork":false,"description":"The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more","topicNames":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2"],"topicsNotShown":10,"allTopics":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2","mobile-deep-learning","mobilenetv3","efficientnet","augmix","randaugment","nfnets","normalization-free-training","vision-transformer-models","convnext","maxvit"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":28,"issueCount":64,"starsCount":30105,"forksCount":4595,"license":"Apache License 2.0","participation":[3,3,2,7,1,1,1,0,1,12,18,11,24,19,14,0,3,3,9,4,7,16,7,10,1,18,11,4,5,2,1,5,6,2,2,9,5,16,8,0,1,1,9,6,2,7,11,4,7,12,18,15],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:34:46.383Z"}},{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","topicNames":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp"],"topicsNotShown":10,"allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":242,"issueCount":836,"starsCount":126470,"forksCount":25033,"license":"Apache License 2.0","participation":[47,78,81,68,67,50,34,61,84,48,64,54,74,75,73,38,58,37,60,52,66,59,71,67,70,59,47,45,75,61,34,5,28,64,52,52,52,35,67,42,60,60,56,57,51,56,58,78,67,70,49,61],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:28:18.216Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":18,"starsCount":1392,"forksCount":86,"license":"Apache License 2.0","participation":[0,0,0,15,18,38,22,7,7,12,3,8,3,0,5,3,2,0,0,3,10,10,2,12,1,34,10,11,8,9,4,0,1,0,11,5,16,2,8,6,11,2,3,7,1,2,8,9,2,10,2,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T19:15:54.414Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","topicNames":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"topicsNotShown":0,"allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":8,"issueCount":134,"starsCount":8083,"forksCount":884,"license":"Apache License 2.0","participation":[9,13,5,4,5,12,5,22,15,16,10,17,7,6,6,15,1,9,18,9,2,8,1,3,1,4,10,3,7,13,5,0,3,4,4,20,10,11,11,6,12,2,3,11,4,9,18,11,23,9,6,18],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:13:35.181Z"}},{"type":"Public","name":"autotrain-advanced","owner":"huggingface","isFork":false,"description":"🤗 AutoTrain Advanced","topicNames":["natural-language-processing","natural-language-understanding","huggingface","autotrain","python","machine-learning","deep-learning"],"topicsNotShown":0,"allTopics":["natural-language-processing","natural-language-understanding","huggingface","autotrain","python","machine-learning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":21,"starsCount":3415,"forksCount":410,"license":"Apache License 2.0","participation":[0,8,5,7,2,0,8,4,5,29,4,15,22,7,20,2,3,2,9,4,3,16,4,5,1,5,11,13,26,40,25,3,0,0,1,4,4,9,12,11,10,22,14,17,9,8,4,12,33,14,27,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T18:37:02.817Z"}},{"type":"Public","name":"trl","owner":"huggingface","isFork":false,"description":"Train transformer language models with reinforcement learning.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":17,"issueCount":60,"starsCount":8309,"forksCount":1000,"license":"Apache License 2.0","participation":[3,9,15,6,14,4,7,9,12,10,8,3,9,8,15,12,19,10,4,6,12,10,12,8,11,5,3,12,12,3,21,4,20,8,12,13,10,0,10,11,5,4,16,13,8,7,13,19,3,8,5,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T18:25:35.673Z"}},{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","topicNames":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling"],"topicsNotShown":3,"allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":121,"issueCount":332,"starsCount":22962,"forksCount":4737,"license":"Apache License 2.0","participation":[33,35,25,21,12,19,44,31,45,37,43,28,32,54,37,33,48,41,35,32,24,29,34,42,51,36,25,52,23,36,49,31,40,45,26,24,36,45,33,35,39,36,47,29,32,14,23,23,30,21,34,18],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:11:38.533Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":21,"issueCount":108,"starsCount":7112,"forksCount":821,"license":"Apache License 2.0","participation":[6,28,31,10,14,16,13,25,15,6,14,6,11,15,11,14,11,0,7,5,9,10,7,8,18,17,3,11,10,5,6,4,14,8,11,6,12,14,9,13,16,14,5,16,8,11,15,7,12,14,10,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T17:59:08.507Z"}},{"type":"Public","name":"controlnet_aux","owner":"huggingface","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":37,"starsCount":317,"forksCount":68,"license":"Apache License 2.0","participation":[9,25,8,3,2,4,0,0,0,0,4,0,4,3,7,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T17:42:07.356Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":19,"issueCount":11,"starsCount":3168,"forksCount":224,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,2,1,7,22,47,68,37,82,34,65,79,18,19,19,15,14],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T16:35:13.258Z"}},{"type":"Public","name":"dataset-viewer","owner":"huggingface","isFork":false,"description":"Lightweight web API for visualizing and exploring any dataset - computer vision, speech, text, and tabular - stored on the Hugging Face Hub","topicNames":["nlp","data","machine-learning","api-rest","datasets","huggingface"],"topicsNotShown":0,"allTopics":["nlp","data","machine-learning","api-rest","datasets","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":141,"starsCount":624,"forksCount":62,"license":"Apache License 2.0","participation":[13,20,25,32,25,28,28,6,25,26,47,15,22,14,22,18,16,28,36,30,21,15,26,17,18,20,6,8,17,12,5,0,18,16,14,14,17,23,11,25,21,17,19,10,12,21,23,8,12,13,9,33],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:33:52.013Z"}},{"type":"Public","name":"parler-tts","owner":"huggingface","isFork":false,"description":"Inference and training library for high-quality TTS models.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":30,"starsCount":2624,"forksCount":266,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,27,19,10,5,2,3,0,48,16,4,8,0,11,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T16:32:00.193Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","topicNames":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"topicsNotShown":0,"allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":40,"starsCount":14205,"forksCount":1319,"license":"Apache License 2.0","participation":[0,15,8,13,16,5,8,17,11,7,7,2,11,8,6,10,6,11,10,9,5,4,5,12,20,11,13,17,22,12,6,4,7,15,6,12,11,16,18,12,9,11,4,9,2,10,7,4,17,5,5,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T15:13:56.827Z"}},{"type":"Public","name":"huggingface_hub","owner":"huggingface","isFork":false,"description":"The official Python client for the Huggingface Hub.","topicNames":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"topicsNotShown":0,"allTopics":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":18,"issueCount":130,"starsCount":1728,"forksCount":443,"license":"Apache License 2.0","participation":[4,7,4,7,5,8,7,0,3,3,2,3,10,10,12,26,12,10,1,11,17,5,13,11,9,8,23,8,4,16,6,1,12,8,8,3,5,2,7,9,12,8,7,12,7,16,10,18,16,6,0,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T15:07:19.413Z"}},{"type":"Public","name":"datasets","owner":"huggingface","isFork":false,"description":"🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools","topicNames":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest"],"topicsNotShown":2,"allTopics":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest","nlp","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":81,"issueCount":622,"starsCount":18545,"forksCount":2536,"license":"Apache License 2.0","participation":[6,13,6,11,7,4,7,14,11,14,3,3,7,8,8,8,7,3,4,2,8,12,5,8,5,10,10,10,5,12,12,4,3,5,0,14,10,7,6,3,20,6,7,3,3,4,13,4,3,7,2,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T15:02:04.998Z"}},{"type":"Public","name":"lighteval","owner":"huggingface","isFork":false,"description":"LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.","topicNames":["evaluation","evaluation-metrics","evaluation-framework","huggingface"],"topicsNotShown":0,"allTopics":["evaluation","evaluation-metrics","evaluation-framework","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":41,"starsCount":376,"forksCount":47,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,20,12,16,2,12,12,9,6,8,7,3,5,2,4,5,2,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T14:05:43.292Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":17,"issueCount":29,"starsCount":821,"forksCount":70,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,1,0,0,0,0,0,0,0,0,0,0,8,23,9,25,34,101,89,50,98,61,27,73,21,28,19,21,34,62,43,29,12,16,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T13:44:07.619Z"}},{"type":"Public","name":"optimum","owner":"huggingface","isFork":false,"description":"🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools","topicNames":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch"],"topicsNotShown":2,"allTopics":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch","quantization","onnx"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":88,"issueCount":269,"starsCount":2208,"forksCount":375,"license":"Apache License 2.0","participation":[3,6,7,9,7,11,11,6,12,16,7,9,8,10,15,11,6,4,4,5,10,9,7,13,6,3,1,6,7,15,1,0,7,7,5,10,5,5,10,9,7,1,5,7,2,8,7,0,7,1,2,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T15:04:36.795Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","topicNames":["transformers","bert","fine-tuning","hpu","habana"],"topicsNotShown":0,"allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":56,"issueCount":13,"starsCount":111,"forksCount":125,"license":"Apache License 2.0","participation":[6,4,11,8,6,5,7,5,10,8,3,11,12,20,9,11,15,14,14,8,9,7,21,9,10,17,14,22,11,10,4,4,9,9,12,10,9,9,14,10,10,12,23,7,12,10,9,10,24,24,8,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T19:57:21.581Z"}},{"type":"Public","name":"lm-evaluation-harness","owner":"huggingface","isFork":true,"description":"A framework for few-shot evaluation of language models.","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":0,"starsCount":2,"forksCount":1384,"license":"MIT License","participation":[6,35,88,101,62,101,68,128,36,35,89,160,67,69,63,43,86,38,16,33,32,32,3,53,19,47,84,71,45,35,22,8,6,18,17,9,16,10,8,20,13,15,12,8,4,5,2,2,5,19,6,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:12:08.349Z"}},{"type":"Public","name":"optimum-nvidia","owner":"huggingface","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":40,"starsCount":792,"forksCount":76,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,2,2,10,11,10,17,37,23,10,9,2,0,2,1,0,0,0,0,1,6,5,0,2,16,1,25,11,1,0,3,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T21:17:31.303Z"}},{"type":"Public","name":"optimum-benchmark","owner":"huggingface","isFork":false,"description":"A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Optimum's hardware optimizations & quantization schemes.","topicNames":["benchmark","pytorch","openvino","onnxruntime","text-generation-inference","neural-compressor","tensorrt-llm"],"topicsNotShown":0,"allTopics":["benchmark","pytorch","openvino","onnxruntime","text-generation-inference","neural-compressor","tensorrt-llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":7,"starsCount":200,"forksCount":33,"license":"Apache License 2.0","participation":[13,27,25,14,9,10,10,14,10,5,24,70,18,49,47,28,13,9,0,2,20,6,13,6,3,5,4,12,3,0,0,1,2,5,2,0,1,1,6,9,3,1,9,5,5,3,6,2,2,4,7,18],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T20:45:26.621Z"}},{"type":"Public","name":"distil-whisper","owner":"huggingface","isFork":false,"description":"Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.","topicNames":["audio","speech-recognition","whisper"],"topicsNotShown":0,"allTopics":["audio","speech-recognition","whisper"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":51,"starsCount":3238,"forksCount":228,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,25,2,36,0,0,7,1,2,0,0,0,3,4,2,0,0,0,0,0,0,1,30,11,2,3,7,13,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T19:59:26.596Z"}},{"type":"Public","name":"tgi-gaudi","owner":"huggingface","isFork":true,"description":"Large Language Model Text Generation Inference on Habana Gaudi","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":1,"starsCount":16,"forksCount":884,"license":"Apache License 2.0","participation":[9,13,5,4,5,12,5,22,15,16,10,17,7,6,6,15,1,9,18,9,2,8,1,3,1,4,10,4,11,13,7,1,4,8,11,20,10,11,15,18,25,3,9,14,9,15,15,1,4,6,3,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T18:25:03.383Z"}},{"type":"Public","name":"competitions","owner":"huggingface","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":102,"forksCount":8,"license":"Apache License 2.0","participation":[1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,18,21,10,4,12,4,1,34,5,14,3,4,1,0,8,1,0,0,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T10:41:56.794Z"}},{"type":"Public","name":"quanto","owner":"huggingface","isFork":false,"description":"A pytorch Quantization Toolkit","topicNames":["pytorch","quantization"],"topicsNotShown":0,"allTopics":["pytorch","quantization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":14,"starsCount":608,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,29,14,18,22,0,3,8,2,39,51,33,14,4,6,3,12,6,11,6,20,31,12,21,10,13,7,3,16,5,1,5,8,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T16:55:59.417Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":31,"forksCount":5,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,2,3,1,11,3,0,5,6,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-20T08:30:47.560Z"}},{"type":"Public","name":"dataspeech","owner":"huggingface","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":9,"starsCount":201,"forksCount":22,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,7,0,0,0,0,0,0,28,7,0,41,4,2,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-22T18:23:08.129Z"}},{"type":"Public","name":"open_asr_leaderboard","owner":"huggingface","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":7,"starsCount":34,"forksCount":10,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,6,1,4,2,0,0,12,13,7,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,6,0,0,0,0,0,0,0,0,1,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-17T20:16:53.746Z"}},{"type":"Public","name":"diffusion-fast","owner":"huggingface","isFork":false,"description":"Faster generation with text-to-image diffusion models.","topicNames":["pytorch","diffusion-models","text-to-image-generation","diffusers","sdxl"],"topicsNotShown":0,"allTopics":["pytorch","diffusion-models","text-to-image-generation","diffusers","sdxl"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":150,"forksCount":7,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,83,4,19,40,3,1,0,0,3,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-16T03:11:19.988Z"}}],"repositoryCount":110,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}