{"payload":{"pageCount":4,"repositories":[{"type":"Public","name":"lm-evaluation-harness","owner":"huggingface","isFork":true,"description":"A framework for few-shot evaluation of language models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":0,"starsCount":2,"forksCount":1398,"license":"MIT License","participation":[35,88,101,62,101,68,128,36,35,89,160,67,69,63,43,86,38,16,33,32,32,3,53,19,47,84,71,45,35,22,8,6,18,17,9,16,10,8,20,13,15,12,8,4,5,2,2,5,19,6,3,10],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T09:00:30.392Z"}},{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":242,"issueCount":846,"starsCount":126811,"forksCount":25103,"license":"Apache License 2.0","participation":[78,81,68,67,50,34,61,84,48,64,54,74,75,73,38,58,37,60,52,66,59,71,67,70,59,47,45,75,61,34,5,28,64,52,52,52,35,67,42,60,60,56,57,51,56,58,78,67,70,49,61,75],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:49:58.149Z"}},{"type":"Public","name":"dataset-viewer","owner":"huggingface","isFork":false,"description":"Lightweight web API for visualizing and exploring any dataset - computer vision, speech, text, and tabular - stored on the Hugging Face Hub","allTopics":["nlp","data","machine-learning","api-rest","datasets","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":146,"starsCount":627,"forksCount":62,"license":"Apache License 2.0","participation":[20,25,32,25,28,28,6,25,26,47,15,22,14,22,18,16,28,36,30,21,15,26,17,18,20,6,8,17,12,5,0,18,16,14,14,17,23,11,25,21,17,19,10,12,21,23,8,12,13,9,33,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:39:45.940Z"}},{"type":"Public","name":"huggingface_hub","owner":"huggingface","isFork":false,"description":"The official Python client for the Huggingface Hub.","allTopics":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":126,"starsCount":1748,"forksCount":451,"license":"Apache License 2.0","participation":[7,4,7,5,8,7,0,3,3,2,3,10,10,12,26,12,10,1,11,17,5,13,11,9,8,23,8,4,16,6,1,12,8,8,3,5,2,7,9,12,8,7,12,7,16,10,18,16,6,0,4,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:28:29.396Z"}},{"type":"Public","name":"optimum","owner":"huggingface","isFork":false,"description":"🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools","allTopics":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch","quantization","onnx"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":85,"issueCount":272,"starsCount":2219,"forksCount":384,"license":"Apache License 2.0","participation":[6,7,9,7,11,11,6,12,16,7,9,8,10,15,11,6,4,4,5,10,9,7,13,6,3,1,6,7,15,1,0,7,7,5,10,5,5,10,9,7,1,5,7,2,8,7,0,7,1,2,2,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:25:22.446Z"}},{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":119,"issueCount":340,"starsCount":23093,"forksCount":4762,"license":"Apache License 2.0","participation":[35,25,21,12,19,44,31,45,37,43,28,32,54,37,33,48,41,35,32,24,29,34,42,51,36,25,52,23,36,49,31,40,45,26,24,36,45,33,35,39,36,47,29,32,14,23,23,30,21,34,18,26],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:59:51.401Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":10,"starsCount":3274,"forksCount":235,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,2,1,7,22,47,68,37,82,34,65,79,18,19,19,15,14,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:05:47.393Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":134,"starsCount":8117,"forksCount":893,"license":"Apache License 2.0","participation":[13,5,4,5,12,5,22,15,16,10,17,7,6,6,15,1,9,18,9,2,8,1,3,1,4,10,3,7,13,5,0,3,4,4,20,10,11,11,6,12,2,3,11,4,9,18,11,23,9,6,18,17],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T08:36:52.909Z"}},{"type":"Public","name":"quanto","owner":"huggingface","isFork":false,"description":"A pytorch Quantization Toolkit","allTopics":["pytorch","quantization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":9,"starsCount":613,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,29,14,18,22,0,3,8,2,39,51,33,14,4,6,3,12,6,11,6,20,31,12,21,10,13,7,3,18,6,1,6,9,4,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T07:27:28.325Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":24,"issueCount":30,"starsCount":830,"forksCount":71,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,1,0,0,0,0,0,0,0,0,0,0,8,23,9,25,34,101,89,50,98,61,27,73,21,28,19,21,34,62,43,29,12,16,6,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T03:55:34.921Z"}},{"type":"Public","name":"datasets","owner":"huggingface","isFork":false,"description":"🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools","allTopics":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest","nlp","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":80,"issueCount":625,"starsCount":18591,"forksCount":2542,"license":"Apache License 2.0","participation":[13,6,11,7,4,7,14,11,14,3,3,7,8,8,8,7,3,4,2,8,12,5,8,5,10,10,10,5,12,12,4,3,5,0,14,10,7,6,3,20,6,7,3,3,4,13,4,3,7,2,9,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T03:17:07.784Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":38,"starsCount":14333,"forksCount":1331,"license":"Apache License 2.0","participation":[15,8,13,16,5,8,17,11,7,7,2,11,8,6,10,6,11,10,9,5,4,5,12,20,11,13,17,22,12,6,4,7,15,6,12,11,16,18,12,9,11,4,9,2,10,7,4,17,5,5,11,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T23:57:27.971Z"}},{"type":"Public","name":"trl","owner":"huggingface","isFork":false,"description":"Train transformer language models with reinforcement learning.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":58,"starsCount":8384,"forksCount":1011,"license":"Apache License 2.0","participation":[9,15,6,14,4,7,9,12,10,8,3,9,8,15,12,19,10,4,6,12,10,12,8,11,5,3,12,12,3,21,4,20,8,12,13,10,0,10,11,5,4,16,13,8,7,13,19,3,8,5,1,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T21:44:21.899Z"}},{"type":"Public","name":"autotrain-advanced","owner":"huggingface","isFork":false,"description":"🤗 AutoTrain Advanced","allTopics":["natural-language-processing","natural-language-understanding","huggingface","autotrain","python","machine-learning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":18,"starsCount":3458,"forksCount":414,"license":"Apache License 2.0","participation":[8,5,7,2,0,8,4,5,29,4,15,22,7,20,2,3,2,9,4,3,16,4,5,1,5,11,13,26,40,25,3,0,0,1,4,4,9,12,11,10,22,14,17,9,8,4,12,33,14,27,13,18],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T19:50:15.600Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":52,"issueCount":15,"starsCount":113,"forksCount":129,"license":"Apache License 2.0","participation":[4,11,8,6,5,7,5,10,8,3,11,12,20,9,11,15,14,14,8,9,7,21,9,10,17,14,22,11,10,4,4,9,9,12,10,9,9,14,10,10,12,23,7,12,10,9,10,24,24,8,9,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T07:24:52.820Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":17,"starsCount":1411,"forksCount":89,"license":"Apache License 2.0","participation":[0,0,15,18,38,22,7,7,12,3,8,3,0,5,3,2,0,0,3,10,10,2,12,1,34,10,11,8,9,4,0,1,0,11,5,16,2,8,6,11,2,3,7,1,2,8,9,2,10,2,6,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T16:11:49.032Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":32,"forksCount":5,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,1,0,2,3,1,11,3,0,5,6,0,3,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T07:14:27.513Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":17,"issueCount":104,"starsCount":7154,"forksCount":833,"license":"Apache License 2.0","participation":[28,31,10,14,16,13,25,15,6,14,6,11,15,11,14,11,0,7,5,9,10,7,8,18,17,3,11,10,5,6,4,14,8,11,6,12,14,9,13,16,14,5,16,8,11,15,7,12,14,10,5,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T14:10:14.071Z"}},{"type":"Public","name":"dataspeech","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":9,"starsCount":204,"forksCount":22,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,7,0,0,0,0,0,0,28,7,0,41,4,2,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T12:00:25.403Z"}},{"type":"Public","name":"pytorch-image-models","owner":"huggingface","isFork":false,"description":"The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more","allTopics":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2","mobile-deep-learning","mobilenetv3","efficientnet","augmix","randaugment","nfnets","normalization-free-training","vision-transformer-models","convnext","maxvit"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":28,"issueCount":64,"starsCount":30211,"forksCount":4598,"license":"Apache License 2.0","participation":[3,2,7,1,1,1,0,1,12,18,11,24,19,14,0,3,3,9,4,7,16,7,10,1,18,11,4,5,2,1,5,6,2,2,9,5,16,8,0,1,1,9,6,2,7,11,4,7,12,18,15,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T05:10:59.449Z"}},{"type":"Public","name":"tgi-gaudi","owner":"huggingface","isFork":true,"description":"Large Language Model Text Generation Inference on Habana Gaudi","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":3,"starsCount":17,"forksCount":893,"license":"Apache License 2.0","participation":[13,5,4,5,12,5,22,15,16,10,17,7,6,6,15,1,9,18,9,2,8,1,3,1,4,10,4,11,13,7,1,4,8,11,20,10,11,15,18,25,3,9,14,9,15,15,1,4,6,3,2,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-28T22:51:08.437Z"}},{"type":"Public","name":"lighteval","owner":"huggingface","isFork":false,"description":"LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.","allTopics":["evaluation","evaluation-metrics","evaluation-framework","huggingface"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":41,"starsCount":387,"forksCount":47,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,20,12,16,2,12,12,9,6,8,7,3,5,2,4,5,2,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-27T11:10:53.355Z"}},{"type":"Public","name":"optimum-benchmark","owner":"huggingface","isFork":false,"description":"A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Optimum's hardware optimizations & quantization schemes.","allTopics":["benchmark","pytorch","openvino","onnxruntime","text-generation-inference","neural-compressor","tensorrt-llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":6,"starsCount":203,"forksCount":33,"license":"Apache License 2.0","participation":[27,25,14,9,10,10,14,10,5,24,70,18,49,47,28,13,9,0,2,20,6,13,6,3,5,4,12,3,0,0,1,2,5,2,0,1,1,6,9,3,1,9,5,5,3,6,2,2,4,7,18,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-26T19:04:44.621Z"}},{"type":"Public","name":"alignment-handbook","owner":"huggingface","isFork":false,"description":"Robust recipes to align language models with human and AI preferences","allTopics":["transformers","llm","rlhf"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":81,"starsCount":4005,"forksCount":334,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,5,1,0,0,0,0,4,0,2,10,0,36,4,1,3,1,0,0,0,5,2,1,2,6,0,0,0,2,1,1,2,1,1,1,0,2,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-25T20:15:27.798Z"}},{"type":"Public","name":"cosmopedia","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":4,"starsCount":258,"forksCount":22,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,34,7,0,1,0,2,0,0,1,0,2,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-25T17:55:30.417Z"}},{"type":"Public","name":"optimum-nvidia","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":39,"starsCount":793,"forksCount":76,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,2,2,10,11,10,17,37,23,10,9,2,0,2,1,0,0,0,0,1,6,5,0,2,16,1,25,11,1,0,3,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T12:46:57.176Z"}},{"type":"Public","name":"hf-rocm-benchmark","owner":"huggingface","isFork":false,"description":"A reproducible benchmark of Text Generation Inference and Transformers as of April 2024 on AMD Instinct MI250 and MI300","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,2,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T15:35:24.453Z"}},{"type":"Public","name":"parler-tts","owner":"huggingface","isFork":false,"description":"Inference and training library for high-quality TTS models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":30,"starsCount":2670,"forksCount":271,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,27,19,10,5,2,3,0,48,16,4,8,0,11,1,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T05:09:21.997Z"}},{"type":"Public","name":"jat","owner":"huggingface","isFork":false,"description":"Distributed online training of a general multi-task Deep RL Agent","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":128,"forksCount":7,"license":"Apache License 2.0","participation":[29,28,41,5,7,3,0,3,1,0,1,4,9,9,4,0,0,0,1,1,0,0,6,1,1,8,0,2,0,0,0,1,1,0,0,0,0,0,0,0,0,0,4,0,2,5,6,1,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T11:06:52.354Z"}},{"type":"Public","name":"controlnet_aux","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":37,"starsCount":323,"forksCount":70,"license":"Apache License 2.0","participation":[25,8,3,2,4,0,0,0,0,4,0,4,3,7,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,2,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T03:25:01.356Z"}}],"repositoryCount":111,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}