{"payload":{"pageCount":8,"repositories":[{"type":"Public","name":"chat-ui","owner":"huggingface","isFork":false,"description":"Open source codebase powering the HuggingChat app","allTopics":["typescript","svelte","hacktoberfest","tailwindcss","huggingface","svelte-kit","sveltekit","llm","chatgpt"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":20,"issueCount":195,"starsCount":6576,"forksCount":920,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T21:52:14.811Z"}},{"type":"Public","name":"text-generation-inference","owner":"huggingface","isFork":false,"description":"Large Language Model Text Generation Inference","allTopics":["nlp","bloom","deep-learning","inference","pytorch","falcon","transformer","gpt","starcoder"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":14,"issueCount":147,"starsCount":8192,"forksCount":910,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T20:16:50.806Z"}},{"type":"Public","name":"lm-evaluation-harness","owner":"huggingface","isFork":true,"description":"A framework for few-shot evaluation of language models.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":0,"starsCount":2,"forksCount":1428,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T20:09:32.797Z"}},{"type":"Public","name":"diffusers","owner":"huggingface","isFork":false,"description":"🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.","allTopics":["deep-learning","pytorch","image-generation","flax","hacktoberfest","diffusion","text2image","image2image","jax","score-based-generative-modeling","stable-diffusion","stable-diffusion-diffusers","latent-diffusion-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":122,"issueCount":346,"starsCount":23269,"forksCount":4793,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T20:08:24.012Z"}},{"type":"Public","name":"tokenizers","owner":"huggingface","isFork":false,"description":"💥 Fast State-of-the-Art Tokenizers optimized for Research and Production","allTopics":["nlp","natural-language-processing","transformers","gpt","language-model","bert","natural-language-understanding"],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":7,"issueCount":37,"starsCount":8585,"forksCount":738,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T18:39:09.044Z"}},{"type":"Public","name":"transformers","owner":"huggingface","isFork":false,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":["python","seq2seq","flax","language-models","nlp-library","hacktoberfest","jax","pytorch-transformers","model-hub","nlp","machine-learning","natural-language-processing","deep-learning","tensorflow","pytorch","transformer","speech-recognition","pretrained-models","language-model","bert"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":242,"issueCount":863,"starsCount":127297,"forksCount":25228,"license":"Apache License 2.0","participation":[72,67,68,29,48,71,79,48,67,66,71,62,62,50,48,41,68,57,52,64,82,61,69,45,38,74,75,50,3,20,54,53,49,55,41,52,48,55,65,67,56,50,44,61,73,79,63,47,61,73,55,74],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T17:51:53.973Z"}},{"type":"Public","name":"optimum-habana","owner":"huggingface","isFork":false,"description":"Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)","allTopics":["transformers","bert","fine-tuning","hpu","habana"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":45,"issueCount":18,"starsCount":114,"forksCount":132,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T17:25:10.073Z"}},{"type":"Public","name":"optimum","owner":"huggingface","isFork":false,"description":"🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools","allTopics":["training","optimization","intel","transformers","tflite","onnxruntime","graphcore","habana","inference","pytorch","quantization","onnx"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":90,"issueCount":281,"starsCount":2244,"forksCount":390,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T16:15:52.759Z"}},{"type":"Public","name":"lerobot","owner":"huggingface","isFork":false,"description":"🤗 LeRobot: End-to-end Learning for Real-World Robotics in Pytorch","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":25,"issueCount":15,"starsCount":3437,"forksCount":255,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:54:18.023Z"}},{"type":"Public","name":"peft","owner":"huggingface","isFork":false,"description":"🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.","allTopics":["python","adapter","transformers","pytorch","lora","diffusion","parameter-efficient-learning","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":11,"issueCount":35,"starsCount":14533,"forksCount":1363,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:42:07.005Z"}},{"type":"Public","name":"evaluate","owner":"huggingface","isFork":false,"description":"🤗 Evaluate: A library for easily evaluating machine learning models and datasets.","allTopics":["evaluation","machine-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":52,"issueCount":140,"starsCount":1855,"forksCount":231,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:17:27.597Z"}},{"type":"Public","name":"text-embeddings-inference","owner":"huggingface","isFork":false,"description":"A blazing fast inference solution for text embeddings models","allTopics":["ai","ml","embeddings","huggingface","llm"],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":7,"issueCount":66,"starsCount":2146,"forksCount":124,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T15:00:52.059Z"}},{"type":"Public","name":"datasets","owner":"huggingface","isFork":false,"description":"🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools","allTopics":["machine-learning","natural-language-processing","computer-vision","deep-learning","tensorflow","numpy","speech","pandas","datasets","hacktoberfest","nlp","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":75,"issueCount":636,"starsCount":18628,"forksCount":2552,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T14:52:19.562Z"}},{"type":"Public","name":"llm.nvim","owner":"huggingface","isFork":false,"description":"LLM powered development for Neovim","allTopics":["lua","neovim","neovim-plugin"],"primaryLanguage":{"name":"Lua","color":"#000080"},"pullRequestCount":2,"issueCount":11,"starsCount":590,"forksCount":39,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T11:04:49.755Z"}},{"type":"Public","name":"optimum-intel","owner":"huggingface","isFork":false,"description":"🤗 Optimum Intel: Accelerate inference with Intel optimization tools","allTopics":["optimization","intel","transformers","pruning","distillation","onnx","openvino","diffusers","inference","quantization"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":17,"issueCount":10,"starsCount":342,"forksCount":95,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T10:21:21.036Z"}},{"type":"Public","name":"datatrove","owner":"huggingface","isFork":false,"description":"Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":20,"starsCount":1627,"forksCount":95,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T08:30:57.989Z"}},{"type":"Public","name":"pytorch-image-models","owner":"huggingface","isFork":false,"description":"The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (ViT), MobileNet-V3/V2, RegNet, DPN, CSPNet, Swin Transformer, MaxViT, CoAtNet, ConvNeXt, and more","allTopics":["pytorch","imagenet","image-classification","resnet","pretrained-models","mixnet","pretrained-weights","distributed-training","dual-path-networks","mobilenet-v2","mobile-deep-learning","mobilenetv3","efficientnet","augmix","randaugment","nfnets","normalization-free-training","vision-transformer-models","convnext","maxvit"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":25,"issueCount":54,"starsCount":30319,"forksCount":4608,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T07:00:47.102Z"}},{"type":"Public","name":"nanotron","owner":"huggingface","isFork":false,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":26,"issueCount":31,"starsCount":897,"forksCount":77,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T06:00:54.062Z"}},{"type":"Public","name":"ratchet","owner":"huggingface","isFork":false,"description":"A cross-platform browser ML framework.","allTopics":[],"primaryLanguage":{"name":"Rust","color":"#dea584"},"pullRequestCount":4,"issueCount":23,"starsCount":386,"forksCount":19,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-08T05:51:43.669Z"}},{"type":"Public","name":"huggingface_hub","owner":"huggingface","isFork":false,"description":"The official Python client for the Huggingface Hub.","allTopics":["machine-learning","natural-language-processing","deep-learning","models","pytorch","pretrained-models","hacktoberfest","model-hub"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":13,"issueCount":127,"starsCount":1777,"forksCount":456,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T22:24:45.402Z"}},{"type":"Public","name":"optimum-nvidia","owner":"huggingface","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":39,"starsCount":825,"forksCount":79,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T21:45:52.966Z"}},{"type":"Public","name":"huggingface.js","owner":"huggingface","isFork":false,"description":"Utilities to use the Hugging Face Hub API","allTopics":["machine-learning","inference","hub","api-client","huggingface"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":25,"issueCount":88,"starsCount":1251,"forksCount":151,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T17:32:20.145Z"}},{"type":"Public","name":"accelerate","owner":"huggingface","isFork":false,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":108,"starsCount":7207,"forksCount":841,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T16:33:14.875Z"}},{"type":"Public","name":"blog","owner":"huggingface","isFork":false,"description":"Public repo for HF blog posts","allTopics":["hacktoberfest"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":65,"issueCount":135,"starsCount":2093,"forksCount":643,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T16:29:58.435Z"}},{"type":"Public","name":"optimum-neuron","owner":"huggingface","isFork":false,"description":"Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":15,"issueCount":84,"starsCount":172,"forksCount":47,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T16:14:49.386Z"}},{"type":"Public","name":"cookbook","owner":"huggingface","isFork":false,"description":"Open-source AI cookbook","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":19,"issueCount":16,"starsCount":1415,"forksCount":184,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T15:58:06.586Z"}},{"type":"Public","name":"tgi-gaudi","owner":"huggingface","isFork":true,"description":"Large Language Model Text Generation Inference on Habana Gaudi","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":3,"starsCount":17,"forksCount":910,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T15:19:05.520Z"}},{"type":"Public","name":"doc-builder","owner":"huggingface","isFork":false,"description":"The package used to build the documentation of our Hugging Face repos","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":45,"starsCount":72,"forksCount":30,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T12:58:19.068Z"}},{"type":"Public","name":"optimum-tpu","owner":"huggingface","isFork":false,"description":"Google TPU optimizations for transformers models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":1,"starsCount":35,"forksCount":7,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T12:34:49.943Z"}},{"type":"Public","name":"notebooks","owner":"huggingface","isFork":false,"description":"Notebooks using the Hugging Face libraries 🤗","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":57,"issueCount":118,"starsCount":3381,"forksCount":1454,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T12:06:50.570Z"}}],"repositoryCount":226,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}