{"payload":{"pageCount":4,"repositories":[{"type":"Public","name":"ChatRTX","owner":"NVIDIA","isFork":false,"description":"A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":8,"issueCount":19,"starsCount":2453,"forksCount":269,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T03:38:57.299Z"}},{"type":"Public","name":"numbast","owner":"NVIDIA","isFork":false,"description":"Numbast is a tool to build an automated pipeline that converts CUDA APIs into Numba bindings.","allTopics":["cuda","numba"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":17,"starsCount":13,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,15,0,6,1,8,0,4,0,2,1,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T03:32:20.890Z"}},{"type":"Public","name":"NeMo","owner":"NVIDIA","isFork":false,"description":"A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)","allTopics":["machine-translation","tts","speech-synthesis","neural-networks","deeplearning","speaker-recognition","asr","multimodal","speech-translation","large-language-models","speaker-diariazation","generative-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":89,"issueCount":54,"starsCount":10364,"forksCount":2216,"license":"Apache License 2.0","participation":[14,12,22,10,17,21,11,17,14,22,23,17,21,23,16,21,31,25,13,24,6,12,4,13,4,6,20,14,4,8,15,17,28,13,13,21,22,39,24,20,41,19,13,34,30,27,28,38,28,30,20,28],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:54:20.868Z"}},{"type":"Public","name":"cloudai","owner":"NVIDIA","isFork":false,"description":"CloudAI Benchmark Framework","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":0,"starsCount":16,"forksCount":9,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,22,56,32,69],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T01:44:40.984Z"}},{"type":"Public","name":"warp","owner":"NVIDIA","isFork":false,"description":"A Python framework for high performance GPU simulation and graphics","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":48,"starsCount":1753,"forksCount":150,"license":"Other","participation":[25,26,57,38,10,17,41,9,5,32,24,20,41,7,4,18,29,14,28,32,63,27,32,23,1,26,23,24,0,2,19,39,29,12,6,10,21,48,50,45,40,14,9,14,20,37,35,40,26,21,35,33],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:50:27.828Z"}},{"type":"Public","name":"NeMo-Curator","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for data curation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":28,"starsCount":291,"forksCount":30,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,4,3,2,0,6,1,6,3,4,2,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:27:35.697Z"}},{"type":"Public","name":"Megatron-LM","owner":"NVIDIA","isFork":false,"description":"Ongoing research training transformer models at scale","allTopics":["transformers","model-para","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":128,"issueCount":311,"starsCount":8980,"forksCount":2024,"license":"Other","participation":[12,4,22,14,59,28,38,27,35,51,61,59,32,70,34,36,58,133,101,131,62,33,41,76,24,41,36,59,10,6,39,31,56,61,42,19,5,18,23,41,60,92,44,50,63,30,17,35,25,21,14,31],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:06:11.886Z"}},{"type":"Public","name":"NeMo-speech-data-processor","owner":"NVIDIA","isFork":false,"description":"A toolkit for processing speech data and creating speech datasets","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":1,"starsCount":63,"forksCount":20,"license":"Apache License 2.0","participation":[3,0,2,0,0,17,7,14,0,0,0,0,0,0,0,0,0,0,0,1,1,7,13,3,0,1,0,0,0,0,1,0,0,0,0,0,0,1,0,0,0,5,0,0,1,0,1,1,3,1,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:01:18.903Z"}},{"type":"Public","name":"NeMo-Framework-Launcher","owner":"NVIDIA","isFork":false,"description":"NeMo Megatron launcher and tools","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":23,"starsCount":403,"forksCount":118,"license":"Apache License 2.0","participation":[15,8,3,4,4,11,12,12,14,14,1,12,13,27,8,1,8,10,27,13,7,6,11,17,10,24,36,23,11,12,23,22,27,19,16,55,25,33,35,17,34,5,6,23,11,29,13,22,54,34,1,19],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:44:19.457Z"}},{"type":"Public","name":"modulus","owner":"NVIDIA","isFork":false,"description":"Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods","allTopics":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":17,"issueCount":95,"starsCount":733,"forksCount":153,"license":"Apache License 2.0","participation":[8,1,4,4,2,2,9,9,11,5,1,2,1,3,3,11,7,4,3,4,1,10,5,10,5,10,5,4,0,4,4,5,7,6,5,3,7,7,8,2,3,2,4,10,9,3,5,9,2,3,4,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:32:40.998Z"}},{"type":"Public","name":"NVFlare","owner":"NVIDIA","isFork":false,"description":"NVIDIA Federated Learning Application Runtime Environment","allTopics":["python"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":24,"starsCount":556,"forksCount":150,"license":"Apache License 2.0","participation":[9,2,7,1,16,6,11,12,13,9,14,12,10,17,19,17,12,5,6,11,16,16,9,9,9,10,9,24,5,3,17,7,9,7,11,2,3,4,2,5,9,4,10,8,7,14,9,8,8,5,9,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:11:57.574Z"}},{"type":"Public","name":"TransformerEngine","owner":"NVIDIA","isFork":false,"description":"A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.","allTopics":["python","machine-learning","deep-learning","gpu","cuda","pytorch","jax","fp8"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":31,"issueCount":96,"starsCount":1521,"forksCount":236,"license":"Apache License 2.0","participation":[7,8,8,2,2,17,5,7,6,9,5,5,3,2,3,11,6,11,6,10,2,7,6,6,4,12,6,9,0,4,9,6,15,11,6,7,8,5,10,4,5,4,9,8,12,10,11,4,9,10,6,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:19:53.815Z"}},{"type":"Public","name":"spark-rapids-benchmarks","owner":"NVIDIA","isFork":false,"description":"Spark RAPIDS Benchmarks – benchmark sets and utilities for the RAPIDS Accelerator for Apache Spark","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":21,"starsCount":32,"forksCount":26,"license":"Apache License 2.0","participation":[0,0,1,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,3,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T19:39:21.066Z"}},{"type":"Public","name":"NeMo-Guardrails","owner":"NVIDIA","isFork":false,"description":"NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":27,"issueCount":160,"starsCount":3579,"forksCount":307,"license":"Other","participation":[14,36,46,18,20,27,21,28,15,18,15,21,84,37,29,34,55,71,58,46,51,58,68,53,42,39,47,36,1,11,21,43,26,47,37,56,36,57,29,36,37,81,23,0,18,30,5,12,8,6,14,28],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T19:33:06.343Z"}},{"type":"Public","name":"NeMo-Aligner","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for efficient model alignment","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":40,"starsCount":290,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,9,2,12,2,3,2,0,2,1,2,5,1,2,0,2,4,3,2,3,2,4,3,1,3,4,0,2,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T21:33:36.230Z"}},{"type":"Public","name":"swift","owner":"NVIDIA","isFork":true,"description":"OpenStack Storage (Swift). Mirror of code maintained at opendev.org.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1059,"license":"Apache License 2.0","participation":[13,4,19,10,5,3,13,1,23,8,15,4,5,13,5,14,3,5,8,3,5,6,5,6,2,5,2,2,1,3,10,3,10,4,9,7,5,8,8,15,7,4,4,1,0,4,13,4,15,7,3,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T21:08:42.172Z"}},{"type":"Public","name":"GenerativeAIExamples","owner":"NVIDIA","isFork":false,"description":"Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.","allTopics":["microservice","gpu-acceleration","nemo","tensorrt","rag","triton-inference-server","large-language-models","llm","llm-inference","retrieval-augmented-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":20,"starsCount":1671,"forksCount":269,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,1,1,2,0,4,0,2,2,0,1,2,2,0,4,2,0,0,1,2,1,2,1,1,0,0,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T17:52:56.688Z"}},{"type":"Public","name":"Megatron-Energon","owner":"NVIDIA","isFork":false,"description":"Megatron's multi-modal data loader","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":0,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,0,3,5,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T10:17:17.102Z"}},{"type":"Public","name":"earth2studio","owner":"NVIDIA","isFork":false,"description":"Open-source deep-learning framework for exploring, building and deploying AI weather/climate workflows.","allTopics":["weather","ai","deep-learning","climate-science"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":36,"forksCount":8,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,12,13,7,4,8,3,3,4,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T03:51:58.833Z"}},{"type":"Public","name":"workbench-example-hybrid-rag","owner":"NVIDIA","isFork":false,"description":"An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":47,"forksCount":126,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,7,0,0,0,0,0,0,0,2,1,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T23:53:26.713Z"}},{"type":"Public","name":"Stable-Diffusion-WebUI-TensorRT","owner":"NVIDIA","isFork":false,"description":"TensorRT Extension for Stable Diffusion Web UI","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":139,"starsCount":1808,"forksCount":138,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T13:49:56.975Z"}},{"type":"Public","name":"NeMo-text-processing","owner":"NVIDIA","isFork":false,"description":"NeMo text processing for ASR and TTS","allTopics":["text-normalization","inverse-text-n"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":5,"starsCount":227,"forksCount":73,"license":"Apache License 2.0","participation":[4,1,1,3,1,1,0,0,2,2,0,2,2,1,0,1,2,1,0,3,2,1,0,1,0,0,1,0,0,0,0,0,1,0,0,1,1,0,1,4,1,0,1,0,1,0,3,3,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T21:51:08.664Z"}},{"type":"Public","name":"nv-cloud-function-helpers","owner":"NVIDIA","isFork":false,"description":"Functions that simplify common tasks with NVIDIA Cloud Functions","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":0,"starsCount":8,"forksCount":2,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,5,9,5,0,1,0,0,0,0,1,0,0,0,0,6,2,1,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T18:47:08.264Z"}},{"type":"Public","name":"TensorRT-Model-Optimizer","owner":"NVIDIA","isFork":false,"description":"TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization and sparsity. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed on NVIDIA GPUs.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":8,"starsCount":235,"forksCount":13,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T07:44:32.173Z"}},{"type":"Public","name":"tao_tensorflow2_backend","owner":"NVIDIA","isFork":false,"description":"TAO Toolkit deep learning networks with TensorFlow 2.x backend","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":5,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,1,3,0,0,0,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-29T01:23:40.767Z"}},{"type":"Public","name":"air_sdk","owner":"NVIDIA","isFork":false,"description":"A Python SDK library for interacting with NVIDIA Air","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":5,"forksCount":4,"license":"Other","participation":[0,0,0,0,0,1,10,0,6,3,0,1,0,0,2,0,1,0,0,1,2,2,5,0,1,1,0,1,0,1,2,2,1,0,1,0,1,0,0,1,0,0,1,0,1,0,1,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T17:24:35.791Z"}},{"type":"Public","name":"cuda-python","owner":"NVIDIA","isFork":false,"description":"CUDA Python Low-level Bindings","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":10,"starsCount":788,"forksCount":60,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T16:20:17.052Z"}},{"type":"Public","name":"audio-flamingo","owner":"NVIDIA","isFork":false,"description":"PyTorch implementation of Audio Flamingo: A Novel Audio Language Model with Few-Shot Learning and Dialogue Abilities.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":9,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-20T18:40:22.567Z"}},{"type":"Public","name":"hpc-container-maker","owner":"NVIDIA","isFork":false,"description":"HPC Container Maker","allTopics":["docker","containers","hpc","singularity"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":11,"starsCount":439,"forksCount":86,"license":"Apache License 2.0","participation":[3,0,0,0,0,0,0,0,0,0,0,0,4,0,1,0,0,3,0,0,1,0,0,3,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-04T22:00:21.650Z"}},{"type":"Public","name":"modulus-sym","owner":"NVIDIA","isFork":false,"description":"Framework providing pythonic APIs, algorithms and utilities to be used with Modulus core to physics inform model training as well as higher level abstraction for domain experts","allTopics":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":47,"starsCount":125,"forksCount":52,"license":"Apache License 2.0","participation":[1,1,1,4,0,0,4,5,1,1,0,1,1,0,1,3,4,0,0,0,2,1,2,4,1,2,1,1,0,0,1,1,2,2,0,0,2,1,3,0,0,0,1,2,1,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-16T16:07:46.025Z"}}],"repositoryCount":109,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}