{"payload":{"pageCount":16,"repositories":[{"type":"Public","name":"spark-rapids","owner":"NVIDIA","isFork":false,"description":"Spark RAPIDS plugin - accelerate Apache Spark with GPUs","allTopics":["big-data","gpu","rapids","spark"],"primaryLanguage":{"name":"Scala","color":"#c22d40"},"pullRequestCount":26,"issueCount":1308,"starsCount":742,"forksCount":220,"license":"Apache License 2.0","participation":[33,8,10,12,12,21,23,48,27,60,12,20,18,20,14,21,42,14,20,27,27,23,30,60,33,44,35,15,11,26,18,7,22,46,19,21,15,14,24,9,10,8,13,8,11,6,11,6,20,14,26,18],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T03:16:12.083Z"}},{"type":"Public","name":"spark-rapids-jni","owner":"NVIDIA","isFork":false,"description":"RAPIDS Accelerator JNI For Apache Spark","allTopics":[],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":7,"issueCount":69,"starsCount":31,"forksCount":56,"license":"Apache License 2.0","participation":[18,5,14,10,13,12,18,20,15,16,15,16,12,15,8,43,16,19,15,10,15,7,22,24,17,19,25,11,1,4,12,13,23,27,15,19,12,14,17,21,10,26,8,24,22,16,13,13,8,21,27,19],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T03:13:11.004Z"}},{"type":"Public","name":"NeMo","owner":"NVIDIA","isFork":false,"description":"A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)","allTopics":["machine-translation","tts","speech-synthesis","neural-networks","deeplearning","speaker-recognition","asr","multimodal","speech-translation","large-language-models","speaker-diariazation","generative-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":89,"issueCount":54,"starsCount":10363,"forksCount":2216,"license":"Apache License 2.0","participation":[14,12,22,10,17,21,11,17,14,22,23,17,21,23,16,21,31,25,13,24,6,12,4,13,4,6,20,14,4,8,15,17,28,13,13,21,22,39,24,20,41,19,13,34,30,27,28,38,28,30,20,28],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:54:20.868Z"}},{"type":"Public","name":"cuda-quantum","owner":"NVIDIA","isFork":false,"description":"C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows","allTopics":["python","cpp","quantum","quantum-computing","quantum-programming-language","quantum-algorithms","quantum-machine-learning","unitaryhack"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":34,"issueCount":231,"starsCount":406,"forksCount":135,"license":"Other","participation":[13,8,26,17,28,33,26,10,16,14,23,6,15,21,6,13,26,6,16,24,12,20,29,11,4,23,11,12,4,1,11,7,17,9,18,16,9,17,27,23,20,20,8,7,11,12,10,21,22,8,5,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:35:54.772Z"}},{"type":"Public","name":"cloud-native-stack","owner":"NVIDIA","isFork":false,"description":"Run cloud native workloads on NVIDIA GPUs","allTopics":[],"primaryLanguage":{"name":"Markdown","color":"#083fa1"},"pullRequestCount":2,"issueCount":3,"starsCount":117,"forksCount":46,"license":"Apache License 2.0","participation":[1,0,2,1,3,3,0,1,0,1,3,1,3,0,0,0,4,0,0,0,0,0,0,0,0,0,2,1,0,0,0,0,1,0,0,0,0,0,0,0,0,7,3,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:34:09.502Z"}},{"type":"Public","name":"gpu-operator","owner":"NVIDIA","isFork":false,"description":"NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes","allTopics":["kubernetes","gpu","cuda","nvidia"],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":11,"issueCount":229,"starsCount":1215,"forksCount":231,"license":"Apache License 2.0","participation":[11,6,8,2,8,12,8,21,6,9,8,25,17,46,18,44,67,40,20,15,8,15,19,12,2,21,4,4,1,5,0,0,4,6,5,7,4,11,16,13,16,5,0,7,10,14,11,7,5,0,1,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:17:00.235Z"}},{"type":"Public","name":"Fuser","owner":"NVIDIA","isFork":false,"description":"A Fusion Code Generator for NVIDIA GPUs (commonly known as \"nvFuser\")","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":95,"issueCount":220,"starsCount":228,"forksCount":42,"license":"Other","participation":[16,6,29,5,23,25,16,17,13,27,31,27,19,19,25,48,23,22,27,30,28,40,30,51,30,35,42,24,4,2,28,18,14,15,23,19,30,24,16,38,32,23,16,20,26,35,18,17,21,21,15,37],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:11:45.884Z"}},{"type":"Public","name":"open-gpu-kernel-modules","owner":"NVIDIA","isFork":false,"description":"NVIDIA Linux open GPU kernel module source","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":37,"issueCount":121,"starsCount":14099,"forksCount":1157,"license":"Other","participation":[1,0,0,1,0,1,0,1,1,0,1,0,0,0,0,1,0,0,1,0,3,0,0,1,0,0,0,0,0,0,0,0,2,0,0,0,0,1,0,0,2,0,0,0,1,0,1,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:07:00.421Z"}},{"type":"Public","name":"cccl","owner":"NVIDIA","isFork":false,"description":"CUDA C++ Core Libraries","allTopics":["modern-cpp","parallel-computing","gpu-acceleration","cuda-kernels","gpu-computing","parallel-algorithm","parallel-programming","nvidia-gpu","gpu-programming","cuda-library","cpp-programming","cuda-programming","accelerated-computing","cuda-cpp","cpp","hpc","gpu","cuda","nvidia"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":32,"issueCount":613,"starsCount":867,"forksCount":112,"license":"Other","participation":[11,43,73,30,36,33,18,23,27,34,34,13,12,39,27,33,30,50,18,26,16,32,13,13,11,8,11,7,1,1,4,7,13,7,6,15,14,14,18,17,8,9,7,13,19,7,9,23,15,5,10,14],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T02:05:38.665Z"}},{"type":"Public","name":"NV-Kernels","owner":"NVIDIA","isFork":false,"description":"Ubuntu kernels which are optimized for NVIDIA server systems","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":9,"issueCount":0,"starsCount":7,"forksCount":7,"license":"Other","participation":[2095,1821,1363,1250,1562,1809,1669,2168,2222,1947,1574,1610,1108,1758,2127,1714,1831,1708,1772,1796,1300,1097,1265,1528,1635,2080,1942,2045,959,717,1237,1193,1655,1972,1832,1973,1850,2034,1704,1439,1216,1671,1605,1763,1705,1604,1465,1057,606,341,220,73],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T01:46:22.367Z"}},{"type":"Public","name":"cloudai","owner":"NVIDIA","isFork":false,"description":"CloudAI Benchmark Framework","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":0,"starsCount":16,"forksCount":9,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,22,56,34,70],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T01:44:40.984Z"}},{"type":"Public","name":"spark-rapids-tools","owner":"NVIDIA","isFork":false,"description":"User tools for Spark RAPIDS","allTopics":[],"primaryLanguage":{"name":"Scala","color":"#c22d40"},"pullRequestCount":6,"issueCount":191,"starsCount":42,"forksCount":34,"license":"Apache License 2.0","participation":[1,2,3,10,9,7,6,5,0,10,4,7,5,15,1,7,6,4,6,3,5,6,1,5,2,2,4,3,4,5,2,3,5,13,1,5,2,7,4,2,3,15,6,11,12,7,10,8,2,7,6,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T01:14:03.620Z"}},{"type":"Public","name":"numbast","owner":"NVIDIA","isFork":false,"description":"Numbast is a tool to build an automated pipeline that converts CUDA APIs into Numba bindings.","allTopics":["cuda","numba"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":17,"starsCount":13,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,15,0,6,1,8,0,4,0,2,1,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:54:59.890Z"}},{"type":"Public","name":"warp","owner":"NVIDIA","isFork":false,"description":"A Python framework for high performance GPU simulation and graphics","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":48,"starsCount":1753,"forksCount":150,"license":"Other","participation":[25,26,57,38,10,17,41,9,5,32,24,20,41,7,4,18,29,14,28,32,63,27,32,23,1,26,23,24,0,2,19,39,29,12,6,10,21,48,50,45,40,14,9,14,20,37,35,40,26,21,35,33],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:50:27.828Z"}},{"type":"Public","name":"nim-deploy","owner":"NVIDIA","isFork":false,"description":"A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":1,"issueCount":1,"starsCount":24,"forksCount":11,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,18],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:37:01.164Z"}},{"type":"Public","name":"TransformerEngine","owner":"NVIDIA","isFork":false,"description":"A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.","allTopics":["python","machine-learning","deep-learning","gpu","cuda","pytorch","jax","fp8"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":31,"issueCount":96,"starsCount":1521,"forksCount":236,"license":"Apache License 2.0","participation":[7,8,8,2,2,17,5,7,6,9,5,5,3,2,3,11,6,11,6,10,2,7,6,6,4,12,6,9,0,4,9,6,15,11,6,7,8,5,10,4,5,4,9,8,12,10,11,4,9,10,6,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:19:53.815Z"}},{"type":"Public","name":"DALI","owner":"NVIDIA","isFork":false,"description":"A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.","allTopics":["python","machine-learning","deep-learning","neural-network","mxnet","gpu","image-processing","pytorch","gpu-tensorflow","data-processing","data-augmentation","audio-processing","paddle","image-augmentation","fast-data-pipeline"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":37,"issueCount":179,"starsCount":4953,"forksCount":607,"license":"Apache License 2.0","participation":[11,3,8,6,8,5,4,8,14,6,7,10,9,10,3,5,9,13,6,8,11,14,5,14,20,16,16,5,0,2,6,4,9,10,7,8,13,4,12,13,10,12,3,11,10,10,4,4,10,5,8,10],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-06T00:03:24.303Z"}},{"type":"Public","name":"AMGX","owner":"NVIDIA","isFork":false,"description":"Distributed multigrid linear solver library on GPU","allTopics":[],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":5,"issueCount":84,"starsCount":456,"forksCount":134,"license":null,"participation":[0,0,1,1,0,0,0,0,0,0,0,0,0,2,0,0,6,5,2,7,0,0,4,1,0,6,4,0,0,0,1,0,3,0,0,0,3,0,1,0,0,0,0,0,0,0,1,0,0,0,0,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:55:31.481Z"}},{"type":"Public","name":"NeMo-Curator","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for data curation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":28,"starsCount":291,"forksCount":30,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,4,3,2,0,6,1,6,3,4,2,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:27:35.697Z"}},{"type":"Public","name":"gpu-driver-container","owner":"NVIDIA","isFork":false,"description":"The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":2,"issueCount":4,"starsCount":48,"forksCount":27,"license":"Apache License 2.0","participation":[3,0,5,0,4,0,2,4,10,2,2,6,7,5,1,4,2,0,2,7,5,3,6,2,1,1,0,0,0,0,0,4,3,4,6,1,3,8,9,0,4,8,1,0,1,1,0,1,1,3,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:13:08.875Z"}},{"type":"Public","name":"Megatron-LM","owner":"NVIDIA","isFork":false,"description":"Ongoing research training transformer models at scale","allTopics":["transformers","model-para","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":128,"issueCount":311,"starsCount":8980,"forksCount":2024,"license":"Other","participation":[12,4,22,14,59,28,38,27,35,51,61,59,32,70,34,36,58,133,101,131,62,33,41,76,24,41,36,59,10,6,39,31,56,61,42,19,5,18,23,41,60,92,44,50,63,30,17,35,25,21,14,31],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:06:11.886Z"}},{"type":"Public","name":"NeMo-speech-data-processor","owner":"NVIDIA","isFork":false,"description":"A toolkit for processing speech data and creating speech datasets","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":1,"starsCount":63,"forksCount":20,"license":"Apache License 2.0","participation":[3,0,2,0,0,17,7,14,0,0,0,0,0,0,0,0,0,0,0,1,1,7,13,3,0,1,0,0,0,0,1,0,0,0,0,0,0,1,0,0,0,5,0,0,1,0,1,1,3,1,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T23:01:18.903Z"}},{"type":"Public","name":"NeMo-Framework-Launcher","owner":"NVIDIA","isFork":false,"description":"NeMo Megatron launcher and tools","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":10,"issueCount":23,"starsCount":403,"forksCount":118,"license":"Apache License 2.0","participation":[15,8,3,4,4,11,12,12,14,14,1,12,13,27,8,1,8,10,27,13,7,6,11,17,10,24,36,23,11,12,23,22,27,19,16,55,25,33,35,17,34,5,6,23,11,29,13,22,54,34,1,19],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:44:19.457Z"}},{"type":"Public","name":"modulus","owner":"NVIDIA","isFork":false,"description":"Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods","allTopics":["machine-learning","deep-learning","physics","pytorch","nvidia-gpu"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":17,"issueCount":95,"starsCount":733,"forksCount":153,"license":"Apache License 2.0","participation":[8,1,4,4,2,2,9,9,11,5,1,2,1,3,3,11,7,4,3,4,1,10,5,10,5,10,5,4,0,4,4,5,7,6,5,3,7,7,8,2,3,2,4,10,9,3,5,9,2,3,4,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:32:40.998Z"}},{"type":"Public","name":"JAX-Toolbox","owner":"NVIDIA","isFork":false,"description":"JAX-Toolbox","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":37,"issueCount":111,"starsCount":188,"forksCount":34,"license":"Apache License 2.0","participation":[4,2,12,3,9,17,3,7,7,3,5,19,5,11,5,11,9,12,8,8,9,4,5,8,6,5,10,7,2,1,19,12,10,13,8,13,17,15,11,11,6,6,1,8,10,0,11,11,8,6,6,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:22:22.902Z"}},{"type":"Public","name":"NVFlare","owner":"NVIDIA","isFork":false,"description":"NVIDIA Federated Learning Application Runtime Environment","allTopics":["python"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":24,"starsCount":556,"forksCount":150,"license":"Apache License 2.0","participation":[9,2,7,1,16,6,11,12,13,9,14,12,10,17,19,17,12,5,6,11,16,16,9,9,9,10,9,24,5,3,17,7,9,7,11,2,3,4,2,5,9,4,10,8,7,14,9,8,8,5,9,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:11:57.574Z"}},{"type":"Public","name":"VisRTX","owner":"NVIDIA","isFork":false,"description":"NVIDIA OptiX based implementation of ANARI","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":5,"starsCount":234,"forksCount":25,"license":"Other","participation":[10,0,0,2,5,0,4,4,4,4,8,3,3,2,9,4,7,0,1,1,0,5,6,5,6,6,2,2,0,0,0,0,0,3,5,1,2,3,5,0,2,3,1,3,3,1,7,0,3,1,8,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:03:33.941Z"}},{"type":"Public","name":"NeMo-text-processing","owner":"NVIDIA","isFork":false,"description":"NeMo text processing for ASR and TTS","allTopics":["text-normalization","inverse-text-n"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":5,"starsCount":227,"forksCount":73,"license":"Apache License 2.0","participation":[4,1,1,3,1,1,0,0,2,2,0,2,2,1,0,1,2,1,0,3,2,1,0,1,0,0,1,0,0,0,0,0,1,0,0,1,1,0,1,4,1,0,1,0,1,0,3,3,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T21:51:08.664Z"}},{"type":"Public","name":"NeMo-Aligner","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for efficient model alignment","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":40,"starsCount":290,"forksCount":32,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,9,2,12,2,3,2,0,2,1,2,5,1,2,0,2,4,3,2,3,2,4,3,1,3,4,0,2,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T21:33:36.230Z"}},{"type":"Public","name":"spark-rapids-benchmarks","owner":"NVIDIA","isFork":false,"description":"Spark RAPIDS Benchmarks – benchmark sets and utilities for the RAPIDS Accelerator for Apache Spark","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":21,"starsCount":32,"forksCount":26,"license":"Apache License 2.0","participation":[0,0,1,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,3,0,1,0,1,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T19:39:21.066Z"}}],"repositoryCount":474,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}