{"payload":{"pageCount":16,"repositories":[{"type":"Public","name":"spark-rapids","owner":"NVIDIA","isFork":false,"description":"Spark RAPIDS plugin - accelerate Apache Spark with GPUs","allTopics":["big-data","gpu","rapids","spark"],"primaryLanguage":{"name":"Scala","color":"#c22d40"},"pullRequestCount":19,"issueCount":1304,"starsCount":737,"forksCount":219,"license":"Apache License 2.0","participation":[34,16,10,15,10,18,17,42,32,54,29,15,19,20,11,21,38,23,14,22,33,24,15,55,36,47,33,35,9,27,18,14,7,34,40,19,22,10,17,19,10,8,6,12,13,5,11,7,16,13,14,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T03:40:07.982Z"}},{"type":"Public","name":"stdexec","owner":"NVIDIA","isFork":false,"description":"`std::execution`, the proposed C++ framework for asynchronous and parallel programming.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":10,"issueCount":75,"starsCount":1313,"forksCount":141,"license":"Apache License 2.0","participation":[20,8,9,3,4,4,14,10,1,5,19,41,24,18,15,8,31,5,25,22,7,38,8,22,19,15,8,10,3,12,29,16,9,8,26,9,12,6,18,13,16,8,6,8,0,4,12,10,67,9,4,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T03:21:33.323Z"}},{"type":"Public","name":"ACE","owner":"NVIDIA","isFork":false,"description":"NVIDIA ACE samples, workflows, and resources","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":0,"starsCount":8,"forksCount":3,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T03:07:15.316Z"}},{"type":"Public","name":"TensorRT-LLM","owner":"NVIDIA","isFork":false,"description":"TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":84,"issueCount":561,"starsCount":6997,"forksCount":746,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,5,0,6,8,1,1,3,5,1,1,4,1,0,1,3,2,1,1,6,1,1,3,3,1,1,1,1,1,1,1,1,1,1,1,4,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T02:42:40.922Z"}},{"type":"Public","name":"cuda-quantum","owner":"NVIDIA","isFork":false,"description":"C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows","allTopics":["python","cpp","quantum","quantum-computing","quantum-programming-language","quantum-algorithms","quantum-machine-learning","unitaryhack"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":28,"issueCount":231,"starsCount":402,"forksCount":131,"license":"Other","participation":[15,3,17,21,33,18,38,20,3,22,25,5,11,20,11,10,19,18,7,21,15,23,16,30,5,14,15,8,13,3,5,7,13,12,14,16,15,15,20,27,17,24,9,13,5,12,12,13,28,12,7,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T02:40:45.017Z"}},{"type":"Public","name":"Fuser","owner":"NVIDIA","isFork":false,"description":"A Fusion Code Generator for NVIDIA GPUs (commonly known as \"nvFuser\")","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":106,"issueCount":217,"starsCount":228,"forksCount":42,"license":"Other","participation":[17,10,20,16,15,20,28,14,13,25,23,31,23,17,12,45,37,25,20,31,31,33,32,42,34,44,34,41,7,3,16,27,11,16,16,25,15,36,23,19,41,28,10,29,11,37,25,19,22,20,19,16],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T02:19:30.209Z"}},{"type":"Public","name":"TransformerEngine","owner":"NVIDIA","isFork":false,"description":"A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.","allTopics":["python","machine-learning","deep-learning","gpu","cuda","pytorch","jax","fp8"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":29,"issueCount":96,"starsCount":1509,"forksCount":231,"license":"Apache License 2.0","participation":[9,8,9,6,2,10,11,4,5,11,8,4,2,3,1,9,6,8,12,7,6,3,6,9,3,9,8,12,1,0,8,8,11,10,13,6,6,6,6,8,5,7,3,13,4,14,6,13,4,9,9,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T01:41:50.416Z"}},{"type":"Public","name":"vgpu-device-manager","owner":"NVIDIA","isFork":false,"description":"NVIDIA vGPU Device Manager manages NVIDIA vGPU devices on top of Kubernetes","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":6,"issueCount":0,"starsCount":79,"forksCount":13,"license":"Apache License 2.0","participation":[0,0,0,0,0,5,3,2,0,0,0,0,0,0,0,0,2,0,0,4,0,0,0,0,2,0,0,0,4,3,0,0,0,0,0,0,0,0,0,0,28,0,0,0,10,7,2,0,2,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T01:27:42.762Z"}},{"type":"Public","name":"nim-deploy","owner":"NVIDIA","isFork":false,"description":"A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":1,"issueCount":0,"starsCount":7,"forksCount":4,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T00:54:50.356Z"}},{"type":"Public","name":"vagrant-swift-all-in-one","owner":"NVIDIA","isFork":false,"description":"Vagrant Swift All In One","allTopics":[],"primaryLanguage":{"name":"Ruby","color":"#701516"},"pullRequestCount":22,"issueCount":3,"starsCount":78,"forksCount":58,"license":"Apache License 2.0","participation":[4,0,3,0,0,0,0,0,0,0,0,0,0,0,2,2,4,0,3,0,0,0,0,0,0,0,0,0,0,0,6,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,1,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T00:50:38.813Z"}},{"type":"Public","name":"nvidia-hpcg","owner":"NVIDIA","isFork":false,"description":"NVIDIA HPCG is based on the HPCG benchmark and optimized for performance on NVIDIA accelerated HPC systems.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":21,"forksCount":0,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T00:23:34.344Z"}},{"type":"Public","name":"NeMo","owner":"NVIDIA","isFork":false,"description":"A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)","allTopics":["machine-translation","tts","speech-synthesis","neural-networks","deeplearning","speaker-recognition","asr","multimodal","speech-translation","large-language-models","speaker-diariazation","generative-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":90,"issueCount":51,"starsCount":10324,"forksCount":2204,"license":"Apache License 2.0","participation":[23,9,15,22,8,22,17,11,16,23,19,16,23,26,9,18,32,26,22,15,19,8,7,12,6,6,11,17,12,4,10,20,22,20,9,15,33,19,38,20,37,28,7,28,35,25,25,37,30,33,22,22],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-03T00:13:43.314Z"}},{"type":"Public","name":"DALI","owner":"NVIDIA","isFork":false,"description":"A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.","allTopics":["python","machine-learning","deep-learning","neural-network","mxnet","gpu","image-processing","pytorch","gpu-tensorflow","data-processing","data-augmentation","audio-processing","paddle","image-augmentation","fast-data-pipeline"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":36,"issueCount":180,"starsCount":4948,"forksCount":606,"license":"Apache License 2.0","participation":[6,9,7,9,2,10,3,4,13,14,0,11,9,11,5,4,11,10,6,5,16,8,11,11,13,24,14,9,3,0,3,6,10,6,8,7,13,7,6,14,12,11,7,10,10,9,7,2,10,6,5,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T23:04:56.470Z"}},{"type":"Public","name":"cuCollections","owner":"NVIDIA","isFork":false,"description":"","allTopics":["datastructures","cpp","gpu","cuda","hashmap","cpp17","hashset","hashtable"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":15,"issueCount":65,"starsCount":440,"forksCount":77,"license":"Apache License 2.0","participation":[1,1,1,6,2,2,1,1,6,1,1,4,3,3,2,0,2,3,4,1,0,1,0,4,4,1,1,2,7,0,4,1,2,3,4,1,2,0,0,0,4,3,0,3,3,3,1,1,3,2,3,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T19:49:20.116Z"}},{"type":"Public","name":"NeMo-Framework-Launcher","owner":"NVIDIA","isFork":false,"description":"NeMo Megatron launcher and tools","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":12,"issueCount":23,"starsCount":401,"forksCount":117,"license":"Apache License 2.0","participation":[10,7,8,3,5,6,17,9,14,11,10,12,5,17,18,9,0,9,35,7,7,9,4,16,12,21,22,36,22,4,19,22,25,24,15,58,23,31,27,28,40,5,5,14,18,19,24,17,30,54,15,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T18:07:40.614Z"}},{"type":"Public","name":"k8s-dra-driver","owner":"NVIDIA","isFork":false,"description":"Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":12,"issueCount":15,"starsCount":175,"forksCount":30,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,25,20,24,0,18,0,13,25,2,0,27,0,0,3,0,1,0,0,0,24,1,1,9,1,8,11,9,1,2,1,1,5,3,0,11,1,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T17:28:32.349Z"}},{"type":"Public","name":"knavigator","owner":"NVIDIA","isFork":false,"description":"knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":2,"issueCount":0,"starsCount":33,"forksCount":5,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,0,3,46,15,19,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T16:34:35.297Z"}},{"type":"Public","name":"cccl","owner":"NVIDIA","isFork":false,"description":"CUDA C++ Core Libraries","allTopics":["modern-cpp","parallel-computing","gpu-acceleration","cuda-kernels","gpu-computing","parallel-algorithm","parallel-programming","nvidia-gpu","gpu-programming","cuda-library","cpp-programming","cuda-programming","accelerated-computing","cuda-cpp","cpp","hpc","gpu","cuda","nvidia"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":33,"issueCount":612,"starsCount":858,"forksCount":112,"license":"Other","participation":[95,4,97,38,13,54,16,26,29,28,40,18,14,18,25,42,29,20,56,27,14,34,12,14,10,14,8,13,1,1,3,2,16,8,4,10,13,13,21,18,11,7,7,10,22,12,3,14,20,14,8,10],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T16:26:11.584Z"}},{"type":"Public","name":"Megatron-LM","owner":"NVIDIA","isFork":false,"description":"Ongoing research training transformer models at scale","allTopics":["transformers","model-para","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":130,"issueCount":306,"starsCount":8940,"forksCount":2009,"license":"Other","participation":[36,5,10,19,40,36,43,22,43,23,73,67,38,46,51,30,51,88,122,136,91,44,36,76,22,34,46,38,42,9,19,40,46,62,44,26,18,12,18,27,54,70,58,61,72,47,10,39,14,30,18,14],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T16:23:03.886Z"}},{"type":"Public","name":"NV-Kernels","owner":"NVIDIA","isFork":false,"description":"Ubuntu kernels which are optimized for NVIDIA server systems","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":6,"issueCount":0,"starsCount":7,"forksCount":5,"license":"Other","participation":[2052,1972,1574,1234,1437,1719,1657,1922,2152,2318,1676,1685,1305,1082,2156,1974,1707,1784,1701,1918,1550,1201,1145,1285,1611,1905,2013,2025,1803,531,973,1249,1173,1962,2090,1617,2160,2053,1610,1730,1204,1264,1802,1752,1759,1481,1654,1311,867,426,298,146],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T15:00:12.639Z"}},{"type":"Public","name":"Stable-Diffusion-WebUI-TensorRT","owner":"NVIDIA","isFork":false,"description":"TensorRT Extension for Stable Diffusion Web UI","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":15,"issueCount":139,"starsCount":1800,"forksCount":138,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,8,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T13:49:56.975Z"}},{"type":"Public","name":"JAX-Toolbox","owner":"NVIDIA","isFork":false,"description":"JAX-Toolbox","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":39,"issueCount":111,"starsCount":188,"forksCount":34,"license":"Apache License 2.0","participation":[6,3,5,10,6,8,15,4,7,5,6,12,12,3,9,7,9,13,15,4,9,7,6,3,9,6,8,5,8,0,11,15,7,14,11,13,15,9,19,10,10,3,5,3,11,5,1,16,6,10,6,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T13:35:29.497Z"}},{"type":"Public","name":"cloudai","owner":"NVIDIA","isFork":false,"description":"CloudAI Benchmark Framework","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":0,"starsCount":13,"forksCount":8,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,49,33,47],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T10:22:09.803Z"}},{"type":"Public","name":"nvidia-container-toolkit","owner":"NVIDIA","isFork":false,"description":"Build and run containers leveraging NVIDIA GPUs","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":13,"issueCount":250,"starsCount":1720,"forksCount":198,"license":"Apache License 2.0","participation":[20,7,9,16,19,5,22,0,5,24,12,21,7,8,0,4,1,0,1,5,5,9,1,12,38,12,11,13,0,0,2,12,2,10,26,18,16,3,20,3,13,10,3,10,19,12,6,3,6,4,5,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T08:42:23.478Z"}},{"type":"Public","name":"mig-parted","owner":"NVIDIA","isFork":false,"description":"MIG Partition Editor for NVIDIA GPUs","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":6,"issueCount":15,"starsCount":150,"forksCount":36,"license":"Apache License 2.0","participation":[0,0,0,0,0,7,0,0,0,0,1,0,0,4,0,0,1,1,0,5,2,6,0,0,3,0,0,0,0,0,0,0,0,27,0,2,0,0,10,0,1,0,0,1,20,17,6,1,2,3,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T08:13:50.621Z"}},{"type":"Public","name":"NeMo-Aligner","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for efficient model alignment","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":22,"issueCount":40,"starsCount":286,"forksCount":31,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,4,7,8,7,1,3,1,0,3,0,7,0,2,1,0,4,3,3,2,2,3,3,4,2,5,0,1,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T07:34:54.045Z"}},{"type":"Public","name":"k8s-device-plugin","owner":"NVIDIA","isFork":false,"description":"NVIDIA device plugin for Kubernetes","allTopics":["kubernetes"],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":31,"issueCount":147,"starsCount":2475,"forksCount":577,"license":"Apache License 2.0","participation":[11,0,0,2,5,12,0,2,1,0,10,0,0,0,0,0,0,0,0,7,12,2,3,20,4,2,8,19,0,0,0,20,5,31,12,17,27,29,42,17,35,2,12,16,34,20,21,11,11,7,6,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T07:25:11.617Z"}},{"type":"Public","name":"NeMo-Curator","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for data curation","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":28,"starsCount":279,"forksCount":28,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,4,2,4,1,1,5,4,4,2,6,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T02:30:45.545Z"}},{"type":"Public","name":"cloud-native-docs","owner":"NVIDIA","isFork":false,"description":"Documentation repository for NVIDIA Cloud Native Technologies","allTopics":["kubernetes","containers","kubernetes-operator"],"primaryLanguage":{"name":"CSS","color":"#563d7c"},"pullRequestCount":6,"issueCount":2,"starsCount":12,"forksCount":13,"license":"Apache License 2.0","participation":[20,22,4,2,12,10,9,2,16,2,2,4,15,9,4,5,9,8,2,12,6,2,0,4,0,8,14,8,2,6,0,10,7,10,15,12,6,1,1,18,5,2,3,5,2,7,10,35,2,3,1,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T01:03:03.006Z"}},{"type":"Public","name":"holodeck","owner":"NVIDIA","isFork":false,"description":"Holodeck is a project to create test environments optimised for GPU projects.","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":11,"issueCount":4,"starsCount":7,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,0,11,6,0,7,8,2,24,0,0,2,0,0,0,0,0,13,19,9,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-02T00:59:12.813Z"}}],"repositoryCount":473,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}