{"payload":{"pageCount":17,"repositories":[{"type":"Public","name":"warp","owner":"NVIDIA","isFork":false,"description":"A Python framework for high performance GPU simulation and graphics","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":51,"starsCount":4104,"forksCount":227,"license":"Other","participation":[15,31,13,39,41,45,37,19,23,3,26,23,22,0,6,22,36,30,11,2,13,26,53,51,49,27,10,15,7,35,35,33,37,29,18,39,49,20,11,9,56,22,17,26,29,9,30,15,33,27,6,24],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:25:26.632Z"}},{"type":"Public","name":"NeMo","owner":"NVIDIA","isFork":false,"description":"A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)","allTopics":["machine-translation","tts","speech-synthesis","neural-networks","deeplearning","speaker-recognition","asr","multimodal","speech-translation","large-language-models","speaker-diariazation","generative-ai"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":87,"issueCount":50,"starsCount":11549,"forksCount":2419,"license":"Apache License 2.0","participation":[24,31,23,14,23,4,12,6,11,3,13,14,14,4,10,17,17,28,9,16,19,23,43,20,21,44,13,22,29,30,25,33,38,26,27,23,34,38,16,28,34,48,30,30,44,42,41,47,41,30,65,38],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:21:41.692Z"}},{"type":"Public","name":"Fuser","owner":"NVIDIA","isFork":false,"description":"A Fusion Code Generator for NVIDIA GPUs (commonly known as \"nvFuser\")","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":116,"issueCount":253,"starsCount":250,"forksCount":49,"license":"Other","participation":[44,21,24,27,29,30,39,33,43,37,36,45,13,3,10,26,15,13,20,21,18,35,15,22,37,30,21,22,13,31,30,22,20,19,20,17,40,28,20,35,34,28,50,33,42,23,13,30,22,34,17,27],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:08:30.718Z"}},{"type":"Public","name":"TensorRT-Incubator","owner":"NVIDIA","isFork":false,"description":"Experimental projects related to TensorRT","allTopics":[],"primaryLanguage":{"name":"MLIR","color":"#5EC8DB"},"pullRequestCount":14,"issueCount":35,"starsCount":64,"forksCount":9,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,53,22,17,16,16,15,8],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:03:06.301Z"}},{"type":"Public","name":"NeMo-Guardrails","owner":"NVIDIA","isFork":false,"description":"NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":13,"issueCount":183,"starsCount":3990,"forksCount":365,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T23:46:22.811Z"}},{"type":"Public","name":"cuda-quantum","owner":"NVIDIA","isFork":false,"description":"C++ and Python support for the CUDA Quantum programming model for heterogeneous quantum-classical workflows","allTopics":["python","cpp","quantum","quantum-computing","quantum-programming-language","quantum-algorithms","quantum-machine-learning","unitaryhack"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":40,"issueCount":237,"starsCount":477,"forksCount":174,"license":"Other","participation":[13,24,8,21,16,16,19,30,8,9,18,11,12,3,3,9,11,14,8,21,14,13,16,25,26,22,13,11,7,11,11,13,19,21,6,7,7,22,16,12,26,14,26,30,19,20,13,21,19,9,10,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T23:45:07.613Z"}},{"type":"Public","name":"cccl","owner":"NVIDIA","isFork":false,"description":"CUDA Core Compute Libraries","allTopics":["cpp","hpc","gpu","modern-cpp","parallel-computing","cuda","nvidia","gpu-acceleration","cuda-kernels","gpu-computing","parallel-algorithm","parallel-programming","nvidia-gpu","gpu-programming","cuda-library","cpp-programming","cuda-programming","accelerated-computing","cuda-cpp"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":55,"issueCount":762,"starsCount":1138,"forksCount":135,"license":"Other","participation":[24,29,48,30,17,19,26,14,12,14,6,11,6,1,1,4,13,11,3,8,15,12,20,20,10,7,9,7,17,18,4,11,22,15,7,9,18,22,15,10,14,18,14,20,31,19,18,15,25,25,14,16],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T23:33:16.778Z"}},{"type":"Public","name":"earth2studio","owner":"NVIDIA","isFork":false,"description":"Open-source deep-learning framework for exploring, building and deploying AI weather/climate workflows.","allTopics":["weather","ai","deep-learning","climate-science"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":76,"forksCount":24,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,13,15,6,2,9,2,3,6,3,3,4,5,0,5,3,4,3,2,1,2,4,2,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T23:31:02.443Z"}},{"type":"Public","name":"Megatron-LM","owner":"NVIDIA","isFork":false,"description":"Ongoing research training transformer models at scale","allTopics":["transformers","model-para","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":138,"issueCount":133,"starsCount":10013,"forksCount":2256,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T23:12:51.114Z"}},{"type":"Public","name":"MatX","owner":"NVIDIA","isFork":false,"description":"An efficient C++17 GPU numerical computing library with Python-like syntax","allTopics":["hpc","gpu","cuda","gpgpu","gpu-computing"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":6,"issueCount":38,"starsCount":1189,"forksCount":82,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[3,4,2,1,4,4,5,0,4,1,2,9,5,1,5,12,8,1,1,0,3,3,0,0,7,3,3,4,1,1,2,4,0,6,4,21,7,3,1,8,1,6,2,8,1,5,9,6,8,1,6,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:58:24.877Z"}},{"type":"Public","name":"GenerativeAIExamples","owner":"NVIDIA","isFork":false,"description":"Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.","allTopics":["microservice","gpu-acceleration","nemo","tensorrt","rag","triton-inference-server","large-language-models","llm","llm-inference","retrieval-augmented-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":21,"starsCount":2149,"forksCount":422,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,3,2,0,1,2,1,3,0,2,2,0,1,2,2,0,5,1,0,0,1,2,1,2,1,1,0,0,1,0,0,0,0,1,0,1,3,0,5,4,2,2,10,23,4,3,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:54:46.891Z"}},{"type":"Public","name":"TransformerEngine","owner":"NVIDIA","isFork":false,"description":"A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.","allTopics":["python","machine-learning","deep-learning","gpu","cuda","pytorch","jax","fp8"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":35,"issueCount":140,"starsCount":1819,"forksCount":303,"license":"Apache License 2.0","participation":[11,7,12,5,10,1,8,6,6,6,11,5,8,0,5,9,6,15,11,10,4,7,5,10,4,6,2,13,4,14,8,13,4,8,9,9,5,26,13,6,6,10,8,13,9,9,21,12,7,11,6,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:54:34.332Z"}},{"type":"Public","name":"NeMo-Run","owner":"NVIDIA","isFork":false,"description":"A tool to configure, launch and manage your machine learning experiments.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":11,"starsCount":27,"forksCount":4,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:35:01.083Z"}},{"type":"Public","name":"NeMo-Aligner","owner":"NVIDIA","isFork":false,"description":"Scalable toolkit for efficient model alignment","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":30,"issueCount":56,"starsCount":517,"forksCount":56,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:19:01.352Z"}},{"type":"Public","name":"dcgm-exporter","owner":"NVIDIA","isFork":false,"description":"NVIDIA GPU metrics exporter for Prometheus leveraging DCGM","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":9,"issueCount":59,"starsCount":853,"forksCount":152,"license":"Apache License 2.0","participation":[0,0,1,2,2,0,2,2,0,0,0,3,12,1,3,4,0,1,4,6,7,10,10,2,2,9,1,4,0,0,0,0,1,0,2,0,0,1,0,1,0,0,2,1,0,0,0,1,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:11:36.216Z"}},{"type":"Public","name":"aistore","owner":"NVIDIA","isFork":false,"description":"AIStore: scalable storage for AI applications","allTopics":["kubernetes","sds","erasure-coding","object-storage","software-defined","multiple-backends","batch-jobs","distributed-shuffle","linear-scalability","etl-offload","network-of-clusters"],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":0,"issueCount":4,"starsCount":1231,"forksCount":166,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:09:19.537Z"}},{"type":"Public","name":"TorchFort","owner":"NVIDIA","isFork":false,"description":"An Online Deep Learning Interface for HPC programs on NVIDIA GPUs","allTopics":["deep-learning","fortran","pytorch","libtorch"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":150,"forksCount":15,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T21:35:31.937Z"}},{"type":"Public","name":"cudnn-frontend","owner":"NVIDIA","isFork":false,"description":"cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":15,"starsCount":423,"forksCount":84,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T21:22:06.472Z"}},{"type":"Public","name":"VisRTX","owner":"NVIDIA","isFork":false,"description":"NVIDIA OptiX based implementation of ANARI","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":5,"starsCount":239,"forksCount":25,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T21:18:32.969Z"}},{"type":"Public","name":"NeMo-text-processing","owner":"NVIDIA","isFork":false,"description":"NeMo text processing for ASR and TTS","allTopics":["text-normalization","inverse-text-n"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":4,"starsCount":266,"forksCount":85,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T21:14:05.028Z"}},{"type":"Public","name":"torch-harmonics","owner":"NVIDIA","isFork":false,"description":"Differentiable signal processing on the sphere for PyTorch","allTopics":["machine-learning","signal-processing","sphere","pytorch"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":1,"issueCount":0,"starsCount":364,"forksCount":28,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T21:04:45.527Z"}},{"type":"Public","name":"stdexec","owner":"NVIDIA","isFork":false,"description":"`std::execution`, the proposed C++ framework for asynchronous and parallel programming.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":12,"issueCount":86,"starsCount":1520,"forksCount":155,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:54:51.493Z"}},{"type":"Public","name":"cloud-native-docs","owner":"NVIDIA","isFork":false,"description":"Documentation repository for NVIDIA Cloud Native Technologies","allTopics":["kubernetes","containers","kubernetes-operator"],"primaryLanguage":{"name":"CSS","color":"#563d7c"},"pullRequestCount":8,"issueCount":3,"starsCount":15,"forksCount":18,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:41:29.286Z"}},{"type":"Public","name":"spark-rapids-jni","owner":"NVIDIA","isFork":false,"description":"RAPIDS Accelerator JNI For Apache Spark","allTopics":[],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":8,"issueCount":76,"starsCount":36,"forksCount":64,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:35:53.832Z"}},{"type":"Public","name":"BioNeMo","owner":"NVIDIA","isFork":false,"description":"BioNeMo NIMs example notebooks: for optimized inference at scale","allTopics":[],"primaryLanguage":null,"pullRequestCount":1,"issueCount":1,"starsCount":81,"forksCount":12,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:22:13.207Z"}},{"type":"Public","name":"k8s-nim-operator","owner":"NVIDIA","isFork":false,"description":"An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":11,"issueCount":0,"starsCount":19,"forksCount":7,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:12:42.072Z"}},{"type":"Public","name":"ais-k8s","owner":"NVIDIA","isFork":false,"description":"Kubernetes Operator, ansible playbooks, and production scripts for large-scale AIStore deployments on Kubernetes.","allTopics":[],"primaryLanguage":{"name":"Go","color":"#00ADD8"},"pullRequestCount":0,"issueCount":0,"starsCount":67,"forksCount":21,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:09:46.652Z"}},{"type":"Public","name":"NeMo-Curator","owner":"NVIDIA","isFork":false,"description":"Scalable data pre processing and curation toolkit for LLMs","allTopics":["python","data","data-processing","data-preparation","deduplication","data-quality","data-curation","data-prep","fine-tuning","fast-data-processing","data-processing-pipelines","datacuration","large-language-models","llm","llmapps","large-scale-data-processing","datarecipes","semantic-deduplication","llm-data-quality"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":14,"issueCount":51,"starsCount":462,"forksCount":55,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T20:04:34.893Z"}},{"type":"Public","name":"nv-ingest","owner":"NVIDIA","isFork":false,"description":"NVIDIA Ingest is a set of microservices for parsing hundreds of thousands of complex, messy unstructured PDFs and other enterprise documents into metadata and text to embed into retrieval systems.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":30,"starsCount":35,"forksCount":10,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T19:35:57.818Z"}},{"type":"Public","name":"egl-wayland","owner":"NVIDIA","isFork":false,"description":"The EGLStream-based Wayland external platform","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":4,"issueCount":32,"starsCount":293,"forksCount":47,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T19:22:39.156Z"}}],"repositoryCount":509,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"NVIDIA repositories"}