-
onnxruntime Public
Forked from microsoft/onnxruntimeONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
C++ MIT License UpdatedJan 25, 2025 -
whisper Public
Forked from openai/whisperRobust Speech Recognition via Large-Scale Weak Supervision
Python MIT License UpdatedDec 21, 2024 -
transformers Public
Forked from huggingface/transformers🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Python Apache License 2.0 UpdatedOct 21, 2024 -
optimum Public
Forked from huggingface/optimum🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
Python Apache License 2.0 UpdatedSep 23, 2024 -
llama.cpp Public
Forked from ggerganov/llama.cppLLM inference in C/C++
C++ MIT License UpdatedMar 1, 2024 -
pytorch Public
Forked from pytorch/pytorchTensors and Dynamic neural networks in Python with strong GPU acceleration
Python Other UpdatedSep 22, 2023 -
TensorRT Public
Forked from NVIDIA/TensorRTNVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applicat…
C++ Apache License 2.0 UpdatedJan 14, 2023 -
AITemplate Public
Forked from facebookincubator/AITemplateAITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.
Python Apache License 2.0 UpdatedOct 28, 2022