Stars
The Triton TensorRT-LLM Backend
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
Transformer related optimization, including BERT, GPT
Mesh TensorFlow: Model Parallelism Made Easier
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
分享一下鄙人自己做的一些特效字幕,如有写得不当的地方还请多多包涵。
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Ongoing research training transformer models at scale
A toolkit for developing and comparing reinforcement learning algorithms.
Reference BLEU implementation that auto-downloads test sets and reports a version string to facilitate cross-lab comparisons
An elegant PyTorch deep reinforcement learning library.
Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
This is the code for "OpenAI Five vs DOTA 2 Explained" By Siraj Raval on Youtube
cedrickchee / baselines
Forked from openai/baselinesOpenAI Baselines: high-quality implementations of reinforcement learning algorithms
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX
Tutorials for creating and using ONNX models