Skip to content
#

qat

Here are 29 public repositories matching this topic...

Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.

  • Updated Feb 20, 2025
  • Python

mi-optimize is a versatile tool designed for the quantization and evaluation of large language models (LLMs). The library's seamless integration of various quantization methods and evaluation techniques empowers users to customize their approaches according to specific requirements and constraints, providing a high level of flexibility.

  • Updated Nov 28, 2024
  • Python

The project delivers a comprehensive full-stack solution for the Intel® Enterprise AI Foundation on the OpenShift platform to provision Intel AI and Xeon accelerators, integrate AI software, and enable key AI workloads, such as LLM inferencing and fine-tuning for enterprise AI. GPU network provisioning is currently in the planning stage.

  • Updated Feb 20, 2025
  • Python

Improve this page

Add a description, image, and links to the qat topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the qat topic, visit your repo's landing page and select "manage topics."

Learn more