- Tokyo, Japan
⚙️ MLOps
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)
An Engine-Agnostic Deep Learning Framework in Java
[Unmaintained, see README] An ecosystem of Rust libraries for working with large language models
MLOps Tools For Managing & Orchestrating The Machine Learning LifeCycle
Superduper: Build end-to-end AI applications and agent workflows on your existing data infrastructure and preferred tools - without migrating your data.
⚗️ Instill Model contains components for AI model orchestration
prompt2model - Generate Deployable Models from Natural Language Instructions
The Virtual Feature Store. Turn your existing data infrastructure into a feature store.
Kubernetes-friendly ML model management, deployment, and serving.
SkyPilot: Run AI and batch jobs on any infra (Kubernetes or 12+ clouds). Get unified execution, cost savings, and high GPU availability via a simple interface.
A series of Terraform based recipes to provision popular MLOps stacks on the cloud.
A curated list of awesome open source libraries to deploy, monitor, version and scale your machine learning
Sacred is a tool to help you configure, organize, log and reproduce experiments developed at IDSIA.
😎 A curated list of awesome MLOps tools
JVector: the most advanced embedded vector search engine
The open source Solver AI for Java, Python and Kotlin to optimize scheduling and routing. Solve the vehicle routing problem, employee rostering, task assignment, maintenance scheduling and other pl…
An open-source ML pipeline development platform
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable,…
Run any open-source LLMs, such as Llama, Mistral, as OpenAI compatible API endpoint in the cloud.
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
Running large language models on a single GPU for throughput-oriented scenarios.
A voice-enabled chatbot application built using of 🦜️🔗 LangChain, text-to-speech, and speech-to-text models from 🤗 Hugging Face, and 🍱 BentoML.
Interact with your documents using the power of GPT, 100% privately, no data leaks
Large Language Model Text Generation Inference