Highlights
Stars
Contains the official specification for the Serverless Workflow Domain Specific Language. It provides detailed guidelines and standards for defining, executing, and managing workflows in serverless…
🌀 Experience tranquillity while browsing the web without people tracking you!
🤯 Lobe Chat - an open-source, modern-design AI chat framework. Supports Multi AI Providers( OpenAI / Claude 3 / Gemini / Ollama / Qwen / DeepSeek), Knowledge Base (file upload / knowledge managemen…
Backup and migrate Kubernetes applications and their persistent volumes
High Performance Web Platform Based on Nginx and LuaJIT
GeeseFS-based CSI for mounting S3 buckets as PersistentVolumes
Connect your devices into a secure WireGuard®-based overlay network with SSO, MFA and granular access controls.
A reactive notebook for Python — run reproducible experiments, execute as a script, deploy as an app, and version with git.
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
This repository contains a LLM benchmark for the social deduction game `Resistance Avalon'
End-to-end stack for WebRTC. SFU media server and SDKs.
Quill is a modern WYSIWYG editor built for compatibility and extensibility
Implementation of the Transformer variant proposed in "Transformer Quality in Linear Time"
Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.
An extremely fast Python package and project manager, written in Rust.
A configuration as code language with rich validation and tooling.
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
ClearML - Auto-Magical CI/CD to streamline your AI workload. Experiment Management, Data Management, Pipeline, Orchestration, Scheduling & Serving in one MLOps/LLMOps solution
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
中文LLaMA-2 & Alpaca-2大模型二期项目 + 64K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs with 64K long context models)
使用peft库,对chatGLM-6B/chatGLM2-6B实现4bit的QLoRA高效微调,并做lora model和base model的merge及4bit的量化(quantize)。
AirLLM 70B inference with single 4GB GPU
QLoRA: Efficient Finetuning of Quantized LLMs