Lists (8)
Sort Name ascending (A-Z)
Stars
Writing Extension for Text Generation WebUI
Search the web and your self-hosted apps using local AI agents.
Stable Diffusion and Flux in pure C/C++
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
Repo for Hamel's Professional Website
Lazy Predict help build a lot of basic models without much code and helps understand which models works better without any parameter tuning
A community list of common phrases generated by GPT and Claude models
ComfyUI-Manager is an extension designed to enhance the usability of ComfyUI. It offers management functions to install, remove, disable, and enable various custom nodes of ComfyUI. Furthermore, th…
This repository showcases various advanced techniques for Retrieval-Augmented Generation (RAG) systems. RAG systems combine information retrieval with generative models to provide accurate and cont…
A fancy, easy-to-use and reactive self-hosted docker compose.yaml stack-oriented manager
Tika-Python is a Python binding to the Apache Tika™ REST services allowing Tika to be called natively in the Python community.
Multiple NVIDIA GPUs or Apple Silicon for Large Language Model Inference?
WhisperPlus: Faster, Smarter, and More Capable 🚀
ID-based RAG FastAPI: Integration with Langchain and PostgreSQL/pgvector
LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progres…
Script that organizes the Google Takeout archive into one big chronological folder
Connect your devices into a secure WireGuard®-based overlay network with SSO, MFA and granular access controls.
🤯 Lobe Chat - an open-source, modern-design AI chat framework. Supports Multi AI Providers( OpenAI / Claude 3 / Gemini / Ollama / Qwen / DeepSeek), Knowledge Base (file upload / knowledge managemen…
Port of Facebook's LLaMA model in C/C++
Tensor parallelism is all you need. Run LLMs on an AI cluster at home using any device. Distribute the workload, divide RAM usage, and increase inference speed.
A collection of PocketBase community resources.
A highly customizable homepage (or startpage / application dashboard) with Docker and service API integrations.
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory
Easy web analytics. No tracking of personal data.
Turns KoboldAI into a crowdsourced distributed cluster
Drag & drop UI to build your customized LLM flow