Skip to content
View squallliu's full-sized avatar

Block or report squallliu

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
10 stars written in C++
Clear filter

LLM inference in C/C++

C++ 69,475 10,016 Updated Dec 19, 2024

Android real-time display control software

C++ 21,415 2,795 Updated Dec 7, 2024

ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator

C++ 15,029 2,960 Updated Dec 19, 2024

A lightweight library for portable low-level GPU computation using WebGPU.

C++ 3,776 176 Updated Nov 18, 2024

C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)

C++ 2,957 335 Updated Jul 31, 2024

Pure C++ implementation of several models for real-time chatting on your computer (CPU)

C++ 457 36 Updated Dec 19, 2024

Java Bindings for llama.cpp - A Port of Facebook's LLaMA model in C/C++

C++ 318 33 Updated Sep 14, 2024

Public Domain library for rectifying Chinese coordinates

C++ 158 26 Updated May 21, 2024

llama.cpp fork with additional SOTA quants and improved performance

C++ 115 6 Updated Dec 19, 2024

LM inference server implementation based on llama.cpp.

C++ 46 7 Updated Dec 19, 2024