Highlights
- Pro
Stars
GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.
zms1999 / SmartMoE
Forked from laekov/fastmoeA MoE impl for PyTorch, [ATC'23] SmartMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)
A library for mechanistic interpretability of GPT-style language models
Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
Training Sparse Autoencoders on Language Models
[ACL 2024] Unveiling Linguistic Regions in Large Language Models
A framework for few-shot evaluation of language models.
My Nord-themed dotfiles for my Linux Installations (currently NixOS)
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
✱ Interpreting learned feedback patterns in large language models
Official code for the ACL 2024 paper: Chat Vector: A Simple Approach to Equip LLMs with Instruction Following and Model Alignment in New Languages.
Video+code lecture on building nanoGPT from scratch
Interactive TUI for Transformer Model Analysis
KoCommonGEN v2: A Benchmark for Navigating Korean Commonsense Reasoning Challenges in Large Language Models
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
A curated list of Large Language Model (LLM) Interpretability resources.
언어모델을 학습하기 위한 공개 한국어 instruction dataset들을 모아두었습니다.
The Universe of Evaluation. All about the evaluation for LLMs.
SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models