Highlights
- Pro
Stars
code for EMNLP 2024 paper: Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis
GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)
A library for mechanistic interpretability of GPT-style language models
Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
Training Sparse Autoencoders on Language Models
[ACL 2024] Unveiling Linguistic Regions in Large Language Models
A framework for few-shot evaluation of language models.
My Nord-themed dotfiles for my Linux Installations (currently NixOS)
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
✱ Interpreting learned feedback patterns in large language models
Official code for the ACL 2024 paper: Chat Vector: A Simple Approach to Equip LLMs with Instruction Following and Model Alignment in New Languages.
Video+code lecture on building nanoGPT from scratch
Interactive TUI for Transformer Model Analysis
KoCommonGEN v2: A Benchmark for Navigating Korean Commonsense Reasoning Challenges in Large Language Models
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
A curated list of Large Language Model (LLM) Interpretability resources.
언어모델을 학습하기 위한 공개 한국어 instruction dataset들을 모아두었습니다.
SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models
The simplest, fastest repository for training/finetuning medium-sized GPTs.