-
Bidirectional-LLM Public
Code for pretraining a Bi-directional Attention Decoder-only Model
-
-
LitterBrother-Xiao.github.io Public
Forked from RayeRen/acad-homepage.github.ioAcadHomepage: A Modern and Responsive Academic Personal Homepage
SCSS MIT License UpdatedMar 25, 2024 -
rayeren.github.io Public
Forked from RayeRen/rayeren.github.ioMy personal homepage
SCSS MIT License UpdatedMar 18, 2024 -
DongwuLLM Public
Forked from dwzq-com-cn/DongwuLLMThis is the codebase for pre-training, compressing, extending, and distilling LLMs with Megatron-LM.
Python Other UpdatedMar 11, 2024 -
Instruct_XMLR Public
[ICLR 2024] Are Bert Family Good Instruction Followers? A Study on Their Potential And Limitations
-
pythia Public
Forked from EleutherAI/pythiaThe hub for EleutherAI's work on interpretability and learning dynamics
Jupyter Notebook Apache License 2.0 UpdatedFeb 22, 2024 -
LLM-Shearing Public
Forked from princeton-nlp/LLM-Shearing[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Python MIT License UpdatedJan 31, 2024 -
-
-
-
Mask-Predict Public
Forked from facebookresearch/Mask-PredictA masked language modeling objective to train a model to predict any subset of the target words, conditioned on both the input text and a partially masked target translation.