Skip to content

Commit

Permalink
Clean
Browse files Browse the repository at this point in the history
  • Loading branch information
Muennighoff committed Aug 2, 2023
1 parent f0e4280 commit 182b4b8
Showing 1 changed file with 14 additions and 4 deletions.
18 changes: 14 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,9 +16,11 @@ This repository provides an overview of all components from the paper [OctoPack:
- [Run](#run)
- [Creation](#creation)
- [Training](#training)
- [Transformers](#transformers)
- [OctoCoder](#octocoder)
- [OctoGeeX](#octogeex)
- [Megatron-LM](#megatron-lm)
- [SantaCoder Finetuning](#santacoder-finetuning)
- [SantaCoder Pretraining](#santacoder-pretraining)
- [Other](#other)
- [Citation](#citation)

<!-- /TOC -->
Expand Down Expand Up @@ -180,15 +182,23 @@ To create HumanEvalPack, we follow these steps:

## Training

### Transformers
### OctoCoder

The finetuning script to create OctoCoder is at `finetuning/finetune.py`. The folder contains a `README.md` with instructions.

### OctoGeeX

TODO: Qinkai

### SantaCoder Finetuning

### Megatron-LM
TODO: Qian

### SantaCoder Pretraining

TODO: Binyuan

### Other

We did not end up using Megatron-LM fine-tuning for the model in the paper, but implemented it nevertheless. Feel free to follow these instructions to use it:

Expand Down

0 comments on commit 182b4b8

Please sign in to comment.