forked from jackmpcollins/Awesome-LLM
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
mac
committed
Mar 14, 2023
1 parent
a50f5f2
commit 0d956f5
Showing
8 changed files
with
175 additions
and
95 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,6 +1,10 @@ | ||
# Acceleration | ||
Acceleration for LLM training and inference. | ||
> Hardware and software acceleration for LLM training and inference | ||
## Papers | ||
- **High-throughput Generative Inference of Large Language Models with a single GPU** (2023-02) Ying Sheng et al. [Paper](https://github.com/FMInference/FlexGen/blob/main/docs/paper.pdf) | [Github](https://github.com/FMInference/FlexGen) | ||
|
||
### 2023 | ||
|
||
- (2023-02) **High-throughput Generative Inference of Large Language Models with a single GPU** Ying Sheng et al. [Paper](https://github.com/FMInference/FlexGen/blob/main/docs/paper.pdf) | [Github](https://github.com/FMInference/FlexGen) | ||
|
||
## Useful Resources |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,16 @@ | ||
# Application | ||
|
||
> Augment LLM in different aspects including faithfulness, expressiveness, domain-specific knowledge etc. | ||
## Papers | ||
|
||
### 2022 | ||
|
||
- (2022-10) **Help me write a poem: Instruction Tuning as a Vehicle for Collaborative Poetry Writing** [paper](https://arxiv.org/abs/2210.13669) | ||
|
||
### 2023 | ||
|
||
- (2023-03) **Mixture of Soft Prompts for Controllable Data Generation** [paper](https://arxiv.org/pdf/2303.01580.pdf) | ||
|
||
## Useful Resources | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,11 @@ | ||
# Augmentation | ||
|
||
## Papers | ||
|
||
### 2023 | ||
|
||
- (2023-01) **REPLUG: Retrieval-Augmented Black-Box Language Models** [paper](https://arxiv.org/abs/2301.12652) | ||
- (2023-02) **Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback** [paper](https://arxiv.org/abs/2302.12813) | ||
- (2023-02) **Augmented Language Models: a Survey** [paper](https://arxiv.org/abs/2302.07842) | ||
|
||
## Useful Resources |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,8 +1,15 @@ | ||
# Prompt Learning | ||
|
||
## Papers | ||
- **Making Pre-trained Language Models Better Few-shot Learners** (202-12) Tianyu Gao et al. [paper](https://arxiv.org/pdf/2012.15723.pdf) | ||
- **Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing** (2021-07) Pengfei Liu et al. [paper](https://arxiv.org/abs/2107.13586) | ||
|
||
### 2020 | ||
|
||
- (2020-12) **Making Pre-trained Language Models Better Few-shot Learners** [paper](https://arxiv.org/pdf/2012.15723.pdf) | ||
|
||
### 2021 | ||
|
||
- (2021-07) **Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing** [paper](https://arxiv.org/abs/2107.13586) | ||
|
||
> A Systematic Survey | ||
## Useful Resources |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters