Skip to content

Commit

Permalink
Merge pull request FlagAI-Open#327 from Anhforth/merge_aquila
Browse files Browse the repository at this point in the history
update aquila
  • Loading branch information
ftgreat authored Jun 9, 2023
2 parents deeae84 + d9d8d00 commit bb260eb
Show file tree
Hide file tree
Showing 6 changed files with 12 additions and 28 deletions.
5 changes: 3 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -274,7 +274,7 @@ for text_pair in test_data:

## LICENSE

The majority of FlagAI is licensed under the [Apache 2.0 license](LICENSE), however portions of the project are available under separate license terms:
The majority of FlagAI is licensed under the [Apache 2.0 license](https://www.apache.org/licenses/LICENSE-2.0), however portions of the project are available under separate license terms:

* Megatron-LM is licensed under the [Megatron-LM license](https://github.com/NVIDIA/Megatron-LM/blob/main/LICENSE)
* GLM is licensed under the [MIT license](https://github.com/THUDM/GLM/blob/main/LICENSE)
Expand All @@ -283,6 +283,7 @@ The majority of FlagAI is licensed under the [Apache 2.0 license](LICENSE), howe


## News
- [9 June 2023] release v1.7.0, Support Aquila [#324](https://github.com/FlagAI-Open/FlagAI/pull/324);
- [31 Mar 2023] release v1.6.3, Support AltCLIP-m18 [#303](https://github.com/FlagAI-Open/FlagAI/pull/303) and AltDiffusion-m18 [#302](https://github.com/FlagAI-Open/FlagAI/pull/302);
- [17 Mar 2023] release v1.6.2, Support application of new optimizers [#266](https://github.com/FlagAI-Open/FlagAI/pull/266), and added a new gpt model name 'GPT2-base-en' for English;
- [2 Mar 2023] release v1.6.1, Support Galactica model [#234](https://github.com/FlagAI-Open/FlagAI/pull/234); BMInf, a low-resource inference package [#238](https://github.com/FlagAI-Open/FlagAI/pull/238), and examples for p-tuning [#227](https://github.com/FlagAI-Open/FlagAI/pull/238)
Expand All @@ -297,7 +298,7 @@ The majority of FlagAI is licensed under the [Apache 2.0 license](LICENSE), howe
## Platforms supported

<div align="center">
<img src="./examples/aquila/img/merged_platform.jpg" height = "100" align=center />
<img src="./examples/Aquila/img/merged_platform.jpg" height = "100" align=center />
</div>


Expand Down
14 changes: 3 additions & 11 deletions README_zh.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,17 +11,8 @@
**FlagAI飞智**是一个快速、易于使用和可扩展的大模型工具包。 我们的目标是支持在多模态的各种下游任务上训练、微调和部署大规模模型。
<br><br>

<p align="center">
已支持平台
</p>

****
天数智芯 Nvidia
****

## 为什么你需要 FlagAI?


1. **可通过 API 快速下载模型**

提供 API 方便你快速下载模型,并在给定(中/英文)文本上使用这些预训练模型,在从[SuperGLUE](https://super.gluebenchmark.com/)[CLUE](https://github.com/CLUEbenchmark/CLUE) benchmarks收集的广泛使用的数据集上对它们进行微调。
Expand Down Expand Up @@ -270,6 +261,7 @@ for text_pair in test_data:


## 动态
- [9 June 2023] 支持 v1.7.0版本, 增加Aquila [#324](https://github.com/FlagAI-Open/FlagAI/pull/324);
- [31 Mar 2023] 支持v1.6.3版本, 增加AltCLIP-m18模型 [#303](https://github.com/FlagAI-Open/FlagAI/pull/303) 以及 AltDiffusion-m18模型 [#302](https://github.com/FlagAI-Open/FlagAI/pull/302);
- [17 Mar 2023] 支持v1.6.2版本, 可以使用新的优化器 [#266](https://github.com/FlagAI-Open/FlagAI/pull/266), 并增加了英文gpt模型GPT2-base-en;
- [2 Mar 2023] 支持v1.6.1版本, 增加Galactica模型 [#234](https://github.com/FlagAI-Open/FlagAI/pull/234), 大模型推理的低资源工具包BMInf [#238](https://github.com/FlagAI-Open/FlagAI/pull/238), 以及P-tuning样例 [#227](https://github.com/FlagAI-Open/FlagAI/pull/238)
Expand All @@ -284,7 +276,7 @@ for text_pair in test_data:
## 许可 LICENSE


FlagAI飞智大部分项目基于 [Apache 2.0 license](LICENSE),但是请注意部分项目代码基于其他协议:
FlagAI飞智大部分项目基于 [Apache 2.0 license](https://www.apache.org/licenses/LICENSE-2.0),但是请注意部分项目代码基于其他协议:

* Megatron-LM 是基于协议 [Megatron-LM license](https://github.com/NVIDIA/Megatron-LM/blob/main/LICENSE)
* GLM 是基于协议 [MIT license](https://github.com/THUDM/GLM/blob/main/LICENSE)
Expand All @@ -293,7 +285,7 @@ FlagAI飞智大部分项目基于 [Apache 2.0 license](LICENSE),但是请注
## 平台支持

<div align="center">
<img src="./examples/aquila/img/merged_platform.jpg" height = "100" align=center />
<img src="./examples/Aquila/img/merged_platform.jpg" height = "100" align=center />
</div>


Expand Down
4 changes: 1 addition & 3 deletions examples/Aquila/Aquila-chat/README.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
license: [Apache License 2.0](https://model.baai.ac.cn/use-agreement)


# AquilaChat

Expand Down Expand Up @@ -153,7 +151,7 @@ def convo_tokenize(convo_obj, tokenizer):
print('-'*80)
print(f"text is {text}")

from examples.aquila.cyg_conversation import default_conversation
from cyg_conversation import default_conversation

conv = default_conversation.copy()
conv.append_message(conv.roles[0], text)
Expand Down
2 changes: 0 additions & 2 deletions examples/Aquila/Aquila-code/README.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
license: [Apache License 2.0](https://model.baai.ac.cn/use-agreement)


# AquilaCode-7B

Expand Down
2 changes: 0 additions & 2 deletions examples/Aquila/Aquila-pretrain/README.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
license: [Apache License 2.0](https://model.baai.ac.cn/use-agreement)


# Aquila

Expand Down
13 changes: 5 additions & 8 deletions examples/Aquila/README.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,3 @@
license: [Apache License 2.0](https://model.baai.ac.cn/use-agreement)


# Aquila

## 简介/Overview
Expand Down Expand Up @@ -46,7 +43,7 @@ Aquila预训练使用了Pile,[RedPajama-Data-1T](https://huggingface.co/datase

The Aquila-7B model was pretrained on Pile,[RedPajama-Data-1T](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T), [Wikipedia](https://huggingface.co/datasets/wikipedia), [C4](https://huggingface.co/datasets/c4), Wudao Corpus、e-book、Patent, encyclopedia, forum, github etc. Details are given in the figure below.

![Screenshot](../img/data_dist.png)
![Screenshot](./img/data_dist.png)



Expand Down Expand Up @@ -82,11 +79,11 @@ bash dist_trigger_docker.sh hostfile Aquila-pretrain-33B.yaml aquila-33b [实验

接下来会输出下列信息,注意`NODES_NUM`应该与节点数相等,`LOGFILE`是模型运行的日志文件;The following information will be output. Note that `NODES_NUM` should be equal to the number of nodes, and `LOGFILE` is the log file for the model run.

![Screenshot](../img/info.jpg)
![Screenshot](./img/info.jpg)

成功训练之前能看到如下信息(具体参数可能不同); Before successful training, you may see the following information with parameters that may differ:

![Screenshot](../img/info2.jpg)
![Screenshot](./img/info2.jpg)

### 2. 可监督微调/Supervised Fine-tuning(SFT)
#### Step 1: 修改参数/Modify Parameters
Expand All @@ -111,11 +108,11 @@ bash dist_trigger_docker.sh hostfile Aquila-sft.yaml aquila-33b [实验名 exper
```
接下来会输出下列信息,注意`NODES_NUM`应该与节点数相等,`LOGFILE`是模型运行的日志文件;The following information will be output. Note that `NODES_NUM` should be equal to the number of nodes, and `LOGFILE` is the log file for the model run.

![Screenshot](../img/info.jpg)
![Screenshot](./img/info.jpg)

成功训练之前能在日志里看到如下信息(具体参数可能不同); Before successful training, you may see the following information in the log file with parameters that may differ:

![Screenshot](../img/info2.jpg)
![Screenshot](./img/info2.jpg)

### 3. 推理/Inference

Expand Down

0 comments on commit bb260eb

Please sign in to comment.