Skip to content

Commit 1a10e49

Browse files
committed
project readme
1 parent a4cd0b9 commit 1a10e49

File tree

3 files changed

+8
-4
lines changed

3 files changed

+8
-4
lines changed

README.md

+3-1
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,9 @@
4646

4747

4848
## News
49-
🔥🔥🔥 [2024/10/29] Our paper [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) has been accepted by EMNLP-2024, which achieves balanced convergence across various tasks.
49+
🔥🔥🔥 [2024/11/01] We released **MFTCoder v0.5** mainly for MFTCoder-accelerate, which is now supporting preference alignment methods like **DPO/RPO/ORPO** in the new **xxpo** module, adding full-parameter continue-training in the additional **mpt** module along with its **offline_tokenization** module, updating selfpaced method to new convergence balance(CoBa) method for MFT in the original **pefts** module.
50+
51+
🔥🔥🔥 [2024/11/01] Our paper [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) has been accepted by EMNLP-2024, which achieves balanced convergence across various tasks.
5052

5153
🔥🔥🔥 [2024/05/20] We released **MFTCoder v0.4**, mainly for MFTCoder-accelerate. It supports **QLoRA + DeepSpeed Zero3** and **QLoRA + FSDP** as options allowing you training very large models. It now supports new models like Qwen2, Qwen2-MoE, Starcoder2, Gemma, etc.
5254

README_cn.md

+3-1
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,9 @@
4545

4646

4747
## 新闻
48-
🔥🔥🔥 [2024/10/29] 我们的论文 [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) 已被 EMNLP 2024 接收,可以实现多任务收敛均衡。
48+
🔥🔥🔥 [2024/11/01] **MFTCoder-v0.5**发布,新增**xxpo**模块支持偏好对齐DPO/RPO/ORPO;新增**mpt****offline_tokenization**模块支持全量参数的加训;在原本的**pefts**模块(MFT)更新selfpaced收敛均衡技术并更名CoBa。
49+
50+
🔥🔥🔥 [2024/11/01] 我们的论文 [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) 已被 EMNLP 2024 接收,可以实现多任务收敛均衡。
4951

5052
🔥🔥🔥 [2024/05/20] **MFTCoder-v0.4**发布。新增支持**QLoRA+ DeepSpeed Zero3**, **QLoRA + FSDP**训练模式,可以更好的支持微调更大的模型,比如Qwen1.5-70B等。新增对Qwen2, Qwen2-MoE, Starcoder2, Gemma等模型的支持。
5153

requirements.txt

+2-2
Original file line numberDiff line numberDiff line change
@@ -3,8 +3,8 @@ pandas==2.2.1
33
torch==2.1.0
44
tensorboard==2.11.0
55
deepspeed==0.14.0
6-
transformers==4.40.2
7-
accelerate==0.28.0
6+
transformers==4.44.2
7+
accelerate==0.31.0
88
peft==0.10.0
99
BitsAndBytes==0.43.0
1010
xformers==0.0.22.post7

0 commit comments

Comments
 (0)