Merge pull request #369 from bug-orz/master · alibaba/EasyNLP@a4ee956

Original file line numberDiff line numberDiff line change

@@ -0,0 +1,46 @@

1+

# TAPIR: Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning

2+
3+

![image](https://github.com/user-attachments/assets/1c48a0ce-bc47-468a-9762-c952b1494d0e)

4+
5+
6+

## 📖Introduction

7+
8+

Our paper "Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning" introduces a framework called Task-Aware Curriculum Planning for Instruction Refinement (TAPIR). TAPIR is designed to improve the instruction-following capabilities of large language models (LLMs) by addressing the challenges of task distribution and instruction difficulty during training. The framework uses an oracle LLM to select difficult instructions for a student LLM and adjusts task distributions to balance the student's capabilities. TAPIR also incorporates curriculum planning to escalate task difficulty levels progressively.

9+
10+

## 🧠Models

11+
12+

Download Tapir 7B:

13+
14+

```

15+

bash dl_tapir_7B.sh

16+

```

17+
18+

**Please use official Llama2 template:**

19+
20+

>[INST] \<\<SYS>> {{ .System }} \<\</SYS>>

21+

>

22+

>{{ .Prompt }}

23+

>

24+

>[/INST]

25+
26+

## 🗃️Data

27+
28+

Download Tapir_Instruct_70k Dataset:

29+
30+

https://atp-modelzoo-sh.oss-cn-shanghai.aliyuncs.com/release/tutorials/TAPIR-Distillation/Tapir_Instruct.json

31+
32+

## 📜 Citation

33+
34+

If you find our work helpful, please cite it!

35+
36+

```

37+

@misc{TAPIR,

38+

title={Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning},

39+

author={Yuanhao Yue and Chengyu Wang and Jun Huang and Peng Wang},

40+

year={2024},

41+

eprint={2405.13448},

42+

archivePrefix={arXiv},

43+

primaryClass={cs.CL},

44+

url={https://arxiv.org/abs/2405.13448},

45+

}

46+

```