|
| 1 | +# TAPIR: Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning |
| 2 | + |
| 3 | + |
| 4 | + |
| 5 | + |
| 6 | +## 📖Introduction |
| 7 | + |
| 8 | +Our paper "Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning" introduces a framework called Task-Aware Curriculum Planning for Instruction Refinement (TAPIR). TAPIR is designed to improve the instruction-following capabilities of large language models (LLMs) by addressing the challenges of task distribution and instruction difficulty during training. The framework uses an oracle LLM to select difficult instructions for a student LLM and adjusts task distributions to balance the student's capabilities. TAPIR also incorporates curriculum planning to escalate task difficulty levels progressively. |
| 9 | + |
| 10 | +## 🧠Models |
| 11 | + |
| 12 | +Download Tapir 7B: |
| 13 | + |
| 14 | +``` |
| 15 | +bash dl_tapir_7B.sh |
| 16 | +``` |
| 17 | + |
| 18 | +**Please use official Llama2 template:** |
| 19 | + |
| 20 | +>[INST] \<\<SYS>> {{ .System }} \<\</SYS>> |
| 21 | +> |
| 22 | +>{{ .Prompt }} |
| 23 | +> |
| 24 | +>[/INST] |
| 25 | + |
| 26 | +## 🗃️Data |
| 27 | + |
| 28 | +Download Tapir_Instruct_70k Dataset: |
| 29 | + |
| 30 | +https://atp-modelzoo-sh.oss-cn-shanghai.aliyuncs.com/release/tutorials/TAPIR-Distillation/Tapir_Instruct.json |
| 31 | + |
| 32 | +## 📜 Citation |
| 33 | + |
| 34 | +If you find our work helpful, please cite it! |
| 35 | + |
| 36 | +``` |
| 37 | +@misc{TAPIR, |
| 38 | + title={Distilling Instruction-following Abilities of Large Language Models with Task-aware Curriculum Planning}, |
| 39 | + author={Yuanhao Yue and Chengyu Wang and Jun Huang and Peng Wang}, |
| 40 | + year={2024}, |
| 41 | + eprint={2405.13448}, |
| 42 | + archivePrefix={arXiv}, |
| 43 | + primaryClass={cs.CL}, |
| 44 | + url={https://arxiv.org/abs/2405.13448}, |
| 45 | +} |
| 46 | +``` |