Instructiontuned
Instruction tuning refers to the process of fine-tuning a pre-trained language model on a curated dataset of instruction–response pairs with the aim of improving the model’s ability to follow user instructions. The resulting models are often described as instruction-tuned. This approach is a form of supervised fine-tuning that emphasizes alignment with human intent rather than purely optimizing next-token prediction.
Datasets used for instruction tuning are diverse and task-spanning, covering areas such as summarization, translation, reasoning,
Instruction tuning is related to, but distinct from, reinforcement learning with human feedback (RLHF). In practice,
Limitations include dependence on the quality and breadth of the instruction dataset, potential amplification of biases
The term appears in research on open-domain instruction-following models and is associated with models capable of