Content pfp
Content
@
https://warpcast.com/~/channel/base-seoul
0 reply
0 recast
0 reaction

Luuu pfp
Luuu
@luuu
#dailychallenge LLM Fine-tuning Methodology - LoRA Tuning, P-Tuning, Instruction Tuning - Pre-trained model gets fine-tuned with task-based data - LoRA is best for parameter-efficient fine-tuning - P-Tuning optimizes prompt engineering - Instruction Tuning focuses on making models better at understanding and executing human instructions - LoRA Tuning& P-Tuning are both PEFT(Parameter-Efficient Fine-Tuning). - This means it doesn’t change the whole parameter but trains additional small parameters. - Instruction Tuning is used to enhance the model itself. - It uses a dataset of ‘instructions and answers’ - While PEFT is a methodology to adjust the model to fit specific tasks, instruction tuning is used for performance enhancement. - PEFT is Tuning. Instruction Tuning is Training
0 reply
0 recast
3 reactions