1Masalah: Fine-Tuning Model Gede Itu Berat Banget
As we pre-train larger models, full fine-tuning, which retrains all model parameters, becomes less feasible. Using GPT-3 175B as an example – deploying independent instances of fine-tuned models, each with 175B parameters, is prohibitively expensive.
Masalahnya, model AI jaman sekarang makin raksasa. Kalau kita mau ajarin tugas baru pake cara lama (full fine-tuning), kita harus update semua isi otaknya. Bayangin kalau punya banyak tugas, kita harus simpen banyak copy model yang ukurannya bergiga-giga. Boros banget di storage sama duit!
Masalahnya, model AI jaman sekarang makin raksasa. Kalau kita mau ajarin tugas baru pake cara lama (full fine-tuning), kita harus update semua isi otaknya. Bayangin kalau punya banyak tugas, kita harus simpen banyak copy model yang ukurannya bergiga-giga. Boros banget di storage sama duit!
As we pre-train larger models, full fine-tuning, which retrains all model parameters, becomes less feasible. Using GPT-3 175B as an example – deploying independent instances of fine-tuned models, each with 175B parameters, is prohibitively expensive.
Kayak lu punya kamus setebal gaban, terus tiap mau nambahin satu kata baru, lu harus cetak ulang seluruh bukunya dari awal. Capek dan mahal kan?