On-device training of small LLM LoRA weights #21447
Replies: 1 comment 2 replies
-
On-device training does provide regular finetuning abilities by enabling users to freeze most layers and unfreeze some layers. AFAIK, we haven't yet tested on-device training with finetuning the LoRA weights, but it should be possible with the following:
If you give it a try, I'd be curious to hear how it goes. For your second question, on-device training is currently enabled on CUDA, but not other device GPUs. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Bit of a beginner with the onnx runtime framework, but I noticed that it is one of the rare frameworks that allow on-device retraining of models. So here’s my question:
Does a framework like ONNX runtime offer such a thing for LLM models, by converting from existing transformer models and then finetuning them on device? Instead of re-training the whole model, we would only finetune the LoRA weights.
I understand the hardware restrictions of edge devices, but I would still like to know if such a thing is still feasible with this framework.
Also is it possible that on-device learning can be used with the underlying device GPU (like for Android…)?
Open to any answers, tips or discussion about this. Thank you!
Beta Was this translation helpful? Give feedback.
All reactions