Lastly, I think we need to be careful in what significance we give to seeing the number of trainable parameters going down… You can arbitrarily reduce the number of “trainable parameters” in a model simply by choosing to freeze parts of the model (you just set requires_grad = False on a weight matrix), and I think that’s all that’s happening here. Don’t conflate that with “parameter efficient fine tuning” techniques like LoRA, where you get to train fewer parameters while stil...| Hugging Face Forums
Instruction tuning the OPT-125M model by training it on the Open Assistant Guanaco dataset using Hugging Face Transformers.| DebuggerCafe
Instruction tuning the GPT2 model on the Alpaca dataset using the Hugging Face Transformers library and the SFT Trainer pipine.| DebuggerCafe
Contact DebuggerCafe for Machine Learning, Deep Learning, and AI.| DebuggerCafe
We’re on a journey to advance and democratize artificial intelligence through open source and open science.| huggingface.co
We’re on a journey to advance and democratize artificial intelligence through open source and open science.| huggingface.co
We’re on a journey to advance and democratize artificial intelligence through open source and open science.| huggingface.co