by KhoomeiK on 3/10/24, 12:40 PM with 28 comments
by kayson on 3/11/24, 1:21 AM
by katzenversteher on 3/11/24, 9:23 AM
I have a PC that is able to run e.g. Mistral Instruct 7B Q4 inference with around 30 token/s.
How (computation and memory) expensive would it be to also run backpropagation in addition to inference?
I'm aware that the models are typically fed with much more and better data than what is typically provided during normal conversations but on the other hand if I could finetune my local model a teeny tiny bit during during / after each conversation I have with it anyways, it would after a while be perfectly customize for me.
I'm also aware that this could be problematic for models that are used by multiple users but my intended use case would be personal use by a single user.
by internet101010 on 3/10/24, 5:45 PM
by potatoman22 on 3/11/24, 7:38 AM
by dennisy on 3/10/24, 7:29 PM
by KhoomeiK on 3/10/24, 5:32 PM
by adawg4 on 3/10/24, 10:54 PM
by zeroq on 3/10/24, 11:24 PM
by 3abiton on 3/10/24, 3:18 PM
by raidicy on 3/10/24, 3:19 PM
by ponderchan on 3/21/24, 8:07 AM
by neodypsis on 3/10/24, 7:51 PM
by SuhanaJabin on 3/11/24, 3:32 PM