Idle thought: I wonder if we'll start seeing "training@home" training runs for open-source LLMs. Anyone care to run some numbers or sanity checks on whether this is possible in principle?
The folding@home project has been hugely successful, reaching at least exaFLOPS of compute.
"Training@home" would have to efficiently do partial gradient updates on extremely heterogeneous hardware with widely varying network properties; I'm not sure if this has any chance of producing base models competitive with e.g. Llama. In terms of ops alone, a 1 exaFLOPS network would have taken 10^7 seconds = ~half a year to train Llama 70b, and I imagine the costs of distributing jobs to such a network and coordinating on weight updates would make this much more expensive. So, probably not going to be competitive?
JP Addison likes this.
Kevin Gibbons
in reply to Ben Weinstein-Raun • •Ben Weinstein-Raun likes this.
Ben Weinstein-Raun
in reply to Kevin Gibbons • •I would guess that there will be reasons to at least want an LLM trained on an open corpus, whether it's community-trained or not.
Example reasons include ensuring that the model isn't secretly trying to get you to buy McDonalds, and the possibility that companies start releasing un-fine-tunable models.