What does HackerNews think of petals?
🌸 Run 100B+ language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
I am currently running my 3090 GPU on there to help out, you can check out https://health.petals.dev/
If you have a spare GPU, consider contributing: https://github.com/bigscience-workshop/petals . I am not associated with them.
https://github.com/bigscience-workshop/petals
I don't understand how petals can work though. I thought LLMs were typically quite monolithic.
Colab: https://colab.research.google.com/drive/1uCphNY7gfAUkdDrTx21...
Project description: https://github.com/bigscience-workshop/petals
https://github.com/bigscience-workshop/petals is a project that does this kind of thing for running inference - I tried it out in Google Collab and it seemed to work pretty well.
Model training is much harder though, because it requires a HUGE amount of high bandwidth data exchange between the machines doing the training - way more than is feasible to send over anything other than a local network connection.
[0] https://petals.ml/ [1] https://github.com/bigscience-workshop/petals [2] https://github.com/yandex-research/swarm [3] https://twitter.com/m_ryabinin/status/1625175933492641814
I'm contributing to the project by running a node in my garage with a single RTX 3060ti in it, and you can too: https://github.com/bigscience-workshop/petals
It's early days, but the tech is super promising.
for inference, there is https://github.com/bigscience-workshop/petals
however, both are only in the research phase. start tinkering!
Since my other account is shadow banned for some unexplained reason, I just wanted to mention the petal project. It's an attempt to bittorrent style distribute the load of running these large models. Good luck!