- cross-posted to:
- technology@hexbear.net
- technology@lemmy.ml
- cross-posted to:
- technology@hexbear.net
- technology@lemmy.ml
You must log in or register to comment.
Very unconfirmed so far but if there was indeed a new, and reasonably good, LLM trained without nvidia chips it would signal that alternatives exist, at least for the smaller models, and hopefully would drive down prices of chips needed for training or using such models by individuals and small groups as well, allowing for a lot more specialization of small LLMs for all sorts of tasks.
It’s worth noting that there are also distributed efforts like Petals that avoid the need to have a big data centre to train and run models https://github.com/bigscience-workshop/petals