Digitized pfp
Digitized
@digitized
1/ Nous Research is advancing decentralized AI. Last month, they announced the pre-training of a 15B parameter language model.
1 reply
0 recast
0 reaction

Digitized pfp
Digitized
@digitized
2/ Unlike conventional AI, the model was trained over the internet leveraging a network of hardware contributed by Oracle, Lambda API, Northern Data Group, Crusoe Cloud, and the Andromeda Cluster.
1 reply
0 recast
0 reaction

Digitized pfp
Digitized
@digitized
3/ Their approach uses Nous DisTrO, a distributed system that consolidates varied GPU resources over the internet. Early results show a loss curve and convergence rate that “meets or exceeds” centralized training.
1 reply
0 recast
0 reaction

Digitized pfp
Digitized
@digitized
4/ You can follow and watch the training run live on their website: https://t.co/xspVGkccDd
1 reply
0 recast
0 reaction

Digitized pfp
Digitized
@digitized
5/ The underlying research, DeMo, laid the groundwork for Nous DisTrO. Explore it here: 🔹Paper Link: https://arxiv.org/abs/2411.19870 🔹Code: bloc97/DeMo: https://github.com/bloc97/DeMo
1 reply
0 recast
0 reaction

Digitized pfp
Digitized
@digitized
6/ Key to Nous DisTrO is: 🔹 A networking stack that drastically cuts down inter-GPU communication–up to 10,000x less data transfer; and 🔹 Psyche, a decentralized network that autonomously coordinates compute for large-scale training.
1 reply
0 recast
0 reaction