Content pfp
Content
@
0 reply
0 recast
0 reaction

𝚐π”ͺ𝟾𝚑𝚑𝟾 pfp
𝚐π”ͺ𝟾𝚑𝚑𝟾
@gm8xx8
Nous Research has released DisTrO, a system of distributed optimizers that significantly reduces inter-GPU communication, allowing for efficient training of large neural networks on slower internet connections. It also minimizes dependency on a single entity for computation, promoting a more secure and equitable environment for training LLMs. pre-training is back bby! https://github.com/NousResearch/DisTrO
2 replies
0 recast
4 reactions

Stephan pfp
Stephan
@stephancill
I'm supporting you through /microsub! 895 $DEGEN (Please mute the keyword "ms!t" if you prefer not to see these casts.)
0 reply
0 recast
1 reaction