Content pfp
Content
@
0 reply
0 recast
0 reaction

𝚐𝔪𝟾𝚡𝚡𝟾 pfp
𝚐𝔪𝟾𝚡𝚡𝟾
@gm8xx8
Announcing Together Inference Engine 2.0 with new Turbo and Lite endpoints “provides decoding throughput 4x faster than open-source vLLM” burr https://www.together.ai/blog/together-inference-engine-2
0 reply
0 recast
0 reaction

TuanNguyen pfp
TuanNguyen
@tuannguyen93
gm
0 reply
0 recast
0 reaction