Content pfp
Content
@
0 reply
0 recast
0 reaction

Eric Platon pfp
Eric Platon
@ic
Recent Apple silicone with unified memory seems to have a strategic advantage over a traditional CPU/GPU pairing. E.g. M2 Ultra with 192Gb RAM which can be allocated to any activity like ML. And the M3 series is apparently just pushing on this track. Anyone with concrete feedback on all this?
2 replies
0 recast
1 reaction

Eric Platon pfp
Eric Platon
@ic
Personal context is that one finger on ordering a pricey Nvidia-based shared machine, while I am basically giving up on Apple. But as a shared machine, an M3 may be faster and even more cost effective?
1 reply
0 recast
0 reaction

Eric Platon pfp
Eric Platon
@ic
And asking to smart and trusted community here, as the real question is about, well, software support—difficult to get solid feedback without Warpcast.
1 reply
0 recast
0 reaction

vincent pfp
vincent
@pixel
GGML/llama.cpp is riding this Apple machine wave. Not a hardcore tinkerer, haven't tried it, but have seen lots of demos on the repo: https://github.com/ggerganov/llama.cpp
3 replies
0 recast
1 reaction