Content pfp
Content
@
https://warpcast.com/~/channel/aichannel
0 reply
0 recast
0 reaction

Kasra Rahjerdi pfp
Kasra Rahjerdi
@jc4p
my new default for at home local llm: https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-gguf -- Google provided quantizing/GGUF
3 replies
3 recasts
55 reactions

christopher pfp
christopher
@christopher
What kind of machine specs?
1 reply
0 recast
8 reactions

Kasra Rahjerdi pfp
Kasra Rahjerdi
@jc4p
12GB RTX 4060 Ti -- that specific GGUF uses 7GB of my VRAM, the 4B quantizes down to 3GB and the 1B to 720MB, i like this one (the 12B quant) cause it answers more than just my direct questions, i use the 1B or 4GB if i need a pipeline for classification or something, this one i talk to the most
1 reply
0 recast
5 reactions

Leeward Bound pfp
Leeward Bound
@leewardbound
alpha, ty, pretty close to aping in on a high end jetson and this informs the decision
0 reply
0 recast
2 reactions