Content
@
0 reply
0 recast
0 reaction
Timi
@timigod.eth
Increasingly capable local models (like Llama 3.2 1b & 3b) make me very optimistic about inference at the edge - mostly because of cost. However, I worry about how limited they’ll be if Apple continues to disallow any kind of meaningful background processing on iOS.
1 reply
0 recast
0 reaction
Timi
@timigod.eth
I’m playing around with a NotebookLM podcast generator type knock off and even if there was a local tts model that was lightweight enough, the user would still have to wait ~3-5 mins without closing the app for a podcast to be generated.
2 replies
0 recast
0 reaction