Curious, can anyone having 128gb ram macs tell their story - is it usable for coding and running model locally? How does latency compare to say copilot?
A rambly "thinking" model like this is way too slow for coding assistance imo, although maybe it could take on larger assignments than you could get out of a chat or coding model.
One of my friends says that the reason why this and several other important problems haven't been solved is because the problems themselves are just too valuable. Think of all the money flowing around...