ramsudharsan75
Apprentice
Recently there has been a lot of buzz in the AI race. With recent launches like Qwen 2.5-coder models demanding around 20 GBs of VRAM and new MCP superpowers, I'm feeling the itch to run these models locally instead of paying subscription fees to services like Github Copilot or Cursor. These services also limit the number of tokens per request thus restricting their full potential.
I know you can use your API keys in extensions like Cline to unlock them, but that would mean you'll shell out a lot of bucks very quickly, hence this discussion. There are a couple of options, either buy a GPU to run them in Windows (5090 32GB) or buy a Mac Studio (36GB). In the former option, it will consume more power raising the electricity bills than the latter (but by how much?) and would need other components thus raising the setup cost but will be able to game on it if needed
.
One question keeps popping up though - will there be any Apple alternative that will have unified memory similar to it by AMD or Intel in the near future? It's already several years now but I'm a bit disappointed by the lack of hardware by non-Apple CPU manufacturers. Let me know your thoughts. Interesting times ahead for sure with the smaller AI models.
I know you can use your API keys in extensions like Cline to unlock them, but that would mean you'll shell out a lot of bucks very quickly, hence this discussion. There are a couple of options, either buy a GPU to run them in Windows (5090 32GB) or buy a Mac Studio (36GB). In the former option, it will consume more power raising the electricity bills than the latter (but by how much?) and would need other components thus raising the setup cost but will be able to game on it if needed

One question keeps popping up though - will there be any Apple alternative that will have unified memory similar to it by AMD or Intel in the near future? It's already several years now but I'm a bit disappointed by the lack of hardware by non-Apple CPU manufacturers. Let me know your thoughts. Interesting times ahead for sure with the smaller AI models.