Does anyone here use cloud VMs with GPUs to run AI workloads?
I'm trying to run the new Facebook/Meta LLM called Llama 3.1 and its 405b sized model using ollama.com which in GPU mode requires at least 230 GB of total GPU memory.
I have been trying to launch VMs in various cloud providers, like...