Not what I am suggesting. However, having trained a few different things on a modest M4 Pro chip (so not even their absolute most powerful chips mind you), and using it for local-first AI inference, I can see the value. A single server could serve an LLM for a small business and cost a lot less than running the same inference through a 5090 in terms of power usage.
I could also see universities giving this type of compute access to students for cheaper to work on more basic less resource intensive models.
I could also see universities giving this type of compute access to students for cheaper to work on more basic less resource intensive models.