📌 Recent Posts:
Local AI inferencing or cloud based?
I am trying to figure out where the pendulum for AI LLM inference hosting will swing to. Will we have more cloud service usage of the cloud hyperscalers, or will inferencing be on local devices such as laptops or in corporate datacenter? Here is my thinking. I believe Moore’s law will be relevant for a few more years, implying that an affordable supply of local capacity will grow. I also believe that there is a minimum LLM size to make them functional, in the same way that there is a minimum number of bytes for a decent picture or audible sound. But above a certain size, there is going to be a diminishing return. Depending on the scenario, we are talking about billions to trillions of parameters in the LLM. ...