28 Comments
Dec 11, 2023Liked by Herb Greenberg

I think it's still something of an open question whether or not there's a material advantage to running LLMs locally, but my initial thought here would be no for a few reasons.

First, it's a computationally- and data-intensive computation with utterly trivial network requirements. The query/response is simple plain-text, with the data requirements being the model itself and whatever supporting data may be desirable to flesh out the results or enable further computation. Why shell out top-dollar to compete with tech company wallets for constrained hardware supply when you can pay a subscription fee? Other than the privacy of your question/response, you're not gaining much here and paying a lot of money for it.

Second, running LLM computations on the edge means moving the models themselves to the edge. We're still in the early days, so the models/weights are quite valuable and expensive to produce. If there's very little practical incentive to run the models locally to begin with (see above), why risk something taking the model and running when you can instead just host the model in the cloud?

Third, we're still largely in the "hope" phase of AI/LLMs. There's a lot of hope that this will generate real economic return commensurate with the cost, but I'm not sure to what extent this return has been realized. In my experience it's a net-productivity improvement, but for newer or more junior people I can almost see that reversing since they're not in as good of a position to judge when they're getting a bad answer.

TLDR, I think we're still in the "hope" phase of the technology itself, with a lot of hype and still a mixed bag on real-world return. Added to that, the incentives to push this compute to the edge seems weak at best and likely there's more incentive to keep it off the edge for a while from a data-value perspective. There might be a time for this trade, but I think you're early.

Expand full comment
author

Great note, thanks!

Expand full comment
Dec 12, 2023Liked by Herb Greenberg

Well there's Llama - the open source AI models that are catching the closed LLMs and being used for highly specific training cases. Would that be a reason to ramp compute at the edge?

I'm an AI skeptic but my 4 year old workstation machine's XEON chips and once overabundance of DRAM [32GB] is starting to groan under the weight of ever heavier software and higher interactive loads.

Could the replacement cycle boom - AI or no? THANK YOU FOR ANY THOUGHTS!

Expand full comment
author

Well, arguably AI is a better marketing spin than "gaming"...

Expand full comment

To support the "marketing spin" angle, let me share some personal understanding here. It appears that even a currently available consumer-grade pc like the M1 16GB MacBook Air is not bad at running the Mistral 7B model locally (e.g., using @LMStudioAI https://lmstudio.ai). See the experiences shared in these tweets: https://twitter.com/chakkaradeep/status/1727574808240824455 and https://twitter.com/skirano/status/1727453513670709721 .

Mistral 7B is an open-source trimmed-down version of the Llama 2 model (an alternative to chatGPT) that can be fitted into an ordinary pc to do inferencing (i.e., answering users' questions, which is far, far less demanding than training an LLM). IMAO, the current development of such open-source LLMs is progressing at lightning speed. For those who know how this works, they don't need to upgrade their current pcs to do fantanstic things.

But for ordinary consumers, there might be a FOMO if they don't follow others to upgrade their pcs when everyone is talking about AI. At the end of the day, perhaps behaviors move stock prices more than fundamentals do.

Expand full comment

Just out now on the open-source "new Mistral 8x7B model outperforming gpt-3.5 and llama2 70B. "

https://twitter.com/jerryjliu0/status/1734388929003139175

Expand full comment

Totally agree. Was this shift in perception to the edge just based on AMD's comments? What will a $3,000 AI PC even provide?

Expand full comment

AI is like the oh boy moment, I can crunch a spreadsheet on my PC. Did nothing for productivity. And costs are a lot lower, considering that first computer might have cost you a grand in 1990's dollars. AI is going to start a data stampede. Ask Bing a question and he goes to the same sources you already knew. Its' faster (and less critical). Ultimately are you better at picking stocks (or race horses - James Quinn wrote a book on that). So if you thought information overload was a problem, get ready. Bing will help you sort through the data? (see lack of critical parsing of sources), same old garbage in garbage out admonitions. And Bill Gates will get rich all over again.

Expand full comment

Haha, AI is good boy with bad attitude, hope will come out with good attitude one day!

Expand full comment
Dec 11, 2023Liked by Herb Greenberg

I want to see a side by side video of a PC and an "AI PC" running and someone explain and demo the difference, not just jargon. Searching BestBuy does not turn up anything significant. So far it just looks like a hype term. Where are the actual products?

Expand full comment

"local large language models",

aka "small large language models"

aka anyone else see the contradiction?

Expand full comment
Dec 12, 2023Liked by Herb Greenberg

Could narrower data sets may eliminate the hallucination effects and improve on the corrosion of large public data training?

Expand full comment

Quite unlikely, for a few reasons. One is that the *size* if the yraining sets are what gives it the illusion of intelligence. For another, if we already know how to reduce the data set to "eliminate the hallucination effect", the what the heck do we need the LLM for?

Expand full comment

Perhaps retrieval-augmented generation (RAG) is a way to mitigate hallucination.

See some offerings by solution providers like these:

https://twitter.com/Gradient_AI_/status/1729189923809554723

https://twitter.com/ecardenas300/status/1713577279975051429

Expand full comment

Yes it appears contradictory. To me:

- large language models = 100s billions to trillions of parameters

- small or local large language models = billions to tens of billions of parameters

I.e. much smaller but still large.

Expand full comment

Alas, you *really* don't understand the scale of resources needed for these LLMs

Expand full comment

Care to elaborate? Resources in terms of RAM, GPU, and memory bandwidth?

Expand full comment

All of the above by many orders of magnitude.

Expand full comment
Dec 11, 2023Liked by Herb Greenberg

Great read on the PC's comeback tour - who knew those old boxes had an encore in them? Thanks for the tech-savvy crystal ball! Keep those insights coming!

Expand full comment
Dec 11, 2023Liked by Herb Greenberg

One day during our conversation, my daughter said, "What will someone do with Software without Hardware? Cloud is a concept that involve data servers construcred with chips, even AI, virtual etc"

Expand full comment

This is huge, something I’ve written about before over at Seeking Alpha. I’d recommend checking out some of Dell’s investor calls where executives spell out in detail a lot of the hardware greenfield in front of the company. Cheers!

Expand full comment
Dec 11, 2023Liked by Herb Greenberg

I went to Dell.com and searched on "AI PC" and got a lot of ordinary PCs. Ask Dell to show you one instead of just talk about it.

Expand full comment

Um, well, yes, you probably won’t get far searching for an AI PC given that no PC maker has marketed an AI PC (as if that would be a thing). An AI capable PC (big difference) would be equipped to handle the computing workload required for on-device processing (aka, edge computing). And please, don’t bother visiting the Dell website and searching for ‘AI capable PC.’

Expand full comment

"would be" - future tense. Just pointing out that they do not exist. No point at getting excited until they exist.

Expand full comment
Dec 12, 2023Liked by Herb Greenberg

Fascinating article if only because AI or not, for the first time in a long time I'm seeing a 4 year old machine wheezing under the relentless weight of nonstop Teams/Zoom calls, screensharing and the nonstop running of various pieces of software pulling 10s of thousands of price updates every second. Can't recall the last time I saw a solid [$3.5k] machine struggle after only 4 years!!

Expand full comment

Sir, I tried a lot in reaching out to the team earlier to take suggestions from the experts of all fields before initiating the widespread, consider an article drafted by an interconnected humain brain's neuron network at the same time with that of inbuilt neural network trained LLMs shifting the focus to write what it shows.

Tested on 50 critical thinkers and newcomers out of which 40 adapted LLM with 30 having similar words, sentences, 10 combined them with their own, 10 are completely different, now question the brain and body development, leisure activity, risk taking, out of box thinking, prediction level, misuse of technology taking upper hand as the saying goes with AI helping to focus on other works, how much and in what way that benefitted all sectors as not all humans are machines to do round robin kinda parallel processing at a time!

It is not good for the managements simply turning away from fundamental core subjects to bridge the gap between institutions and industries since Data constitute Information constitute Knowledge constitute Intelligence to be implemented on Systems.

Do focus on how many objectives it reached so far from the day one it was put into existence and why there is still recession, inconvenience with most of its usage in finance departments of businesses? No comments on "AI PC" as of now.

Expand full comment