Anatoly, Stephen, Dan, Alex, and every subscriber to these lists,
I want to emphasize two points: (1) I am extremely enthusiastic about LLMs and what they
can and cannot do. (2) I am also extremely enthusiastic about the 60+ years of R & D
in AI technologies and what they have and have not done. Many of the most successful AI
developments are no longer called AI because they have become integral components of
computer science. Examples: compilers, databases, computer graphics, and the interfaces
of nearly every appliance we use today: cars, trucks, airplanes, rockets, telephones,
farm equipment, construction equipment, washing machines, etc. For those things, the AI
technology of the 20th century is performing mission-critical operations with a level of
precision and dependability that unaided humans cannot achieve without their help.
Fundamental principle: For any tool of any kind -- hardware or software -- it's
impossible to understand exactly what it can do until the tool is pushed to the limits
where it breaks. At that point, an examination of the pieces shows where its strengths
and weaknesses lie.
For LLMs, some of the breaking points have been published as hallucinations and humorous
nonsense. But more R & D is necessary to determine where the boundaries are, how to
overcome them, work around them, and supplement them with the 60+ years of other AI
tools.
Anatoly> When you target LLM and ANN as its engine, you should consider that this is
very fast moving target. E.g. consider recent work (and imagine what can be done there in
a year or two in graph-of-thoughts architectures) . . .
Yes, that's obvious. The article you cited looks interesting, and there are many
others. They are certainly worth exploring. But I emphasize the question I asked:
Google and OpenAI have been exploring this technology for quite a few years. What
mission-critical applications have they or anybody else discovered and implemented?
So far the only truly successful applications are in MT -- machine translation of
languages, natural and artificial. Can anybody point to any other applications that are
mission critical for any business or government organization anywhere?
Stephen Young> Yup. My 17yo only managed 94% in his Math exam. He got 6% wrong.
Hopeless - he'll never amount to anything.
The LLMs have been successful in passing various tests at levels that match or surpass the
best humans. But that's because they cheat. They have access to a huge amount of
information on the WWW about a huge range of tests. Bur when they are asked routine
questions for which the answers or the methods for generating answers cannot be found,
they make truly stupid mistakes.
No mission-critical system that guides a car, an airplane, a rocket, or a farmer's
plow can depend on such tools.
Dan Brickley> Encouraging members of this forum to delay putting time into learning how
to use LLMs is doing them no favours. All of us love to feel we can see through hype, but
it’s also a brainworm that means we’ll occasionally miss out on things whose hype is
grounded in substance.
Yes, I enthusiastically agree. We must always ask questions. We must study how LLMs
work, what they do, and what their limitations are. If they cannot solve some puzzle,
it's essential to find out why. Noticing a failure on one problem is not an excuse
for giving up. It's a clue for guiding the search.
Alex> I'm researching how LLMs work. And we will really find out where they will
be used after the hype in 3-5 years.
Yes. But that is when everybody else will have won the big contracts to develop the
mission-critical applications.
Now is the time to do the critical research on where the strengths and limitations are.
Right now, the crowd is having fun building toys that exploit the obvious strengths. The
people who are doing the truly fundamental research are exploring the limitations and how
to get around them.
John
Show replies by date