On Sunday, CBS 60 Minutes presented a segment about Khanmigo, an AI tutor that is powered
by LLM technology. It has shown some very impressive results, and the teachers who use it
in their classes have found it very helpful. It doesn't replace teachers. It helps
them by offloading routine testing and tutoring,
https://www.cbsnews.com/video/khanmigo-ai-tutor-60-minutes-video-2024-12-08/
As I have said many times. there are serious limitations to the LLM technology, which
requires evaluation to avoid serious errors and hallucinogenic disasters. Question; How
can Khanmigo and related systems avoid those disasters?
I do not know the details of the Khanmigo implementation. But from the examples they
showed, I suspect that they avoid mistakes by (1) Starting with a large text that was
written, tested, and verified by humans (possibly with some computer aid); (2) For each
topic, the system does Q/A primarily by translation; (3) And the LLM technology was first
developed for translation and Q/A; (4) if the source text is tested and verified, a Q/A
system that is based on that text can usually be very good and dependable.
But the CBS program did show an example where the system made some mistakes.
Summary: This example shows great potential for the LLM technology. But it also shows
the need for evaluation by the traditional AI symbolic methods. Those methods have been
tried and tested for over 50 years, and they are just as important today as they ever
were.
As a reminder: LLMs can be used with a large volume of sources to find information and to
generate hypotheses. But if the source is very large and unverified for accuracy, it can
and does find and generate erroneous or even dangerously false information. That is why
traditional AI methods are essential for evaluating what they find in a large volume of
source data.
Danger; The larger the sources, the more likely that the LLMs will find bad data.
Without evaluation, bigger is definitely not better. I am skeptical about attempts to
create super large volumes of LLM data. Those systems consume enormous amounts of
electricity with a diminishing return on investment.
There is already a backlash by employees of Google and Elon M.
John