Large Language Models is just One Branch of Artificial Intelligence

Superfact 90: Large Language Models (LLMs) such as ChatGPT, Claude, Llama and Gemini are just one type of popular applications of Artificial Intelligence among hundreds of applications of Artificial Intelligence, and LLMs represents just one branch of Artificial Intelligence.

White female AI robot using a microscope in the scientific laboratory
Artificial intelligence and research concept. Shutterstock Asset id: 2314449325 by Stock-Asso

LLMs are currently the most popular “viral” AI. We can all access LLMs in our browsers. This has created the common misconception that Artificial Intelligence is the same as Large Language Models. However, LLMs represent only one branch of narrow AI systems designed to perform specific tasks.

Applications of Artificial Intelligence other than what Large Language Models are used for include robotics, robot motion planning, advanced control systems using AI, self-driving cars, image processing, optical character recognition, classification, facial recognition systems, medical imaging diagnostics, game playing such as chess playing computers, financial fraud detection, cybersecurity, investment robots, route optimization, mathematical proof generation, recommendation algorithms, virtual assistants, programming code generation, smart home devices, drug discovery, and that is just for starters. There are probably many applications and types of Artificial Intelligence that we have not yet invented.

C3P0 and R2D2 from Star Wars
Two Robots powered by Artificial Intelligence. Shutterstock Asset id: 558350728 by Willrow Hood.

LLMs use large neural networks with many hidden layers, so called deep learning algorithms, and they employ the Rumelhart backpropagation learning algorithm invented by David Rumelhart, Geoffrey Hinton, and Ronald Williams. Clearly neural networks with multiple hidden layers and using the Rumelhart backpropagation algorithm are incredibly successful but it is just one of many kinds of Artificial Intelligence algorithms, and who knows what we will see in the future. Related to this post is my previous post Artificial Intelligence is Not New. We have only just begun.

I consider this a super fact because it is true, kind of important, and I believe that the multitude of Artificial Intelligence algorithms and applications is a surprise to many.

The many Artificial Intelligence Algorithms

AI Humanoid Face Concept. Technology Digital Robot Head Side View with Circuit Board Components. Tech Blue Background. Artificial Intelligence Agent or Assistant Concept. Vector Digital Illustration.
Shutterstock Asset id: 2645975149

Due to the great improvement and success of Neural Networks, they have become very popular and Large Language Models use very large Neural Networks with multiple hidden layers (employing the Rumelhart back propagation algorithm). You can read more about that here.

However, there are many other AI algorithms, hundreds, maybe thousands. One example is genetic algorithms. These are types of algorithms that mimics evolution. They iteratively select a set of the best candidate solutions, then combine them (crossover), and also add random changes (mutation) to generate new solutions. Then select the best solutions and then you do it again. Selecting the best solutions corresponds to natural selection. I tried out such algorithms at my work, and over many iterations / generations you can get some impressive results. It is easy to understand how a complex organ such as an eye can evolve in a similar way in nature.

One type of decision tree based machine learning algorithm that I used specifically for classification tasks at work was C4.5 and C5. More specifically I used this type of machine learning algorithm for evaluating the results from automatic mail sorting systems. Basically, how well can a result from a certain machine be trusted. I don’t remember exactly but my classes were something along the line of super reliable, pretty reliable, average, and this result probably sucks. Other examples of this type of machine learning are ID3, Random Forest, Gradient Boosting, and CART. These types of algorithms are still very popular.

One advantage of using decision tree based machine learning over neural networks for the same task is that when a decision has been made you can follow the decision tree backwards and see why a decision / classification was made. In fact, if you have less than 100 parameters you could likely do it over a lunch. When a neural network makes a decision all you have is a large bunch of numbers that were spit out by an algorithm that looped possibly thousands of times and changing all the numbers every time. You can’t backtrack and figure out exactly how a decision was made. You just have to trust the neural network. The advantage of a neural network in this situation is that if it is trained properly, it is likely to have a better result.

Another type of algorithm used in Artificial Intelligence is search algorithms. For robot motion planning I used an algorithm called A* or A-star, which is a very efficient pathfinding algorithm. It comes in dozens of variants and there are hundreds of other types of search algorithms.

These are just a few examples, but there’s also knowledge based agents, AI-agents with reinforcement learning algorithms, algorithms based on Bayes’ Theorem, Vector Machines, Markov Decision Processes, clustering algorithms,  K-nearest neighbor (KNN) algorithm, simulated annealing, hill climbing,  the ant colony optimization algorithm, and of course neural networks and there are also many types of neural networks. I used a relatively unknown form of artificial intelligence called reflex control for my robotics research. The point is, there is zoo of artificial intelligence algorithms out there. Deep learning neural networks are very popular AI algorithms but far from the only ones.

My Personal Experience with Artificial Intelligence

In 1986, when I was in college in Sweden, I took a class in the LISP programming language. LISP was the first Artificial Intelligence programming language, and it was invented in 1958. In 1987, as a university level exchange student, I took a class called Artificial Intelligence at Case Western Reserve University. That same year I also took a class called Pattern Recognition which introduced neural networks to me.

In 1986 a landmark paper was published by David Rumelhart, Geoffrey Hinton, and Ronald Williams introducing the Rumelhart backpropagation algorithm. Geoffrey Hinton received the Nobel Prize in physics in 2024. David Rumelhart and Ronald Williams were both dead and could therefore not receive the Nobel Prize. The Nobel Prize was also given to John J. Hopfield, another pioneer in neural networks. He invented the Hopfield network. You can read more about neural networks and the Nobel Prize in physics in 2024 here.

The Rumelhart backpropagation algorithm was a giant leap forward for neural networks and for Artificial Intelligence and it is the algorithm used by ChatGPT and the other large language models. Geoffrey Hinton is often interviewed in media and often presented as the father of Artificial Intelligence. He is not, but he us arguably partially responsible for the greatest leap forward in neural networks, as well as Artificial Intelligence.

In the pattern recognition class, we used the Rumelhart backpropagation algorithm on a simple neural network to read images with text. Later I did research in the field of Robotics where I implemented various Artificial Intelligence algorithms as mentioned above. I have a PhD in Applied Physics and Electrical Engineering with specialty in Robotics. Later I would use artificial intelligence algorithms in my professional career.

A picture of a large silver colored industrial robot.
I used mostly the seven joint Robotics Research Corporation Robot for my robotics research. The robot was able to detect and avoid colliding with the objects surrounding it. I used echolocation for object detection.

My previous posts on Artificial Intelligence, “Artificial Intelligence is Not New”, and “The Nobel Prize in Physics and Neural Networks”, describe how neural networks work in greater detail.

Note on potential harm of AI

The potential harm of AI is a related and important topic that I did not address in this post and not in my previous post either. However, this is an issue that will be addressed later.




To see the Other Super Facts click here

Unknown's avatar

Author: thomasstigwikman

My name is Thomas Wikman. I am a software/robotics engineer with a background in physics. I am currently retired. I took early retirement. I am a dog lover, and especially a Leonberger lover, a home brewer, craft beer enthusiast, I’m learning French, and I am an avid reader. I live in Dallas, Texas, but I am originally from Sweden. I am married to Claudia, and we have three children. I have two blogs. The first feature the crazy adventures of our Leonberger Le Bronco von der Löwenhöhle as well as information on Leonbergers. The second blog, superfactful, feature information and facts I think are very interesting. With this blog I would like to create a list of facts that are accepted as true among the experts of the field and yet disputed amongst the public or highly surprising. These facts are special and in lieu of a better word I call them super-facts.

5 thoughts on “Large Language Models is just One Branch of Artificial Intelligence”

  1. I am impressed by your vast knowledge and experience, Thomas. Thank you for this eye-opener, I thought that the term “AI” meant only one form of programming and/or technology. Excellent share.

    Like

Leave a reply to Lynette d'Arty-Cross Cancel reply