Large Language Models
Continuing with my readings of ChatGPT and Google Colab: The Easiest, Quickest Way to Start Learning AI by Dr. Kelsey, we get to a chapter dedicated solely to LLMs. Dr. Kelsey explains that "LLMs are artificial intelligence m models that are designed to understand and generate text." LLMs actually form part of the Artificial Neural Networks (ANNs) family we touched in my blog post What Even is Machine Learning? and in more detail in Using AI to Explore Deep Learning. Examples of LLMs I have used just in this class are ChatGPT and DALL-E.
LLMs are a type of Artificial Neural Networks and Deep Learning. To try and keep things straight, I've simplified the levels under AI that highlight the path to LLMs here:
We've already discussed Deep Learning in a previous post, so we will continue onto the next level: Transformer Networks, or Transformers.
I asked ChatGPT again to describe Transformer Networks in simple terms. What I understood is that Transformers are similar to speed reading on asteroids. Transformers models intake data and information and determine the words, sentences, and paragraphs most important to the story, but instead of going word by word and line by line, Transformers can analyze several paragraphs at a time. ChatGPT used the following metaphor to describe Transformers and I think it really puts everything into perspective:
It's like having a smart highlighter that marks which words in a sentence are most important for understanding the overall meaning.Dr. Kelsey highlights three main challenges LLMs face today: computational costs, data quality, and ethical concerns. I own a Mac and I do most of my coding work in R in my Mac. I wanted to share a screenshot of an error code I got while trying to create a Decision Tree algorithm.
This is a Mac... Please engage with Apple and ask them for support... After several years of many reports of installation problems on Mac, it's time to gingerly point out that there have been no similar problems on Windows or Linux.
Needless to say, developers definitely don't use Mac computers! Jokes aside, the computer power needed to train and maintain these models does not come cheap. Not to mention the high energy consumption these models take plus the carbon footprint this implies.
Comments
Post a Comment