Google’s AI Course for Beginners (in 10 minutes)!

YouTube ID: Yq0QkCxoTHM
Source URL: https://www.youtube.com/watch?v=Yq0QkCxoTHM
Status: ✓ Completed

Transcript

Language: en
Provider: ruby_llm_whisper
Segments: 168

View Full Text
If you don't have a technical background, but you still want to learn the basics of artificial intelligence, stick around because we're distilling Google's four -hour AI course for beginners into just 10 minutes. I was initially very skeptical because I thought the course would be too conceptual. We're all about practical tips on this channel. And knowing Google, the course might just disappear after one hour. But I found the underlying concepts actually made me better at using tools like ChatGPT and Google Bard. and cleared up a bunch of misconceptions I didn't know I had about AI, machine learning, and large language models. So starting with the broadest possible question, what is artificial intelligence? It turns out, and I'm so embarrassed to admit I didn't know this, AI is an entire field of study like physics, and machine learning is a subfield of AI, much like how thermodynamics is a subfield of physics. Going down another level, deep learning is a subset of machine learning. of machine learning and deep learning models can be further broken down into something called discriminative models and generative models. Large language models, LLMs, also fall under deep learning and right at the intersection between generative and LLMs is the technology that powers the applications we're all familiar with, ChatGPT and Google Bard. Let me know in the comments if this was news to you as well. Now that we have an understanding of the overall landscape, escape and you see how the different disciplines sit in relation to each other, let's go over the key takeaways you should know for each level. In a nutshell, machine learning is a program that uses input data to train a model. That trained model can then make predictions based on data it has never seen before. For example, if you train a model based on Nike sales data, you can then use that model to predict how well a new shoe from Adidas would sell. based on Adidas sales data. Two of the most common types of machine learning models are supervised and unsupervised learning models. The key difference between the two is supervised models use labeled data and unsupervised models use unlabeled data. In this supervised example, we have historical data points that plot the total bill amount at a restaurant against the tip amount. And here the data is labeled. Blue dot equals the order was picked up. up and yellow dot equals the order was delivered. Using a supervised learning model, we can now predict how much tip we can expect for the next order given the bill amount and whether it's picked up or delivered. For unsupervised learning models, we look at the raw data and see if it naturally falls into groups. In this example, we plotted the employee tenure at a company against their income. We see this group of employees have a relatively high High income to years worked ratio versus this group. We can also see all these are unlabeled data. If they were labeled, we would see male, female, years worked, company function, etc. We can now ask this unsupervised learning model to solve a problem like, if a new employee joins, are they on the fast track or not? If they appear on the left, then yes. If they appear on the right, then no. Pro tip, another big difference between the two models is that after a supervised learning model makes a makes a prediction, it will compare that prediction to the training data used to train that model. And if there's a difference, it tries to close that gap. Unsupervised learning models do not do this. By the way, this video is not sponsored, but it is supported by those of you who subscribe to my paid productivity newsletter on Google Tips. Link in the description if you want to learn more. Now we have a basic grasp of machine learning, it's a good time to talk about deep learning, which is just a type of machine learning. that uses something called artificial neural networks don't worry all you have to know for now is that artificial neural networks are inspired by the human brain and looks something like this layers of nodes and neurons and the more layers there are the more powerful the model and because we have these neural networks we can now do something called semi -supervised learning whereby a deep learning model is trained on a small amount of labeled data data, and a large amount of unlabeled data. For example, a bank might use deep learning models to detect fraud. The bank spends a bit of time to tag or label 5% of transactions as either fraudulent or not fraudulent, and they leave the remaining 95% of transactions unlabeled because they don't have the time or resources to label every transaction. The magic happens when the deep learning model uses the 5% of labeled data to learn the basic concepts of the data. of the task. Okay, these transactions are good and these are bad. Okay, apply those learnings to the remaining 95 % of unlabeled data and using this new aggregate data set, the model makes predictions for future transactions. That's pretty cool. And we're not done because deep learning can be divided into two types, discriminative and generative models. Discriminative models learn from the relationship between labels of data points and only has the ability to classify those data points. data points fraud not fraud for example you have a bunch of pictures or data points you purposely label some of them as cats and some of them as dogs a discriminative model will learn from the label cat or dog and if you submit a picture of a dog it will predict the label for that new data point a dog we finally get to generative AI unlike discriminative models generative models learn about the patterns in the training data Data. Then after they receive some input, for example, a text prompt from us, they generate something new based on the patterns they just learned. Going back to the animal example, the pictures or data points are not labeled as cat or dog, so a generative model will look for patterns. Oh, these data points all have two ears, four legs, a tail, likes dog food and barks. When asked to generate something called a dog, the generative model generates a completely new image based on the patterns it just learned. There's a super simple way to determine if something is generative AI or not. If the output is a number, a classification, spam, not spam, or a probability, it is not generative AI. It is Gen AI when the output is natural language, text or speech, an image or audio. Basically, generative AI generates new samples that are similar to the data it was trained on. Moving on to different generative AI model types, most of us are familiar with text-to -text models like ChatGPT and Google Bard. Other common model types include text-to-image models like MidJourney, Dali, and Stable Diffusion. These can not only generate images, but edit images as well. Text-to-video models, surprise surprise, can generate an... and edit video footage. Examples include Google's image and video, cog video, and the very creatively named make a video. Text to 3D models are used to create game assets and a little known example would be OpenAI's shape E model. And finally, text to task models are trained to perform a specific task. For example, if you type at Gmail, summarize my unread emails, Google Bard will look through your inbox and summarize your unread. read emails. Moving over to large language models, don't forget that LLMs are also a subset of deep learning, and although there is some overlap, LLMs and GenAI are not the same thing. An important distinction is that large language models are generally pre-trained with a very large set of data and then fine-tuned for specific purposes. What does that mean? Imagine you have a pet dog. It can be pre-trained with basic commands like like sit, come, down and stay. It's a good boy and a generalist. But if that same good boy goes on to become a police dog, a guide dog or a hunting dog, they need to receive specific training so they're fine-tuned for that specialist role. A similar idea applies to large language models. They're first pre-trained to solve common language problems like text classification, question answering, document summarization. and tech generation. Then using smaller industry-specific data sets, these LLMs are fine-tuned to solve specific problems in retail, finance, healthcare, entertainment, and other fields. In the real world, this might mean a hospital uses a pre-trained large language model from one of the big tech companies and fine-tunes that model with its own first-party medical data to improve diagnostic accuracy from x-rays and other medical data. called tests. This is a win-win scenario because large companies can spend billions developing general purpose large language models, then sell those LLMs to smaller institutions like retail companies, banks, hospitals who don't have the resources to develop their own large language models, but they have the domain specific data sets to fine tune those models. Pro tip, if you do end up taking the full course, I'll link it down below, it's completely free. When you're taking notes, you can right click on the video player. And copy video URL at the current time so can quickly navigate back to that specific part of the video There are five modules total and get a badge after completing each module the content overall is a bit more in the theoretical side So you definitely want to check out this video on how to master prompting next see you on the next video in the meantime Have a great one

Transcript Segments

00:00 - Speaker A | Watch →
If you don't have a technical background,
00:01 - Speaker A | Watch →
but you still want to learn the basics of artificial intelligence,
00:04 - Speaker A | Watch →
stick around because we're distilling Google's four-hour AI course for beginners into just 10 minutes.
00:10 - Speaker A | Watch →
I was initially very skeptical because I thought the course would be too conceptual.
00:14 - Speaker A | Watch →
We're all about practical tips on this channel.
00:16 - Speaker A | Watch →
And knowing Google,
00:17 - Speaker A | Watch →
the course might just disappear after one hour.
00:19 - Speaker A | Watch →
But I found the underlying concepts actually made me better at using tools like ChatGPT and Google Bard.
00:25 - Speaker A | Watch →
and cleared up a bunch of misconceptions I didn't know I had about AI,
00:30 - Speaker A | Watch →
machine learning,
00:31 - Speaker A | Watch →
and large language models.
00:33 - Speaker A | Watch →
So starting with the broadest possible question,
00:35 - Speaker A | Watch →
what is artificial intelligence?
00:38 - Speaker A | Watch →
It turns out, and I'm so embarrassed to admit I didn't know this,
00:40 - Speaker A | Watch →
AI is an entire field of study like physics, and machine learning is a subfield of AI,
00:47 - Speaker A | Watch →
much like how thermodynamics is a subfield of physics. Going down another level,
00:52 - Speaker A | Watch →
deep learning is a subset of machine learning.
00:54 - Speaker A | Watch →
of machine learning and deep learning models can be further broken down into something called discriminative models and generative models.
01:01 - Speaker A | Watch →
Large language models,
01:02 - Speaker A | Watch →
LLMs, also fall under deep learning and right at the intersection between generative and LLMs is the technology that powers the applications we're all familiar with,
01:13 - Speaker A | Watch →
ChatGPT and Google Bard.
01:15 - Speaker A | Watch →
Let me know in the comments if this was news to you as well.
01:17 - Speaker A | Watch →
Now that we have an understanding of the overall landscape,
01:20 - Speaker A | Watch →
escape and you see how the different disciplines sit in relation to each other,
01:23 - Speaker A | Watch →
let's go over the key takeaways you should know for each level.
01:27 - Speaker A | Watch →
In a nutshell,
01:28 - Speaker A | Watch →
machine learning is a program that uses input data to train a model.
01:33 - Speaker A | Watch →
That trained model can then make predictions based on data it has never seen before.
01:38 - Speaker A | Watch →
For example, if you train a model based on Nike sales data,
01:42 - Speaker A | Watch →
you can then use that model to predict how well a new shoe from Adidas would sell.
01:46 - Speaker A | Watch →
based on Adidas sales data.
01:48 - Speaker A | Watch →
Two of the most common types of machine learning models are supervised and unsupervised learning models.
01:54 - Speaker A | Watch →
The key difference between the two is supervised models use labeled data and unsupervised models use unlabeled data.
02:02 - Speaker A | Watch →
In this supervised example,
02:04 - Speaker A | Watch →
we have historical data points that plot the total bill amount at a restaurant against the tip amount.
02:09 - Speaker A | Watch →
And here the data is labeled.
02:11 - Speaker A | Watch →
Blue dot equals the order was picked up.
02:13 - Speaker A | Watch →
up and yellow dot equals the order was delivered.
02:17 - Speaker A | Watch →
Using a supervised learning model,
02:19 - Speaker A | Watch →
we can now predict how much tip we can expect for the next order given the bill amount and whether it's picked up or delivered. For unsupervised learning models,
02:27 - Speaker A | Watch →
we look at the raw data and see if it naturally falls into groups.
02:31 - Speaker A | Watch →
In this example, we plotted the employee tenure at a company against their income. We see this group of employees have a relatively high
02:39 - Speaker A | Watch →
High income to years worked ratio versus this group.
02:42 - Speaker A | Watch →
We can also see all these are unlabeled data.
02:45 - Speaker A | Watch →
If they were labeled,
02:46 - Speaker A | Watch →
we would see male,
02:47 - Speaker A | Watch →
female,
02:48 - Speaker A | Watch →
years worked,
02:49 - Speaker A | Watch →
company function,
02:50 - Speaker A | Watch →
etc.
02:51 - Speaker A | Watch →
We can now ask this unsupervised learning model to solve a problem like,
02:55 - Speaker A | Watch →
if a new employee joins,
02:57 - Speaker A | Watch →
are they on the fast track or not?
02:58 - Speaker A | Watch →
If they appear on the left,
03:00 - Speaker A | Watch →
then yes.
03:01 - Speaker A | Watch →
If they appear on the right,
03:02 - Speaker A | Watch →
then no.
03:03 - Speaker A | Watch →
Pro tip,
03:03 - Speaker A | Watch →
another big difference between the two models is that after a supervised learning model makes a
03:08 - Speaker A | Watch →
makes a prediction,
03:09 - Speaker A | Watch →
it will compare that prediction to the training data used to train that model.
03:13 - Speaker A | Watch →
And if there's a difference,
03:14 - Speaker A | Watch →
it tries to close that gap.
03:17 - Speaker A | Watch →
Unsupervised learning models do not do this.
03:19 - Speaker A | Watch →
By the way, this video is not sponsored,
03:21 - Speaker A | Watch →
but it is supported by those of you who subscribe to my paid productivity newsletter on Google Tips.
03:26 - Speaker A | Watch →
Link in the description if you want to learn more.
03:28 - Speaker A | Watch →
Now we have a basic grasp of machine learning,
03:31 - Speaker A | Watch →
it's a good time to talk about deep learning,
03:33 - Speaker A | Watch →
which is just a type of machine learning.
03:35 - Speaker A | Watch →
that uses something called artificial neural networks don't worry all you have to know for now is that artificial neural networks are inspired by the human brain and looks something like this layers of nodes and neurons and the more layers there are the more powerful the model and because we have these neural networks we can now do something called semi-supervised learning whereby a deep learning model is trained on a small amount of labeled data
04:01 - Speaker A | Watch →
data, and a large amount of unlabeled data.
04:03 - Speaker A | Watch →
For example, a bank might use deep learning models to detect fraud.
04:07 - Speaker A | Watch →
The bank spends a bit of time to tag or label 5% of transactions as either fraudulent or not fraudulent,
04:13 - Speaker A | Watch →
and they leave the remaining 95% of transactions unlabeled because they don't have the time or resources to label every transaction.
04:21 - Speaker A | Watch →
The magic happens when the deep learning model uses the 5% of labeled data to learn the basic concepts of the data.
04:27 - Speaker A | Watch →
of the task.
04:27 - Speaker A | Watch →
Okay,
04:28 - Speaker A | Watch →
these transactions are good and these are bad.
04:30 - Speaker A | Watch →
Okay,
04:30 - Speaker A | Watch →
apply those learnings to the remaining 95% of unlabeled data and using this new aggregate data set, the model makes predictions for future transactions.
04:40 - Speaker A | Watch →
That's pretty cool.
04:42 - Speaker A | Watch →
And we're not done because deep learning can be divided into two types,
04:45 - Speaker A | Watch →
discriminative and generative models.
04:48 - Speaker A | Watch →
Discriminative models learn from the relationship between labels of data points and only has the ability to classify those data points.
04:56 - Speaker A | Watch →
data points fraud not fraud for example you have a bunch of pictures or data points you purposely label some of them as cats and some of them as dogs a discriminative model will learn from the label cat or dog and if you submit a picture of a dog it will predict the label for that new data point a dog we finally get to generative AI unlike discriminative models generative models learn about the patterns in the training data
05:22 - Speaker A | Watch →
Data.
05:22 - Speaker A | Watch →
Then after they receive some input,
05:24 - Speaker A | Watch →
for example, a text prompt from us, they generate something new based on the patterns they just learned.
05:30 - Speaker A | Watch →
Going back to the animal example,
05:31 - Speaker A | Watch →
the pictures or data points are not labeled as cat or dog,
05:35 - Speaker A | Watch →
so a generative model will look for patterns.
05:38 - Speaker A | Watch →
Oh,
05:38 - Speaker A | Watch →
these data points all have two ears, four legs,
05:41 - Speaker A | Watch →
a tail,
05:42 - Speaker A | Watch →
likes dog food and barks.
05:44 - Speaker A | Watch →
When asked to generate something called a dog,
05:46 - Speaker A | Watch →
the generative model generates a completely new image based on the patterns it just learned.
05:53 - Speaker A | Watch →
There's a super simple way to determine if something is generative AI or not. If the output is a number,
05:59 - Speaker A | Watch →
a classification,
06:00 - Speaker A | Watch →
spam,
06:00 - Speaker A | Watch →
not spam,
06:01 - Speaker A | Watch →
or a probability,
06:02 - Speaker A | Watch →
it is not generative AI.
06:04 - Speaker A | Watch →
It is Gen AI when the output is natural language,
06:08 - Speaker A | Watch →
text or speech,
06:09 - Speaker A | Watch →
an image or audio.
06:11 - Speaker A | Watch →
Basically,
06:12 - Speaker A | Watch →
generative AI generates new samples that are similar to the data it was trained on.
06:18 - Speaker A | Watch →
Moving on to different generative AI model types,
06:21 - Speaker A | Watch →
most of us are familiar with text-to-text models like ChatGPT and Google Bard.
06:26 - Speaker A | Watch →
Other common model types include text-to-image models like MidJourney,
06:29 - Speaker A | Watch →
Dali,
06:30 - Speaker A | Watch →
and Stable Diffusion.
06:31 - Speaker A | Watch →
These can not only generate images,
06:34 - Speaker A | Watch →
but edit images as well.
06:36 - Speaker A | Watch →
Text-to-video models,
06:37 - Speaker A | Watch →
surprise surprise,
06:38 - Speaker A | Watch →
can generate an...
06:39 - Speaker A | Watch →
and edit video footage.
06:40 - Speaker A | Watch →
Examples include Google's image and video,
06:43 - Speaker A | Watch →
cog video,
06:43 - Speaker A | Watch →
and the very creatively named make a video.
06:46 - Speaker A | Watch →
Text to 3D models are used to create game assets and a little known example would be OpenAI's shape E model.
06:53 - Speaker A | Watch →
And finally,
06:54 - Speaker A | Watch →
text to task models are trained to perform a specific task.
06:57 - Speaker A | Watch →
For example,
06:58 - Speaker A | Watch →
if you type at Gmail,
06:59 - Speaker A | Watch →
summarize my unread emails,
07:01 - Speaker A | Watch →
Google Bard will look through your inbox and summarize your unread.
07:05 - Speaker A | Watch →
read emails.
07:05 - Speaker A | Watch →
Moving over to large language models,
07:07 - Speaker A | Watch →
don't forget that LLMs are also a subset of deep learning,
07:11 - Speaker A | Watch →
and although there is some overlap,
07:14 - Speaker A | Watch →
LLMs and GenAI are not the same thing.
07:17 - Speaker A | Watch →
An important distinction is that large language models are generally pre-trained with a very large set of data and then fine-tuned for specific purposes.
07:27 - Speaker A | Watch →
What does that mean? Imagine you have a pet dog.
07:29 - Speaker A | Watch →
It can be pre-trained with basic commands like
07:31 - Speaker A | Watch →
like sit,
07:32 - Speaker A | Watch →
come,
07:32 - Speaker A | Watch →
down and stay.
07:33 - Speaker A | Watch →
It's a good boy and a generalist.
07:35 - Speaker A | Watch →
But if that same good boy goes on to become a police dog,
07:39 - Speaker A | Watch →
a guide dog or a hunting dog,
07:40 - Speaker A | Watch →
they need to receive specific training so they're fine-tuned for that specialist role.
07:47 - Speaker A | Watch →
A similar idea applies to large language models.
07:50 - Speaker A | Watch →
They're first pre-trained to solve common language problems like text classification,
07:55 - Speaker A | Watch →
question answering,
07:57 - Speaker A | Watch →
document summarization.
07:58 - Speaker A | Watch →
and tech generation.
07:59 - Speaker A | Watch →
Then using smaller industry-specific data sets,
08:03 - Speaker A | Watch →
these LLMs are fine-tuned to solve specific problems in retail,
08:08 - Speaker A | Watch →
finance,
08:09 - Speaker A | Watch →
healthcare,
08:10 - Speaker A | Watch →
entertainment,
08:10 - Speaker A | Watch →
and other fields.
08:12 - Speaker A | Watch →
In the real world, this might mean a hospital uses a pre-trained large language model from one of the big tech companies and fine-tunes that model with its own first-party medical data to improve diagnostic accuracy from x-rays and other medical data.
08:26 - Speaker A | Watch →
called tests. This is a win-win scenario because large companies can spend billions developing general purpose large language models,
08:33 - Speaker A | Watch →
then sell those LLMs to smaller institutions like retail companies,
08:37 - Speaker A | Watch →
banks,
08:38 - Speaker A | Watch →
hospitals who don't have the resources to develop their own large language models,
08:43 - Speaker A | Watch →
but they have the domain specific data sets to fine tune those models.
08:48 - Speaker A | Watch →
Pro tip,
08:48 - Speaker A | Watch →
if you do end up taking the full course, I'll link it down below, it's completely free.
08:52 - Speaker A | Watch →
When you're taking notes,
08:53 - Speaker A | Watch →
you can right click on the video player.
08:54 - Speaker A | Watch →
And copy video URL at the current time so can quickly navigate back to that specific part of the video There are five modules total and get a badge after completing each module the content overall is a bit more in the theoretical side So you definitely want to check out this video on how to master prompting next see you on the next video in the meantime
09:14 - Speaker A | Watch →
Have a great one
← Back to videos Edit