SlideShare a Scribd company logo
1 of 23
Download to read offline
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
Input example
Output example
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
Task – rephrased
We have a function R784
→ R10
How can we describe this function?
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
Task – rephrased
We have a function R784
→ R10
How can we describe this function?
Machine/deep learning (today’s topic)
tries to answer questions of this type
letting a computer detect patterns in data
Crucial In ML the computer performs tasks without explicit instructions
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
,
▶ Idea Approximate the unknown function R784
→ R10
▶ Neural network = a piecewise linear approximation (matrices + PL maps)
▶ The matrices = a bunch of numbers (weights) and offsets (biases)
▶ The PL maps = usually ReLU
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
Forward
→
Loss
→
Backward
→ → → → ...
▶ Machine learning mantra
▶ Forward = calculate an approximation (start with random inputs)
▶ Loss = compare to real data
▶ Backward = adjust the approximation
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is a neural network (nn)?
▶ NN = a directed graph as above
▶ The task of a nn is to approximate an unknown function
▶ It consist of neurons = entries of vectors, and weights = entries of matrices
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
What is a neural network (nn)?
▶ NN = a directed graph as above
▶ The task of a nn is to approximate an unknown function
▶ It consist of neurons = entries of vectors, and weights = entries of matrices
Example
Here we have two matrices, a 3-by-1 and a 2-by-3 matrix


w
x
y

 and

a11 a12 a13
a21 a22 a23

plus two bias terms as in y = matrix · x + bias
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
What is a neural network (nn)?
▶ NN = a directed graph as above
▶ The task of a nn is to approximate an unknown function
▶ It consist of neurons = entries of vectors, and weights = entries of matrices
Example
Here we have four matrices (plus four biases), whose composition gives a map
R3
→ R4
→ R3
→ R3
→ R2
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
What is a neural network (nn)?
▶ NN = a directed graph as above
▶ The task of a nn is to approximate an unknown function
▶ It consist of neurons = entries of vectors, and weights = entries of matrices
Actually...
we need nonlinear maps as well, say ReLU applied componentwise
Here we have four matrices, whose composition gives a map
R3 ReLU◦matrix
−
−
−
−
−
−
−
→ R4 ReLU◦matrix
−
−
−
−
−
−
−
→ R3 ReLU◦matrix
−
−
−
−
−
−
−
→ R3 ReLU◦matrix
−
−
−
−
−
−
−
→ R2
But ignore that for now
ReLU doesn’t learn anything, its just brings in nonlinearity
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
What is a neural network (nn)?


a1
11 b1
1
a1
12 b1
2
a1
13 b1
3

 ,

a2
11 a2
12 a2
13 b2
1
a2
21 a2
22 a2
23 b2
2

▶ The ak
ij and bk
i are the parameters of our nn
▶ k = number of the layer
▶ Deep = many layers = better approximation
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
What is a neural network (nn)?


a1
11 b1
1
a1
12 b1
2
a1
13 b1
3

 ,

a2
11 a2
12 a2
13 b2
1
a2
21 a2
22 a2
23 b2
2

▶ The ak
ij and bk
i are the parameters of our nn
▶ k = number of the layer
▶ Deep = many layers = better approximation
The point
Many layers → many parameters
These are good for approximating real world problems
Examples
ResNet-152 with 152. layers (used in transformer models such as ChatGPT)
VGG-19 with 19 layers (used in image classification)
GoogLeNet with 22 layers (used in face detection)
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
What is a neural network (nn)?


a1
11 b1
1
a1
12 b1
2
a1
13 b1
3

 ,

a2
11 a2
12 a2
13 b2
1
a2
21 a2
22 a2
23 b2
2

▶ The ak
ij and bk
i are the parameters of our nn
▶ k = number of the layer
▶ Deep = many layers = better approximation
Side fact
Gaming has improved AI!?
A GPU can do e.g. matrix multiplications faster
than a CPU and lots of nn run on GPUs
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
How learning works
▶ Supervised learning Create a dataset with answers, e.g. pictures of
handwritten digits plus their label
▶ There are other forms of learning e.g. unsupervised, which I skip
▶ Split the data into ≈80% training and ≈20% testing data
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Supervised learning Create a dataset with answers, e.g. pictures of
handwritten digits plus their label
▶ There are other forms of learning e.g. unsupervised, which I skip
▶ Split the data into ≈80% training and ≈20% testing data
Idea to keep in mind
How to train students?
There are lectures, exercises etc., the training data
There is a final exam, the testing data
Upon performance, we let them out into the wild
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Forward
Boils down to a bunch of matrix multiplications
followed by the nonlinear activation e.g. ReLU
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Loss
The difference between real values and predictions
Task Minimize loss function
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Backward
This is running gradient descent on the loss function
Slogan Adjust parameters following the steepest descent
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
And what makes it even better:
You can try it yourself
My favorite tool is PyTorch but there are also other methods
Let us see how!
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
Input example
Output example
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
,
▶ Idea Approximate the unknown function R784
→ R10
▶ Neural network = a piecewise linear approximation (matrices + PL maps)
▶ The matrices = a bunch of numbers (weights) and offsets (biases)
▶ The PL maps = usually ReLU
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
Forward
→
Loss
→
Backward
→ → → → ...
▶ Machine learning mantra
▶ Forward = calculate an approximation (start with random inputs)
▶ Loss = compare to real data
▶ Backward = adjust the approximation
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is a neural network (nn)?
▶ NN = a directed graph as above
▶ The task of a nn is to approximate an unknown function
▶ It consist of neurons = entries of vectors, and weights = entries of matrices
Example
Here we have two matrices, a 3-by-1 and a 2-by-3 matrix


w
x
y

 and

a11 a12 a13
a21 a22 a23

plus two bias terms as in y = matrix · x + bias
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
How learning works
▶ Supervised learning Create a dataset with answers, e.g. pictures of
handwritten digits plus their label
▶ There are other forms of learning e.g. unsupervised, which I skip
▶ Split the data into ≈80% training and ≈20% testing data
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Forward
Boils down to a bunch of matrix multiplications
followed by the nonlinear activation e.g. ReLU
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Loss
The difference between real values and predictions
Task Minimize loss function
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Backward
This is running gradient descent on the loss function
Slogan Adjust parameters following the steepest descent
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
There is still much to do...
The mathematics of AI Or: Learning = forward, loss, backward April 2024 5 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
▶ Task Identify handwritten digits
▶ We can see this as a function in the following way:
▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers
▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282
= 784 entries
▶ The output is a vector with 10 entries
▶ We thus have a function R784
→ R10
Input example
Output example
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
,
▶ Idea Approximate the unknown function R784
→ R10
▶ Neural network = a piecewise linear approximation (matrices + PL maps)
▶ The matrices = a bunch of numbers (weights) and offsets (biases)
▶ The PL maps = usually ReLU
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is machine learning?
Forward
→
Loss
→
Backward
→ → → → ...
▶ Machine learning mantra
▶ Forward = calculate an approximation (start with random inputs)
▶ Loss = compare to real data
▶ Backward = adjust the approximation
The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
What is a neural network (nn)?
▶ NN = a directed graph as above
▶ The task of a nn is to approximate an unknown function
▶ It consist of neurons = entries of vectors, and weights = entries of matrices
Example
Here we have two matrices, a 3-by-1 and a 2-by-3 matrix


w
x
y

 and

a11 a12 a13
a21 a22 a23

plus two bias terms as in y = matrix · x + bias
The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
How learning works
▶ Supervised learning Create a dataset with answers, e.g. pictures of
handwritten digits plus their label
▶ There are other forms of learning e.g. unsupervised, which I skip
▶ Split the data into ≈80% training and ≈20% testing data
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Forward
Boils down to a bunch of matrix multiplications
followed by the nonlinear activation e.g. ReLU
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Loss
The difference between real values and predictions
Task Minimize loss function
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
How learning works
▶ Forward Run the nn = function on the training data
▶ Loss Calculate the difference “results - answers” (⇒ loss function)
▶ Backward Change the parameters trying to minimize the loss function
▶ Repeat
Backward
This is running gradient descent on the loss function
Slogan Adjust parameters following the steepest descent
The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
Thanks for your attention!
The mathematics of AI Or: Learning = forward, loss, backward April 2024 5 / 5

More Related Content

Similar to The mathematics of AI (machine learning mathematically)

040 the whole module
040 the whole module040 the whole module
040 the whole moduleedwin caniete
 
WEKA:Algorithms The Basic Methods
WEKA:Algorithms The Basic MethodsWEKA:Algorithms The Basic Methods
WEKA:Algorithms The Basic Methodsweka Content
 
WEKA: Algorithms The Basic Methods
WEKA: Algorithms The Basic MethodsWEKA: Algorithms The Basic Methods
WEKA: Algorithms The Basic MethodsDataminingTools Inc
 
Introduction to Machine Learning for Java Developers
Introduction to Machine Learning for Java DevelopersIntroduction to Machine Learning for Java Developers
Introduction to Machine Learning for Java DevelopersZoran Sevarac, PhD
 
Technology Lesson Plan Assignment: Quadratice Functions
Technology Lesson Plan Assignment: Quadratice FunctionsTechnology Lesson Plan Assignment: Quadratice Functions
Technology Lesson Plan Assignment: Quadratice Functionsdart11746
 
1.1 Real Number Properties
1.1 Real Number Properties1.1 Real Number Properties
1.1 Real Number Propertiessmiller5
 
DeepLearningLecture.pptx
DeepLearningLecture.pptxDeepLearningLecture.pptx
DeepLearningLecture.pptxssuserf07225
 
Introduction to Machine Learning with Spark
Introduction to Machine Learning with SparkIntroduction to Machine Learning with Spark
Introduction to Machine Learning with Sparkdatamantra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
G6 m4-a-lesson 4-t
G6 m4-a-lesson 4-tG6 m4-a-lesson 4-t
G6 m4-a-lesson 4-tmlabuski
 
Deep learning from scratch
Deep learning from scratch Deep learning from scratch
Deep learning from scratch Eran Shlomo
 
Business Mathematics
Business Mathematics Business Mathematics
Business Mathematics Seenleocario
 
Arithmetic instructions
Arithmetic instructionsArithmetic instructions
Arithmetic instructionsLearn By Watch
 
machine_learning.pptx
machine_learning.pptxmachine_learning.pptx
machine_learning.pptxPanchami V U
 
Bahasa Inggris dalam Pembelajaran Matematika
Bahasa Inggris dalam Pembelajaran MatematikaBahasa Inggris dalam Pembelajaran Matematika
Bahasa Inggris dalam Pembelajaran MatematikaNASuprawoto Sunardjo
 

Similar to The mathematics of AI (machine learning mathematically) (20)

040 the whole module
040 the whole module040 the whole module
040 the whole module
 
Guided notes
Guided notesGuided notes
Guided notes
 
Unit5: Learning
Unit5: LearningUnit5: Learning
Unit5: Learning
 
WEKA:Algorithms The Basic Methods
WEKA:Algorithms The Basic MethodsWEKA:Algorithms The Basic Methods
WEKA:Algorithms The Basic Methods
 
WEKA: Algorithms The Basic Methods
WEKA: Algorithms The Basic MethodsWEKA: Algorithms The Basic Methods
WEKA: Algorithms The Basic Methods
 
Introduction to Machine Learning for Java Developers
Introduction to Machine Learning for Java DevelopersIntroduction to Machine Learning for Java Developers
Introduction to Machine Learning for Java Developers
 
Technology Lesson Plan Assignment: Quadratice Functions
Technology Lesson Plan Assignment: Quadratice FunctionsTechnology Lesson Plan Assignment: Quadratice Functions
Technology Lesson Plan Assignment: Quadratice Functions
 
1.1 Real Number Properties
1.1 Real Number Properties1.1 Real Number Properties
1.1 Real Number Properties
 
DeepLearningLecture.pptx
DeepLearningLecture.pptxDeepLearningLecture.pptx
DeepLearningLecture.pptx
 
Introduction to Machine Learning with Spark
Introduction to Machine Learning with SparkIntroduction to Machine Learning with Spark
Introduction to Machine Learning with Spark
 
Lesson 1 matrix
Lesson 1 matrixLesson 1 matrix
Lesson 1 matrix
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
G6 m4-a-lesson 4-t
G6 m4-a-lesson 4-tG6 m4-a-lesson 4-t
G6 m4-a-lesson 4-t
 
Deep learning from scratch
Deep learning from scratch Deep learning from scratch
Deep learning from scratch
 
Integers
IntegersIntegers
Integers
 
Business Mathematics
Business Mathematics Business Mathematics
Business Mathematics
 
Arithmetic instructions
Arithmetic instructionsArithmetic instructions
Arithmetic instructions
 
machine_learning.pptx
machine_learning.pptxmachine_learning.pptx
machine_learning.pptx
 
Bahasa Inggris dalam Pembelajaran Matematika
Bahasa Inggris dalam Pembelajaran MatematikaBahasa Inggris dalam Pembelajaran Matematika
Bahasa Inggris dalam Pembelajaran Matematika
 
Lecture4.pptx
Lecture4.pptxLecture4.pptx
Lecture4.pptx
 

More from Daniel Tubbenhauer

Representation theory of monoidal categories
Representation theory of monoidal categoriesRepresentation theory of monoidal categories
Representation theory of monoidal categoriesDaniel Tubbenhauer
 
Representation theory of algebras
Representation theory of algebrasRepresentation theory of algebras
Representation theory of algebrasDaniel Tubbenhauer
 
Representation theory of monoids
Representation theory of monoidsRepresentation theory of monoids
Representation theory of monoidsDaniel Tubbenhauer
 
Representation theory of monoids and monoidal categories
Representation theory of monoids and monoidal categoriesRepresentation theory of monoids and monoidal categories
Representation theory of monoids and monoidal categoriesDaniel Tubbenhauer
 
Why (categorical) representation theory?
Why (categorical) representation theory?Why (categorical) representation theory?
Why (categorical) representation theory?Daniel Tubbenhauer
 
Some history of quantum groups
Some history of quantum groupsSome history of quantum groups
Some history of quantum groupsDaniel Tubbenhauer
 

More from Daniel Tubbenhauer (11)

Three colors suffice
Three colors sufficeThree colors suffice
Three colors suffice
 
Knots and algebra
Knots and algebraKnots and algebra
Knots and algebra
 
Representation theory of monoidal categories
Representation theory of monoidal categoriesRepresentation theory of monoidal categories
Representation theory of monoidal categories
 
Representation theory of algebras
Representation theory of algebrasRepresentation theory of algebras
Representation theory of algebras
 
Representation theory of monoids
Representation theory of monoidsRepresentation theory of monoids
Representation theory of monoids
 
Representation theory of monoids and monoidal categories
Representation theory of monoids and monoidal categoriesRepresentation theory of monoids and monoidal categories
Representation theory of monoids and monoidal categories
 
Why category theory?
Why category theory?Why category theory?
Why category theory?
 
Why (categorical) representation theory?
Why (categorical) representation theory?Why (categorical) representation theory?
Why (categorical) representation theory?
 
Subfactors in a nutshell
Subfactors in a nutshellSubfactors in a nutshell
Subfactors in a nutshell
 
Temperley-Lieb times four
Temperley-Lieb times fourTemperley-Lieb times four
Temperley-Lieb times four
 
Some history of quantum groups
Some history of quantum groupsSome history of quantum groups
Some history of quantum groups
 

Recently uploaded

Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencyHire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencySheetal Arora
 
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCESTERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCEPRINCE C P
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)Areesha Ahmad
 
Green chemistry and Sustainable development.pptx
Green chemistry  and Sustainable development.pptxGreen chemistry  and Sustainable development.pptx
Green chemistry and Sustainable development.pptxRajatChauhan518211
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfSumit Kumar yadav
 
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 60009654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000Sapana Sha
 
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRStunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRDelhi Call girls
 
Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfSumit Kumar yadav
 
Chromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATINChromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATINsankalpkumarsahoo174
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxgindu3009
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Lokesh Kothari
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSarthak Sekhar Mondal
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoSérgio Sacani
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxkessiyaTpeter
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)Areesha Ahmad
 
Broad bean, Lima Bean, Jack bean, Ullucus.pptx
Broad bean, Lima Bean, Jack bean, Ullucus.pptxBroad bean, Lima Bean, Jack bean, Ullucus.pptx
Broad bean, Lima Bean, Jack bean, Ullucus.pptxjana861314
 
DIFFERENCE IN BACK CROSS AND TEST CROSS
DIFFERENCE IN  BACK CROSS AND TEST CROSSDIFFERENCE IN  BACK CROSS AND TEST CROSS
DIFFERENCE IN BACK CROSS AND TEST CROSSLeenakshiTyagi
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 

Recently uploaded (20)

Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls AgencyHire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
Hire 💕 9907093804 Hooghly Call Girls Service Call Girls Agency
 
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCESTERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)
 
Green chemistry and Sustainable development.pptx
Green chemistry  and Sustainable development.pptxGreen chemistry  and Sustainable development.pptx
Green chemistry and Sustainable development.pptx
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdf
 
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 60009654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
9654467111 Call Girls In Raj Nagar Delhi Short 1500 Night 6000
 
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRStunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
 
Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdf
 
Chromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATINChromatin Structure | EUCHROMATIN | HETEROCHROMATIN
Chromatin Structure | EUCHROMATIN | HETEROCHROMATIN
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptx
 
The Philosophy of Science
The Philosophy of ScienceThe Philosophy of Science
The Philosophy of Science
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on Io
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)
 
Broad bean, Lima Bean, Jack bean, Ullucus.pptx
Broad bean, Lima Bean, Jack bean, Ullucus.pptxBroad bean, Lima Bean, Jack bean, Ullucus.pptx
Broad bean, Lima Bean, Jack bean, Ullucus.pptx
 
DIFFERENCE IN BACK CROSS AND TEST CROSS
DIFFERENCE IN  BACK CROSS AND TEST CROSSDIFFERENCE IN  BACK CROSS AND TEST CROSS
DIFFERENCE IN BACK CROSS AND TEST CROSS
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 

The mathematics of AI (machine learning mathematically)

  • 1.
  • 2. What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
  • 3. What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 Input example Output example The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
  • 4. What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 Task – rephrased We have a function R784 → R10 How can we describe this function? The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
  • 5. What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 Task – rephrased We have a function R784 → R10 How can we describe this function? Machine/deep learning (today’s topic) tries to answer questions of this type letting a computer detect patterns in data Crucial In ML the computer performs tasks without explicit instructions The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
  • 6. What is machine learning? , ▶ Idea Approximate the unknown function R784 → R10 ▶ Neural network = a piecewise linear approximation (matrices + PL maps) ▶ The matrices = a bunch of numbers (weights) and offsets (biases) ▶ The PL maps = usually ReLU The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
  • 7. What is machine learning? Forward → Loss → Backward → → → → ... ▶ Machine learning mantra ▶ Forward = calculate an approximation (start with random inputs) ▶ Loss = compare to real data ▶ Backward = adjust the approximation The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5
  • 8. What is a neural network (nn)? ▶ NN = a directed graph as above ▶ The task of a nn is to approximate an unknown function ▶ It consist of neurons = entries of vectors, and weights = entries of matrices The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 9. What is a neural network (nn)? ▶ NN = a directed graph as above ▶ The task of a nn is to approximate an unknown function ▶ It consist of neurons = entries of vectors, and weights = entries of matrices Example Here we have two matrices, a 3-by-1 and a 2-by-3 matrix   w x y   and a11 a12 a13 a21 a22 a23 plus two bias terms as in y = matrix · x + bias The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 10. What is a neural network (nn)? ▶ NN = a directed graph as above ▶ The task of a nn is to approximate an unknown function ▶ It consist of neurons = entries of vectors, and weights = entries of matrices Example Here we have four matrices (plus four biases), whose composition gives a map R3 → R4 → R3 → R3 → R2 The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 11. What is a neural network (nn)? ▶ NN = a directed graph as above ▶ The task of a nn is to approximate an unknown function ▶ It consist of neurons = entries of vectors, and weights = entries of matrices Actually... we need nonlinear maps as well, say ReLU applied componentwise Here we have four matrices, whose composition gives a map R3 ReLU◦matrix − − − − − − − → R4 ReLU◦matrix − − − − − − − → R3 ReLU◦matrix − − − − − − − → R3 ReLU◦matrix − − − − − − − → R2 But ignore that for now ReLU doesn’t learn anything, its just brings in nonlinearity The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 12. What is a neural network (nn)?   a1 11 b1 1 a1 12 b1 2 a1 13 b1 3   , a2 11 a2 12 a2 13 b2 1 a2 21 a2 22 a2 23 b2 2 ▶ The ak ij and bk i are the parameters of our nn ▶ k = number of the layer ▶ Deep = many layers = better approximation The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 13. What is a neural network (nn)?   a1 11 b1 1 a1 12 b1 2 a1 13 b1 3   , a2 11 a2 12 a2 13 b2 1 a2 21 a2 22 a2 23 b2 2 ▶ The ak ij and bk i are the parameters of our nn ▶ k = number of the layer ▶ Deep = many layers = better approximation The point Many layers → many parameters These are good for approximating real world problems Examples ResNet-152 with 152. layers (used in transformer models such as ChatGPT) VGG-19 with 19 layers (used in image classification) GoogLeNet with 22 layers (used in face detection) The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 14. What is a neural network (nn)?   a1 11 b1 1 a1 12 b1 2 a1 13 b1 3   , a2 11 a2 12 a2 13 b2 1 a2 21 a2 22 a2 23 b2 2 ▶ The ak ij and bk i are the parameters of our nn ▶ k = number of the layer ▶ Deep = many layers = better approximation Side fact Gaming has improved AI!? A GPU can do e.g. matrix multiplications faster than a CPU and lots of nn run on GPUs The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5
  • 15. How learning works ▶ Supervised learning Create a dataset with answers, e.g. pictures of handwritten digits plus their label ▶ There are other forms of learning e.g. unsupervised, which I skip ▶ Split the data into ≈80% training and ≈20% testing data The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 16. How learning works ▶ Supervised learning Create a dataset with answers, e.g. pictures of handwritten digits plus their label ▶ There are other forms of learning e.g. unsupervised, which I skip ▶ Split the data into ≈80% training and ≈20% testing data Idea to keep in mind How to train students? There are lectures, exercises etc., the training data There is a final exam, the testing data Upon performance, we let them out into the wild The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 17. How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 18. How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Forward Boils down to a bunch of matrix multiplications followed by the nonlinear activation e.g. ReLU The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 19. How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Loss The difference between real values and predictions Task Minimize loss function The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 20. How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Backward This is running gradient descent on the loss function Slogan Adjust parameters following the steepest descent The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 21. How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat And what makes it even better: You can try it yourself My favorite tool is PyTorch but there are also other methods Let us see how! The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5
  • 22. What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 Input example Output example The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is machine learning? , ▶ Idea Approximate the unknown function R784 → R10 ▶ Neural network = a piecewise linear approximation (matrices + PL maps) ▶ The matrices = a bunch of numbers (weights) and offsets (biases) ▶ The PL maps = usually ReLU The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is machine learning? Forward → Loss → Backward → → → → ... ▶ Machine learning mantra ▶ Forward = calculate an approximation (start with random inputs) ▶ Loss = compare to real data ▶ Backward = adjust the approximation The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is a neural network (nn)? ▶ NN = a directed graph as above ▶ The task of a nn is to approximate an unknown function ▶ It consist of neurons = entries of vectors, and weights = entries of matrices Example Here we have two matrices, a 3-by-1 and a 2-by-3 matrix   w x y   and a11 a12 a13 a21 a22 a23 plus two bias terms as in y = matrix · x + bias The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5 How learning works ▶ Supervised learning Create a dataset with answers, e.g. pictures of handwritten digits plus their label ▶ There are other forms of learning e.g. unsupervised, which I skip ▶ Split the data into ≈80% training and ≈20% testing data The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Forward Boils down to a bunch of matrix multiplications followed by the nonlinear activation e.g. ReLU The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Loss The difference between real values and predictions Task Minimize loss function The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Backward This is running gradient descent on the loss function Slogan Adjust parameters following the steepest descent The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 There is still much to do... The mathematics of AI Or: Learning = forward, loss, backward April 2024 5 / 5
  • 23. What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is machine learning? ▶ Task Identify handwritten digits ▶ We can see this as a function in the following way: ▶ Convert the pictures into grayscale values, e.g. 28 × 28 grid of numbers ▶ Flatten the result into a vector, e.g. 28 × 28 7→ a vector with 282 = 784 entries ▶ The output is a vector with 10 entries ▶ We thus have a function R784 → R10 Input example Output example The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is machine learning? , ▶ Idea Approximate the unknown function R784 → R10 ▶ Neural network = a piecewise linear approximation (matrices + PL maps) ▶ The matrices = a bunch of numbers (weights) and offsets (biases) ▶ The PL maps = usually ReLU The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is machine learning? Forward → Loss → Backward → → → → ... ▶ Machine learning mantra ▶ Forward = calculate an approximation (start with random inputs) ▶ Loss = compare to real data ▶ Backward = adjust the approximation The mathematics of AI Or: Learning = forward, loss, backward April 2024 2 / 5 What is a neural network (nn)? ▶ NN = a directed graph as above ▶ The task of a nn is to approximate an unknown function ▶ It consist of neurons = entries of vectors, and weights = entries of matrices Example Here we have two matrices, a 3-by-1 and a 2-by-3 matrix   w x y   and a11 a12 a13 a21 a22 a23 plus two bias terms as in y = matrix · x + bias The mathematics of AI Or: Learning = forward, loss, backward April 2024 π / 5 How learning works ▶ Supervised learning Create a dataset with answers, e.g. pictures of handwritten digits plus their label ▶ There are other forms of learning e.g. unsupervised, which I skip ▶ Split the data into ≈80% training and ≈20% testing data The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Forward Boils down to a bunch of matrix multiplications followed by the nonlinear activation e.g. ReLU The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Loss The difference between real values and predictions Task Minimize loss function The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 How learning works ▶ Forward Run the nn = function on the training data ▶ Loss Calculate the difference “results - answers” (⇒ loss function) ▶ Backward Change the parameters trying to minimize the loss function ▶ Repeat Backward This is running gradient descent on the loss function Slogan Adjust parameters following the steepest descent The mathematics of AI Or: Learning = forward, loss, backward April 2024 4 / 5 Thanks for your attention! The mathematics of AI Or: Learning = forward, loss, backward April 2024 5 / 5