TY - CHAP
T1 - Navigating mathematical basics
T2 - A primer for deep learning in science
AU - Liquet, Benoit
AU - Moka, Sarat
AU - Nazarathy, Yoni
PY - 2024
Y1 - 2024
N2 - We present a gentle introduction to elementary mathematical notation with the focus of communicating deep learning principles. This is a “math crash course” aimed at quickly enabling scientists with understanding of the building blocks used in many equations, formulas, and algorithms that describe deep learning. While this short presentation cannot replace solid mathematical knowledge that needs multiple courses and years to solidify, our aim is to allow nonmathematical readers to overcome hurdles of reading texts that also use such mathematical notation. We describe a few basic deep learning models using mathematical notation before we unpack the meaning of the notation. In particular, this text includes an informal introduction to summations, sets, functions, vectors, matrices, gradients, and a few more objects that are often used to describe deep learning. While this is a mathematical crash course, our presentation is kept in the context of deep learning and machine learning models including the sigmoid model, the softmax model, and fully connected feedforward deep neural networks. We also hint at basic mathematical objects appearing in neural networks for images and text data.
AB - We present a gentle introduction to elementary mathematical notation with the focus of communicating deep learning principles. This is a “math crash course” aimed at quickly enabling scientists with understanding of the building blocks used in many equations, formulas, and algorithms that describe deep learning. While this short presentation cannot replace solid mathematical knowledge that needs multiple courses and years to solidify, our aim is to allow nonmathematical readers to overcome hurdles of reading texts that also use such mathematical notation. We describe a few basic deep learning models using mathematical notation before we unpack the meaning of the notation. In particular, this text includes an informal introduction to summations, sets, functions, vectors, matrices, gradients, and a few more objects that are often used to describe deep learning. While this is a mathematical crash course, our presentation is kept in the context of deep learning and machine learning models including the sigmoid model, the softmax model, and fully connected feedforward deep neural networks. We also hint at basic mathematical objects appearing in neural networks for images and text data.
KW - Deep Learning
KW - Machine learning
KW - Mathematics for Data Science
KW - Mathematics of Machine Learning
UR - http://www.scopus.com/inward/record.url?scp=85208989859&partnerID=8YFLogxK
UR - https://doi.org/10.1007/978-3-031-64892-2
U2 - 10.1007/978-3-031-64892-2_5
DO - 10.1007/978-3-031-64892-2_5
M3 - Chapter
C2 - 39523260
AN - SCOPUS:85208989859
SN - 9783031648915
T3 - Advances in Experimental Medicine and Biology
SP - 71
EP - 96
BT - Computational neurosurgery
A2 - Di Ieva, Antonio
A2 - Molina, Eric Suero
A2 - Liu, Sidong
A2 - Russo, Carlo
PB - Springer, Springer Nature
CY - Cham
ER -