I’ve been asked a few times for recommendations on how to get started studying machine learning. I am sharing it here for anyone who is interested, and I will update it as I find new resources. This guide is meant for people with little to no experience.


Linear Algebra

The most important math for machine learning is linear algebra. I recommend the following resources:

“Essence of linear algebra” 3Blue1Brown playlist

This is an excellent playlist that focuses on building intuition for linear algebra. It purposefully avoids computation and instead focuses on the geometric interpretation of linear algebra. I recommend starting here.

Gilbert Strang’s lectures on Linear Algebra

Gilbert Strang is a legendary MIT professor. This was one of the original courses posted to MIT OpenCourseWare and is still one of the best. It is a full-length and comprehensive course on linear algebra.

Probability and Statistics

The second most import math for ML is probability and statistics. I recommend the following resources:

“Bayes theorem, the geometry of changing beliefs” by 3Blue1Brown

This video is a great introduction to Bayes theorem, one of the most important formulas in statistics. It breaks down each part of the formula, a visual proof of why it is true, and when is it useful.

“An Introduction to Statistical Learning: with Applications in Python” by James et al.

This is an excellent book for the foundations of statistics and ML. It is a full book, and one of the most advanced resources on the list, so it is a big time commitment. It assumes a strong background in linear algebra. It covers a broad range of topics, including linear regression, logistic regression, decision trees, random forests, support vector machines, neural networks, clustering, and more.

It is free to read online and also comes with Python code examples. The first edition of this book, ICLR, covers the same material but is for the R programming language. I definitely recommend the new Python version, ISLP.


The third most important math for ML is calculus. It is the basis for understanding backpropagation.

“Essence of calculus” 3Blue1Brown playlist

This is another excellent series from 3Blue1Brown. Like his “Essence of Linear Algebra” playlist, this one focuses on building intuition for calculus rather than computation. It builds up using thought experiments and visuals in a way that makes you feel like you could’ve discovered calculus for yourself. I recommend starting here.

Gilbert Strang’s “Highlights of Calculus” playlist

I love Gilbert Strang and this is another excellent playlist. This isn’t a full course, but it still a deep dive on calculus. It helped me develop a deeper intuition for calculus.



Python is the most popular language for machine learning. This is really the only language you need to know. I recommend the following resources:

“Automate the Boring Stuff with Python” by Al Sweigart

I believe that the best way to learn anything is with project-based learning. The nice thing about projects learning is that concrete goals motivate learning in a way that abstract academic texts can’t.

I like that each chapter is a project that gives you a useful tool that you can use in your daily life. I think everyone should read it for basic programming literacy.

It is also free to read online.

Codecademy Python 3 course

Codecademy is nice because it gives you interactive programming challenges, and helps by pointing out errors. This is a course you can complete in a few days and it will give you a good foundation in Python. It’s also nice that it gives you a virtual Python terminal in your browser, so you don’t have to install anything before you start learning how to code.

Other languages

The other main programming languages for machine learning are R and Julia.

Julia is a newer language purpose-built for data science, and it is gaining popularity. I think it is an elegant alternative to Python, however I can’t recommend learning it because almost everyone uses Python.

R is an older language that is popular in statistics. I really hate R, and I don’t recommend learning it. It is an inelegant language, and it is not as generally useful as Python. I think starting to code with R will teach you terrible coding practices. It does, however, have a tremendous amount of statistical packages. If you are doing a lot of pure statistics, you may want to learn R. Otherwise, I recommend sticking with Python, especially if you are more interested in deep learning.

Machine Learning

ML Engineering

“Machine Learning Engineering for Production (MLOps)” course by Andrew Ng

This playlist focuses more on the engineering side of machine learning. For example, it covers common pitfalls and challenges when deploying ML models in production. I recommend this playlist for anyone working in an ML-adjacent role, as it gives you the bigger picture of the types of challenges with ML in production. The course claims that it is an intermediate level course, but I think it is accessible for beginners.

“A Recipe for Training Neural Networks” by Andrej Karpathy

This is a short blog post that gives some flavor of challenges in training neural networks, and why they are so difficult to diagnose as compared to other software systems.

The other blog posts on Andrej Karpathy’s blog are also excellent!

“Hidden Technical Debt in Machine Learning Systems” by Sculley et al.

This is a seminal academic paper that covers the challenges of ML systems in production, written by a team at Google. One of the main points it makes is that only a small fraction of real-world ML systems are composed of actual ML code. The rest is data collection, data validation, feature extraction, serving infrastructure, monitoring, etc. This is a good paper to read if you want to understand the bigger picture of ML systems, and why they are so difficult to build.

“Deep Learning” Coursera specialization by Andrew Ng

Part 1

Part 2

Part 3

Part 4

Part 5

This is the first course added to Coursera by Andrew Ng, and it is still one of the best. It is a great introduction to deep learning theory, and touches on the engineering side of ML as well. It is a bit dated, but it is still a great resource.

There is a new version of the course (“Machine Learning Specialization”), but I haven’t watched it yet.

“Data-centric ML” by Andrew Ng

This is a short but influential talk by Andrew Ng that argues that there is an over-emphasis on model-centric ML, and that data-centric ML is more important. I recommend this talk for anyone who is interested in ML engineering as well as managers that work with ML teams.

There is an MIT course that is based on this talk: “Data-Centric AI” course by MIT

Deep Learning

“Neural networks” 3Blue1Brown playlist

This is a nice visual introduction to what is a neural network, as well as how backpropagation and gradient descent work. 3Blue1Brown is always great.

“But what is a convolution?” by 3Blue1Brown

Convolutions are a key component of convolutional neural networks, which are the most popular type of neural network for computer vision. This is a nice visual introduction to what is a convolution. Another great video by 3Blue1Brown.

“Visualizing Deep Learning” playlist by vcubingx

More visual intuition building for neural networks. This playlist is a bit more advanced than the 3Blue1Brown playlist, but it is still accessible for beginners.

“Deep Learning” book by Ian Goodfellow et al.

This is an advanced book on deep learning theory. I’ve mostly read select chapters in Part 2 and Part 3, and it was great. It is free to read online.

“NYU Deep Learning” course by Yann LeCun and Alfredo Canziani

This is a full course on deep learning taught at NYU during COVID. It is a great resource for learning the theory of deep learning. It is a bit more advanced than the Andrew Ng course, but it is still accessible for beginners.

Alfredo Canziani also is currently writing a book on deep learning theory from an energy perspective. As of right now, it is not finished, but I am excited for it based on the sneak-peaks he has posted on Twitter.

“The Little Book of Deep Learning” by François Fleuret

This is a short but dense book on deep learning theory (150 pages). I haven’t finished reading it but it has very nice visualization and explanations.

This book is available online for free or you can buy a physical copy for $10.

Machine Learning Street Talk Podcast

This podcast does long 1-3 hour interviews that go into depth on more advanced topics, such as the spline theory of neural networks. I’ve relistened to various of these interviews, such as with Ishan Misra, Randall Balestriero, and Simon Kornblith.

Large Language Models

I split off a new section for LLMs since they are all the rage right now.

“Neural Networks: Zero to Hero” playlist by Andrej Karpathy

This playlist is a “hackers guide” to neural networks, with special attention to language models. It codes end-to-end examples from scratch in pure Python, so while the code is not the fastest it is compact and easy to read. I haven’t watched the whole playlist, but I’ve watched a few videos and they were great (the GPT3 from scratch was fantastic).

Andrew Ng Practical LLM MOOCs

More recently, Andrew Ng has created more beginner-friendly lectures about LLM applications.

  • short courses: These short (~1 hour) videos look like a good introduction to various LLM concepts: vector DBs, inference, prompt engineering, validation, RLHF. The course on creating RAG web apps looks especially interesting.
  • LLM MOOC: A week-end course that goes into a bit more depth than the short videos. Still very beginner-level.

LLMs explained by Letitia

This playlist is an intermediate introduction to LLMs. It covers a bunch of topics like Low-Rank Adaptation, state space models, and transformers.

In general, the AI Coffee Break channel is great for intermediate-level explainers on deep-learning topics, such as graph neural networks, diffusion models, CLIP embeddings, etc.

AI Safety

Robert Miles AI Safety

Robert Miles’ channel is fantastic for beginners learning about AI safety. He does a great job of explaining the orthogonality thesis, instrumental convergence, mesa-optimizers, etc.

Mechanistic Interpretability

Neel Nanda

Neel Nanda is a great person to follow in the mechinterp space. He consistenly has great content, from his interviews, his mechinterp blog, YouTube channel, and outrageously detailed mechinterp glossary.


Distill is a journal that publishes articles on machine learning. It aims to be a more visual, intuitive, and interactive vision of academic publishing for ML. The articles are extremely high quality and are a great resource for learning about advanced topics in ML. The articles have an emphasis on interpretability of ML models, deliving into the “black box” of neural networks by examining the behavior of specific circuits within the network.

Their Transformer Circuits article is one of the best I’ve found for deeply understanding how the residual streams work in transformers.

One of the core Distill authors, Chris Olah, also has a personal blog with some legendary articles.

Advanced Research

The best way to learn about advanced topics in ML is to read papers. I recommend the following resources for finding papers:

  • arXiv Sanity Preserver
  • Twitter. I find Twitter to be much better than Reddit for keeping up with the latest research. You can see who I follow here (mostly computer vision researchers).
  • Semantic Scholar. Great for finding related papers.