Introducing Neural Tangent Kernel (NTK)
A review of a classic deep learning paper by Jascha Sohl-Dickstein et al.
Paper Overview
This paper, called Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent, provides a rigorous mathematical characterization of the training dynamics of wide neural networks, demonstrating that their evolution under gradient descent can be effectively described by a linearized model in function space. This phenomenon emerges from a first-order Taylor expansion of the network function around its initialization.
The core result of the paper is that the Neural Tangent Kernel (NTK) governs the learning dynamics of wide neural networks. Specifically, as width increases, the NTK remains constant during training, leading to an analytically tractable linear differential equation for function evolution. This framework builds upon previous work on Neural Network Gaussian Processes (NNGPs) and extends it to capture the full trajectory of learning rather than just initialization statistics.
Neural Tangent Kernel w/o Tough Math
The Neural Tangent Kernel (NTK) is a mathematical tool used to describe how a neural network learns during training. It helps us understand how small changes in the network's weights affect its output.
Imagine a neural network as a complex function that takes inputs (like an image) and produces outputs (like a classification). During training, the network updates its weights using gradient descent, meaning it adjusts its internal parameters step by step to minimize errors. The NTK captures how these weight updates influence the network’s predictions.
The NTK is a giant matrix that measures how much each weight in the network contributes to changes in the output. If the NTK stays constant during training (which happens in very wide networks), the network behaves like a simple linear model, even though it is a deep, nonlinear function. This allows us to predict the learning behavior of the network using mathematical formulas, rather than running experiments.
Main Paper Contributions:
The paper makes the following key contributions:
Linearization of Training Dynamics: The function space dynamics of wide networks are equivalent to those of a linear model obtained from their first-order Taylor expansion.
Constancy of the NTK: The NTK, which characterizes function evolution under gradient descent, remains approximately constant throughout training, even for finite-width networks.
Gaussian Process Interpretation: The trained network function remains Gaussian-distributed, governed by a mean function that evolves deterministically under gradient descent.
Closed-Form Training Dynamics: For mean squared error (MSE) loss, the function evolution is explicitly solvable in terms of the NTK eigenvalues.
These results provide a precise mathematical foundation for understanding deep learning as a special case of kernel methods, bridging nonlinear neural networks with classical linear models in function space.
Gradient Flow and Linearization
The training of neural networks under gradient descent follows the continuous-time gradient flow equation:
where θ_t represents the network parameters at time t, and L(θ) is the loss function. The key observation is that in the infinite-width limit, the network evolution simplifies dramatically due to the linearization of function dynamics. To formalize this, define the network output function (f is the trained neural net):
The Taylor expansion of f_t(x) around initialization θ_0 gives:
Introducing the weight perturbation vector
the function evolution obeys:
Using this, the function space evolution is:
where Θ_0(X, X) is the Neural Tangent Kernel, defined as:
Recall that NTK is a kernel describing the evolution of deep neural networks during their training by gradient descent. In the above formula it is computed at the network initialization θ_0.
Key Consequences of Linearization
The function dynamics become linear in parameter updates, reducing the complexity of the nonlinear optimization problem. The NTK controls the learning trajectory, and since it remains approximately constant, function evolution is predictable. For small learning rates, gradient flow approximates kernel gradient descent in function space. This linearization is in the infinite-width limit, and empirically holds for moderately wide networks.
Gaussian Process Correspondence and Bayesian Interpretation
A central insight of the paper is the connection between wide neural networks and Gaussian processes (GPs). At initialization, the network function f_0(x) follows a Gaussian process prior:
where K(X, X) is the Neural Network Gaussian Process (NNGP) Kernel, given by:
Under gradient descent, the function distribution remains Gaussian, but now evolves according to the NTK:
where the mean and covariance satisfy:
Closed-Form Training Dynamics and Eigenvalue Analysis
For squared loss
the differential equation governing function evolution is explicitly solvable. Expanding in the eigenbasis of Θ_0 :
the neural network trajectory is:
Each eigenmode decays exponentially with rate λ_i, so the learning speed is determined by the NTK spectrum. Modes with large λ_i converge faster, while small eigenvalues correspond to slow learning components.
Conclusion:
This paper provides a mathematically rigorous characterization of neural network training in the large-width limit, showing that gradient descent effectively transforms deep networks into linear models in function space. By proving that training dynamics are governed by a fixed kernel, the paper establishes a direct mathematical link between deep learning and well-understood mathematical frameworks.

