The Loss Kernel: A Geometric Probe for Deep Learning Interpretability

Maxwell Adam =
Timaeus, University of Melbourne
Zach Furman =
University of Melbourne
Jesse Hoogland
Timaeus
October 1, 2025

Abstract

We introduce the loss kernel, an interpretability method for measuring similarity between data points according to a trained neural network. The kernel is the covariance matrix of per-sample losses computed under a distribution of low-loss-preserving parameter perturbations. We first validate our method on a synthetic multitask problem, showing it separates inputs by task as predicted by theory. We then apply this kernel to Inception-v1 to visualize the structure of ImageNet, and we show that the kernel's structure aligns with the WordNet semantic hierarchy. This establishes the loss kernel as a practical tool for interpretability and data attribution.

Cite as

@article{adam2025the,
  title = {The Loss Kernel: A Geometric Probe for Deep Learning Interpretability},
  author = {Maxwell Adam and Zach Furman and Jesse Hoogland},
  year = {2025},
  abstract = {We introduce the loss kernel, an interpretability method for measuring similarity between data points according to a trained neural network. The kernel is the covariance matrix of per-sample losses computed under a distribution of low-loss-preserving parameter perturbations. We first validate our method on a synthetic multitask problem, showing it separates inputs by task as predicted by theory. We then apply this kernel to Inception-v1 to visualize the structure of ImageNet, and we show that the kernel's structure aligns with the WordNet semantic hierarchy. This establishes the loss kernel as a practical tool for interpretability and data attribution.},
  eprint = {2509.26537},
  archivePrefix = {arXiv},
  url = {https://arxiv.org/abs/2509.26537}
}
Click to copy