Speaker
Prasanth Shyamsundar
(Fermi National Accelerator Laboratory)
Description
The bias-variance tradeoff is a well-recognized phenomenon in statistics and machine learning. In this talk, I will discuss an extension, dubbed the bias-variance-correlation tradeoff. Roughly speaking, as the flexibility of a model decreases, the correlations in the outputs of a trained model for different inputs increases. Such correlations have implications for several applications of machine learning in high energy physics, e.g., the use generative models for event generation. In particular, I will argue that claims in the literature of data amplification by generative models stem from ignoring important correlations between the model's outputs for different inputs.
Author
Prasanth Shyamsundar
(Fermi National Accelerator Laboratory)