"Loss Minimization Yields Multicalibration for Large Neural Networks"
arxiv.org/abs/2304.09424
It seems the paper is not very well known and, hence, worthwhile to post about.
"Loss Minimization Yields Multicalibration for Large Neural Networks"
arxiv.org/abs/2304.09424
It seems the paper is not very well known and, hence, worthwhile to post about.
4. minimizing loss with regularizer that penalizes for large sizes n will avoid the ones where there is significant improvement by making the network a little bigger (i.e., to n+k+2).
4. minimizing loss with regularizer that penalizes for large sizes n will avoid the ones where there is significant improvement by making the network a little bigger (i.e., to n+k+2).
1. neural networks are rich enough so that if there is expected multi-calibration bias for a size n network for points identified by a size k network the two networks can be combined with size n+k+2 to reduce squared loss by the squared bias.
1. neural networks are rich enough so that if there is expected multi-calibration bias for a size n network for points identified by a size k network the two networks can be combined with size n+k+2 to reduce squared loss by the squared bias.
The minimizer of "squared loss minus a regularizer term that penalizes for neural network models by their size" is automatically multi-calibrated with respect to small neural networks.
The minimizer of "squared loss minus a regularizer term that penalizes for neural network models by their size" is automatically multi-calibrated with respect to small neural networks.
- paper reading on calibration: sites.northwestern.edu/hartline/cs-...
- lecture based on data economics: sites.northwestern.edu/hartline/cs-...
- paper reading on calibration: sites.northwestern.edu/hartline/cs-...
- lecture based on data economics: sites.northwestern.edu/hartline/cs-...
Bowing to my AI overloads, I used the example in my lecture today.