Strong converses and high-dimensional statistical estimation problems
Update: 2018-07-24
Description
In many statistical inference problems, we wish to bound the performance of any possible estimator. This can be seen as a converse result, in a standard information-theoretic sense. A standard approach in the statistical literature is based on Fano’s inequality, which typically gives a weak converse. We adapt these arguments by replacing Fano by more recent information-theoretic ideas, based on the work of Polyanskiy, Poor and Verdu. This gives tighter lower bounds that can be easily computed and are asymptotically sharp. We illustrate our technique in three applications: density estimation, active learning of a binary classifier, and compressed sensing, obtaining tighter risk lower bounds in each case.
(joint with Oliver Johnson, see doi:10.1214/18-EJS14)
(joint with Oliver Johnson, see doi:10.1214/18-EJS14)
Comments
Top Podcasts
The Best New Comedy Podcast Right Now – June 2024The Best News Podcast Right Now – June 2024The Best New Business Podcast Right Now – June 2024The Best New Sports Podcast Right Now – June 2024The Best New True Crime Podcast Right Now – June 2024The Best New Joe Rogan Experience Podcast Right Now – June 20The Best New Dan Bongino Show Podcast Right Now – June 20The Best New Mark Levin Podcast – June 2024
In Channel