On The Performance Of The Maximum Likelihood Over Large Models
This dissertation investigates non-parametric regression over large function classes, specifically, non-Donsker classes. We will present the concept of non-Donsker classes and study the statistical performance of Least Squares Estimator (LSE) --- which also serves as the Maximum Likelihood Estimato...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
Massachusetts Institute of Technology
2023
|
Online Access: | https://hdl.handle.net/1721.1/152867 |
_version_ | 1826211908350377984 |
---|---|
author | Kur, Gil |
author2 | Rakhlin, Alexander |
author_facet | Rakhlin, Alexander Kur, Gil |
author_sort | Kur, Gil |
collection | MIT |
description | This dissertation investigates non-parametric regression over large function classes, specifically, non-Donsker classes. We will present the concept of non-Donsker classes and study the statistical performance of Least Squares Estimator (LSE) --- which also serves as the Maximum Likelihood Estimator (MLE) under Gaussian noise --- over these classes. (1) We demonstrate the minimax sub-optimality of the LSE in the non-Donsker regime, extending traditional findings of over these classes. (1) We demonstrate the minimax sub-optimality of the LSE in the non-Donsker regime, extending traditional findings of Birgé and Massart 93' and resolving a longstanding conjecture of Gardner, Markus and Milanfar 06'. (2) We reveal that in the non-Donsker regime, the sub-optimality of LSE arises solely from its elevated bias error term (in terms of the bias and variance decomposition). (3) We introduce the first minimax optimal algorithm for multivariate convex regression with a polynomial runtime in the number of samples -- showing that one can overcome the sub-optimality of the LSE in efficient runtime. (4) We study the minimal error of the LSE both in random and fixed design settings. and Massart 93' and resolving a longstanding conjecture of Gardner, Markus and Milanfar 06'. (2) We reveal that in the non-Donsker regime, the sub-optimality of LSE arises solely from its elevated bias error term (in terms of the bias and variance decomposition). (3) We introduce the first minimax optimal algorithm for multivariate convex regression with a polynomial runtime in the number of samples -- showing that one can overcome the sub-optimality of the LSE in efficient runtime. (4) We study the minimal error of the LSE both in random and fixed design settings. |
first_indexed | 2024-09-23T15:13:21Z |
format | Thesis |
id | mit-1721.1/152867 |
institution | Massachusetts Institute of Technology |
last_indexed | 2024-09-23T15:13:21Z |
publishDate | 2023 |
publisher | Massachusetts Institute of Technology |
record_format | dspace |
spelling | mit-1721.1/1528672023-11-03T03:12:14Z On The Performance Of The Maximum Likelihood Over Large Models Kur, Gil Rakhlin, Alexander Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science This dissertation investigates non-parametric regression over large function classes, specifically, non-Donsker classes. We will present the concept of non-Donsker classes and study the statistical performance of Least Squares Estimator (LSE) --- which also serves as the Maximum Likelihood Estimator (MLE) under Gaussian noise --- over these classes. (1) We demonstrate the minimax sub-optimality of the LSE in the non-Donsker regime, extending traditional findings of over these classes. (1) We demonstrate the minimax sub-optimality of the LSE in the non-Donsker regime, extending traditional findings of Birgé and Massart 93' and resolving a longstanding conjecture of Gardner, Markus and Milanfar 06'. (2) We reveal that in the non-Donsker regime, the sub-optimality of LSE arises solely from its elevated bias error term (in terms of the bias and variance decomposition). (3) We introduce the first minimax optimal algorithm for multivariate convex regression with a polynomial runtime in the number of samples -- showing that one can overcome the sub-optimality of the LSE in efficient runtime. (4) We study the minimal error of the LSE both in random and fixed design settings. and Massart 93' and resolving a longstanding conjecture of Gardner, Markus and Milanfar 06'. (2) We reveal that in the non-Donsker regime, the sub-optimality of LSE arises solely from its elevated bias error term (in terms of the bias and variance decomposition). (3) We introduce the first minimax optimal algorithm for multivariate convex regression with a polynomial runtime in the number of samples -- showing that one can overcome the sub-optimality of the LSE in efficient runtime. (4) We study the minimal error of the LSE both in random and fixed design settings. Ph.D. 2023-11-02T20:23:27Z 2023-11-02T20:23:27Z 2023-09 2023-09-21T14:26:07.968Z Thesis https://hdl.handle.net/1721.1/152867 0000-0001-7386-1686 In Copyright - Educational Use Permitted Copyright retained by author(s) https://rightsstatements.org/page/InC-EDU/1.0/ application/pdf Massachusetts Institute of Technology |
spellingShingle | Kur, Gil On The Performance Of The Maximum Likelihood Over Large Models |
title | On The Performance Of The Maximum Likelihood Over Large Models |
title_full | On The Performance Of The Maximum Likelihood Over Large Models |
title_fullStr | On The Performance Of The Maximum Likelihood Over Large Models |
title_full_unstemmed | On The Performance Of The Maximum Likelihood Over Large Models |
title_short | On The Performance Of The Maximum Likelihood Over Large Models |
title_sort | on the performance of the maximum likelihood over large models |
url | https://hdl.handle.net/1721.1/152867 |
work_keys_str_mv | AT kurgil ontheperformanceofthemaximumlikelihoodoverlargemodels |