High throughput nonparametric probability density estimation.
In high throughput applications, such as those found in bioinformatics and finance, it is important to determine accurate probability distribution functions despite only minimal information about data characteristics, and without using human subjectivity. Such an automated process for univariate dat...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2018-01-01
|
Series: | PLoS ONE |
Online Access: | http://europepmc.org/articles/PMC5947915?pdf=render |
_version_ | 1811219014316195840 |
---|---|
author | Jenny Farmer Donald Jacobs |
author_facet | Jenny Farmer Donald Jacobs |
author_sort | Jenny Farmer |
collection | DOAJ |
description | In high throughput applications, such as those found in bioinformatics and finance, it is important to determine accurate probability distribution functions despite only minimal information about data characteristics, and without using human subjectivity. Such an automated process for univariate data is implemented to achieve this goal by merging the maximum entropy method with single order statistics and maximum likelihood. The only required properties of the random variables are that they are continuous and that they are, or can be approximated as, independent and identically distributed. A quasi-log-likelihood function based on single order statistics for sampled uniform random data is used to empirically construct a sample size invariant universal scoring function. Then a probability density estimate is determined by iteratively improving trial cumulative distribution functions, where better estimates are quantified by the scoring function that identifies atypical fluctuations. This criterion resists under and over fitting data as an alternative to employing the Bayesian or Akaike information criterion. Multiple estimates for the probability density reflect uncertainties due to statistical fluctuations in random samples. Scaled quantile residual plots are also introduced as an effective diagnostic to visualize the quality of the estimated probability densities. Benchmark tests show that estimates for the probability density function (PDF) converge to the true PDF as sample size increases on particularly difficult test probability densities that include cases with discontinuities, multi-resolution scales, heavy tails, and singularities. These results indicate the method has general applicability for high throughput statistical inference. |
first_indexed | 2024-04-12T07:18:37Z |
format | Article |
id | doaj.art-c4d7c76ce65443e1af6bb2e9adbfea67 |
institution | Directory Open Access Journal |
issn | 1932-6203 |
language | English |
last_indexed | 2024-04-12T07:18:37Z |
publishDate | 2018-01-01 |
publisher | Public Library of Science (PLoS) |
record_format | Article |
series | PLoS ONE |
spelling | doaj.art-c4d7c76ce65443e1af6bb2e9adbfea672022-12-22T03:42:24ZengPublic Library of Science (PLoS)PLoS ONE1932-62032018-01-01135e019693710.1371/journal.pone.0196937High throughput nonparametric probability density estimation.Jenny FarmerDonald JacobsIn high throughput applications, such as those found in bioinformatics and finance, it is important to determine accurate probability distribution functions despite only minimal information about data characteristics, and without using human subjectivity. Such an automated process for univariate data is implemented to achieve this goal by merging the maximum entropy method with single order statistics and maximum likelihood. The only required properties of the random variables are that they are continuous and that they are, or can be approximated as, independent and identically distributed. A quasi-log-likelihood function based on single order statistics for sampled uniform random data is used to empirically construct a sample size invariant universal scoring function. Then a probability density estimate is determined by iteratively improving trial cumulative distribution functions, where better estimates are quantified by the scoring function that identifies atypical fluctuations. This criterion resists under and over fitting data as an alternative to employing the Bayesian or Akaike information criterion. Multiple estimates for the probability density reflect uncertainties due to statistical fluctuations in random samples. Scaled quantile residual plots are also introduced as an effective diagnostic to visualize the quality of the estimated probability densities. Benchmark tests show that estimates for the probability density function (PDF) converge to the true PDF as sample size increases on particularly difficult test probability densities that include cases with discontinuities, multi-resolution scales, heavy tails, and singularities. These results indicate the method has general applicability for high throughput statistical inference.http://europepmc.org/articles/PMC5947915?pdf=render |
spellingShingle | Jenny Farmer Donald Jacobs High throughput nonparametric probability density estimation. PLoS ONE |
title | High throughput nonparametric probability density estimation. |
title_full | High throughput nonparametric probability density estimation. |
title_fullStr | High throughput nonparametric probability density estimation. |
title_full_unstemmed | High throughput nonparametric probability density estimation. |
title_short | High throughput nonparametric probability density estimation. |
title_sort | high throughput nonparametric probability density estimation |
url | http://europepmc.org/articles/PMC5947915?pdf=render |
work_keys_str_mv | AT jennyfarmer highthroughputnonparametricprobabilitydensityestimation AT donaldjacobs highthroughputnonparametricprobabilitydensityestimation |