Abstract
We illustrate the use of statistical tools (asymptotic theories of standard error quantification using appro-priate statistical models, bootstrapping, and model comparison techniques) in addition to sensitivity analysis that may be employed to determine the information content in data sets. We do this in the con-text of recent models [S. Prigent, A. Ballesta, F. Charles, N. Lenuzza, P. Gabriel, L.M. Tine, H. Rezaei, and M. Doumic, An efficient kinetic model for assemblies of amyloid fibrils and its application to polyg-lutamine aggregation, PLoS ONE 7 (2012), e43273. doi:10.1371/journal.pone.0043273.] for nucleated polymerization in proteins, about which very little is known regarding the underlying mechanisms; thus, the methodology we develop here may be of great help to experimentalists. We conclude that the inves-tigated data sets will support with reasonable levels of uncertainty only the estimation of the parameters related to the early steps of the aggregation process.
| Original language | English |
|---|---|
| Pages (from-to) | 172-197 |
| Number of pages | 26 |
| Journal | Journal of Biological Dynamics |
| Volume | 9 |
| Issue number | 1 |
| DOIs | |
| Publication status | Published - 1 Jan 2015 |
| Externally published | Yes |
Keywords
- Information content
- Inverse problems
- Polyglutamine and aggregation modelling
- Sensitiv-ity
- Uncertainty quantification