Passer à la navigation principale Passer à la recherche Passer au contenu principal

Neural Conditional Probability for Uncertainty Quantification

  • Vladimir R. Kostic
  • , Karim Lounici
  • , Grégoire Pacreau
  • , Giacomo Turri
  • , Pietro Novelli
  • , Massimiliano Pontil

Résultats de recherche: Contribution à un journalArticle de conférenceRevue par des pairs

Résumé

We introduce Neural Conditional Probability (NCP), an operator-theoretic approach to learning conditional distributions with a focus on statistical inference tasks. NCP can be used to build conditional confidence regions and extract key statistics such as conditional quantiles, mean, and covariance. It offers streamlined learning via a single unconditional training phase, allowing efficient inference without the need for retraining even when conditioning changes. By leveraging the approximation capabilities of neural networks, NCP efficiently handles a wide variety of complex probability distributions. We provide theoretical guarantees that ensure both optimization consistency and statistical accuracy. In experiments, we show that NCP with a 2-hidden-layer network matches or outperforms leading methods. This demonstrates that a a minimalistic architecture with a theoretically grounded loss can achieve competitive results, even in the face of more complex architectures.

langue originaleAnglais
journalAdvances in Neural Information Processing Systems
Volume37
étatPublié - 1 janv. 2024
Evénement38th Conference on Neural Information Processing Systems, NeurIPS 2024 - Vancouver, Canada
Durée: 9 déc. 202415 déc. 2024

Empreinte digitale

Examiner les sujets de recherche de « Neural Conditional Probability for Uncertainty Quantification ». Ensemble, ils forment une empreinte digitale unique.

Contient cette citation