Selected article for: "Markov chain and parameter estimation"

Author: Silvestro, Daniele; Andermann, Tobias
Title: Prior choice affects ability of Bayesian neural networks to identify unknowns
  • Cord-id: l1otrhq1
  • Document date: 2020_5_11
  • ID: l1otrhq1
    Snippet: Deep Bayesian neural networks (BNNs) are a powerful tool, though computationally demanding, to perform parameter estimation while jointly estimating uncertainty around predictions. BNNs are typically implemented using arbitrary normal-distributed prior distributions on the model parameters. Here, we explore the effects of different prior distributions on classification tasks in BNNs and evaluate the evidence supporting the predictions based on posterior probabilities approximated by Markov Chain
    Document: Deep Bayesian neural networks (BNNs) are a powerful tool, though computationally demanding, to perform parameter estimation while jointly estimating uncertainty around predictions. BNNs are typically implemented using arbitrary normal-distributed prior distributions on the model parameters. Here, we explore the effects of different prior distributions on classification tasks in BNNs and evaluate the evidence supporting the predictions based on posterior probabilities approximated by Markov Chain Monte Carlo sampling and by computing Bayes factors. We show that the choice of priors has a substantial impact on the ability of the model to confidently assign data to the correct class (true positive rates). Prior choice also affects significantly the ability of a BNN to identify out-of-distribution instances as unknown (false positive rates). When comparing our results against neural networks (NN) with Monte Carlo dropout we found that BNNs generally outperform NNs. Finally, in our tests we did not find a single best choice as prior distribution. Instead, each dataset yielded the best results under a different prior, indicating that testing alternative options can improve the performance of BNNs.

    Search related documents:
    Co phrase search for related documents
    • accuracy term and machine learning: 1, 2, 3, 4, 5, 6, 7, 8, 9
    • accuracy term and machine learning classifier: 1
    • activation function and adam optimizer: 1, 2, 3
    • activation function and loss function: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12
    • activation function and machine learning: 1, 2, 3, 4, 5, 6
    • adam optimizer and loss function: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10
    • adam optimizer and machine learning: 1, 2
    • additional bias and machine learning: 1
    • additional knowledge and machine learning: 1, 2, 3, 4
    • additional knowledge and machine learning classifier: 1
    • additional measure and machine learning: 1
    • long exact and machine learning: 1
    • loss function and low accuracy: 1
    • loss function and machine learning: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23
    • low accuracy and machine learning: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11