Show simple item record

dc.contributor.authorYang, Wanqian
dc.date.accessioned2020-08-28T10:38:12Z
dc.date.created2020-05
dc.date.issued2020-06-17
dc.date.submitted2020
dc.identifier.citationYang, Wanqian. 2020. Making Decisions Under High Stakes: Trustworthy and Expressive Bayesian Deep Learning. Bachelor's thesis, Harvard College.
dc.identifier.urihttps://nrs.harvard.edu/URN-3:HUL.INSTREPOS:37364721*
dc.description.abstractMachine learning models applied to high-stakes domains must navigate the tradeoff between (i) having enough representation power to learn effectively from high-dimensional, high-volume datasets, while (ii) avoiding cost-prohibitive errors due to model overconfidence and poor generalization. This need has led to growing interest in Bayesian neural networks (BNN), models that perform Bayesian inference on deep neural networks. BNNs are able to quantify predictive uncertainty over an inherently rich hypothesis space, which is crucial for high-stakes decision-making. In this thesis, we present two contributions that tackle the shortcomings of BNNs. First, BNN priors are defined in uninterpretable parameter space, which makes it difficult for end users to express functional prior knowledge independent of training data. We formulate two novel priors that incorporate functional constraints (i.e. what values the output should hold for any given input) that can easily be specified by end users. The resulting model is amenable to black-box inference. We demonstrate its efficacy on two high-stakes domains: (i) enforcing physiologically feasible interventions on a clinical action prediction task, and (ii) enforcing racial fairness constraints on a recidivism prediction task where the training data is biased. Next, variational approximations that are typically used for BNN posterior inference do not come with provable error guarantees, making it difficult to trust their predictive estimates. By exploiting the functional form of BNNs, we bound the predictive mean error of such approximations via maximum mean discrepancy on a reproducing kernel Hilbert space. Our bound is easily estimable and directly useful to end users as it is specified in predictive space.
dc.description.sponsorshipComputer Science
dc.description.sponsorshipComputer Science
dc.format.mimetypeapplication/pdf
dc.language.isoen
dash.licenseLAA
dc.titleMaking Decisions Under High Stakes: Trustworthy and Expressive Bayesian Deep Learning
dc.typeThesis or Dissertation
dash.depositing.authorYang, Wanqian
dc.date.available2020-08-28T10:38:12Z
thesis.degree.date2020
thesis.degree.grantorHarvard College
thesis.degree.grantorHarvard College
thesis.degree.levelUndergraduate
thesis.degree.levelUndergraduate
thesis.degree.nameAB
thesis.degree.nameAB
dc.type.materialtext
thesis.degree.departmentComputer Science
thesis.degree.departmentComputer Science
dash.identifier.vireo
dc.identifier.orcid0000-0002-6372-3607
dash.author.emailywanqian95@gmail.com


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record