quick & clean: computationally efficient methods for value ... · available data (&...
TRANSCRIPT
Quick & clean: computationally efficient methods for Value ofInformation measures
Gianluca Baio
University College LondonDepartment of Statistical Science
(Joint work with Anna Heath and Ioanna Manolopoulou)(Thanks to Mark Strong)
http://www.ucl.ac.uk/statistics/research/statistics-health-economics/http://www.statistica.it/gianluca
https://github.com/giabaio
http://www.statistica.it/gianluca/project/voi/https://www.convoi-group.org/
O’Bayes 2019Objective Bayes Methodology Conference
University of Warwick, Monday 1 July 2019
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 1 / 21
O’Bayes?. . .
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 2 / 21
Outline
1. Value of Information– Basics
2. EVPPI– EVPPI as a (Gaussian Process) regression problem– Faster EVPPI (using INLA/SPDE)– Examples — research prioritisation
3. EVSI (I’ll try really hard to keep time for this!...)– Ridiculously short introduction– Moment-matching– Examples — research prioritisation + study design
4. Conclusions
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 3 / 21
Outline
1. Value of Information– Basics
2. EVPPI– EVPPI as a (Gaussian Process) regression problem– Faster EVPPI (using INLA/SPDE)– Examples — research prioritisation
3. EVSI (I’ll try really hard to keep time for this!...)– Ridiculously short introduction– Moment-matching– Examples — research prioritisation + study design
4. Conclusions
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 3 / 21
Outline
1. Value of Information– Basics
2. EVPPI– EVPPI as a (Gaussian Process) regression problem– Faster EVPPI (using INLA/SPDE)– Examples — research prioritisation
3. EVSI (I’ll try really hard to keep time for this!...)– Ridiculously short introduction– Moment-matching– Examples — research prioritisation + study design
4. Conclusions
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 3 / 21
Outline
1. Value of Information– Basics
2. EVPPI– EVPPI as a (Gaussian Process) regression problem– Faster EVPPI (using INLA/SPDE)– Examples — research prioritisation
3. EVSI (I’ll try really hard to keep time for this!...)– Ridiculously short introduction– Moment-matching– Examples — research prioritisation + study design
4. Conclusions
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 3 / 21
Health technology assessment (HTA)
Objective: Combine costs & benefits of a given intervention (against at least onecomparator) into a rational scheme for allocating resources
Statisticalmodel
Economicmodel
Decisionanalysis
Uncertaintyanalysis
• Estimates relevant populationparameters θ
• Varies with the type ofavailable data (& statisticalapproach!)
• Combines the parameters to obtaina population average measure forcosts and clinical benefits
• Varies with the type of availabledata & statistical model used
• Summarises the economic modelby computing suitable measures of“cost-effectiveness”
• Dictates the best course ofactions, given current evidence
• Standardised process
• Assesses the impact of uncertainty (eg inparameters or model structure) on theeconomic results
• Mandatory in many jurisdictions (includingNICE, in the UK). Often termed “PSA”
• Fundamentally Bayesian!
Uncertaintyanalysis
• Assesses the impact of uncertainty (egin parameters or model structure) onthe economic results
• Mandatory in many jurisdictions(including NICE, in the UK)
• Fundamentally Bayesian!
“Two-stage approach”: Spiegelhalter, Abrams & Myles, 2004
1. Estimation (base-case)
θ
yp(y | θ)
θ = f(Y )
2. Probabilistic sensitivity analysis
⇒
θp(θ) ! g(θ)
Estimation & PSA (one stage)
θ
yp(y | θ)
p(θ) p(θ | y)
“Integrated approach”: Spiegelhalter, Abrams & Myles (2004)Baio, Berardi & Heath (2017)
∆e = fe(θ)
∆c = fc(θ)
. . .
ICER = g(∆e,∆c)
EIB = h(∆e,∆c; k)
. . .
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 4 / 21
Health technology assessment (HTA)
Objective: Combine costs & benefits of a given intervention (against at least onecomparator) into a rational scheme for allocating resources
Statisticalmodel
Economicmodel
Decisionanalysis
Uncertaintyanalysis
• Estimates relevant populationparameters θ
• Varies with the type ofavailable data (& statisticalapproach!)
• Combines the parameters to obtaina population average measure forcosts and clinical benefits
• Varies with the type of availabledata & statistical model used
• Summarises the economic modelby computing suitable measures of“cost-effectiveness”
• Dictates the best course ofactions, given current evidence
• Standardised process
• Assesses the impact of uncertainty (eg inparameters or model structure) on theeconomic results
• Mandatory in many jurisdictions (includingNICE, in the UK). Often termed “PSA”
• Fundamentally Bayesian!
Uncertaintyanalysis
• Assesses the impact of uncertainty (egin parameters or model structure) onthe economic results
• Mandatory in many jurisdictions(including NICE, in the UK)
• Fundamentally Bayesian!
“Two-stage approach”: Spiegelhalter, Abrams & Myles, 2004
1. Estimation (base-case)
θ
yp(y | θ)
θ = f(Y )
2. Probabilistic sensitivity analysis
⇒
θp(θ) ! g(θ)
Estimation & PSA (one stage)
θ
yp(y | θ)
p(θ) p(θ | y)
“Integrated approach”: Spiegelhalter, Abrams & Myles (2004)Baio, Berardi & Heath (2017)
∆e = fe(θ)
∆c = fc(θ)
. . .
ICER = g(∆e,∆c)
EIB = h(∆e,∆c; k)
. . .
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 4 / 21
Health technology assessment (HTA)
Objective: Combine costs & benefits of a given intervention (against at least onecomparator) into a rational scheme for allocating resources
Statisticalmodel
Economicmodel
Decisionanalysis
Uncertaintyanalysis
• Estimates relevant populationparameters θ
• Varies with the type ofavailable data (& statisticalapproach!)
• Combines the parameters to obtaina population average measure forcosts and clinical benefits
• Varies with the type of availabledata & statistical model used
• Summarises the economic modelby computing suitable measures of“cost-effectiveness”
• Dictates the best course ofactions, given current evidence
• Standardised process
• Assesses the impact of uncertainty (eg inparameters or model structure) on theeconomic results
• Mandatory in many jurisdictions (includingNICE, in the UK). Often termed “PSA”
• Fundamentally Bayesian!
Uncertaintyanalysis
• Assesses the impact of uncertainty (egin parameters or model structure) onthe economic results
• Mandatory in many jurisdictions(including NICE, in the UK)
• Fundamentally Bayesian!
“Two-stage approach”: Spiegelhalter, Abrams & Myles, 2004
1. Estimation (base-case)
θ
yp(y | θ)
θ = f(Y )
2. Probabilistic sensitivity analysis
⇒
θp(θ) ! g(θ)
Estimation & PSA (one stage)
θ
yp(y | θ)
p(θ) p(θ | y)
“Integrated approach”: Spiegelhalter, Abrams & Myles (2004)Baio, Berardi & Heath (2017)
∆e = fe(θ)
∆c = fc(θ)
. . .
ICER = g(∆e,∆c)
EIB = h(∆e,∆c; k)
. . .
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 4 / 21
Health technology assessment (HTA)
Objective: Combine costs & benefits of a given intervention (against at least onecomparator) into a rational scheme for allocating resources
Statisticalmodel
Economicmodel
Decisionanalysis
Uncertaintyanalysis
• Estimates relevant populationparameters θ
• Varies with the type ofavailable data (& statisticalapproach!)
• Combines the parameters to obtaina population average measure forcosts and clinical benefits
• Varies with the type of availabledata & statistical model used
• Summarises the economic modelby computing suitable measures of“cost-effectiveness”
• Dictates the best course ofactions, given current evidence
• Standardised process
• Assesses the impact of uncertainty (eg inparameters or model structure) on theeconomic results
• Mandatory in many jurisdictions (includingNICE, in the UK). Often termed “PSA”
• Fundamentally Bayesian!
Uncertaintyanalysis
• Assesses the impact of uncertainty (egin parameters or model structure) onthe economic results
• Mandatory in many jurisdictions(including NICE, in the UK)
• Fundamentally Bayesian!
“Two-stage approach”: Spiegelhalter, Abrams & Myles, 2004
1. Estimation (base-case)
θ
yp(y | θ)
θ = f(Y )
2. Probabilistic sensitivity analysis
⇒
θp(θ) ! g(θ)
Estimation & PSA (one stage)
θ
yp(y | θ)
p(θ) p(θ | y)
“Integrated approach”: Spiegelhalter, Abrams & Myles (2004)Baio, Berardi & Heath (2017)
∆e = fe(θ)
∆c = fc(θ)
. . .
ICER = g(∆e,∆c)
EIB = h(∆e,∆c; k)
. . .
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 4 / 21
Health technology assessment (HTA)
Objective: Combine costs & benefits of a given intervention (against at least onecomparator) into a rational scheme for allocating resources
Statisticalmodel
Economicmodel
Decisionanalysis
Uncertaintyanalysis
• Estimates relevant populationparameters θ
• Varies with the type ofavailable data (& statisticalapproach!)
• Combines the parameters to obtaina population average measure forcosts and clinical benefits
• Varies with the type of availabledata & statistical model used
• Summarises the economic modelby computing suitable measures of“cost-effectiveness”
• Dictates the best course ofactions, given current evidence
• Standardised process
• Assesses the impact of uncertainty (eg inparameters or model structure) on theeconomic results
• Mandatory in many jurisdictions (includingNICE, in the UK). Often termed “PSA”
• Fundamentally Bayesian!
Uncertaintyanalysis
• Assesses the impact of uncertainty (egin parameters or model structure) onthe economic results
• Mandatory in many jurisdictions(including NICE, in the UK)
• Fundamentally Bayesian!
“Two-stage approach”: Spiegelhalter, Abrams & Myles, 2004
1. Estimation (base-case)
θ
yp(y | θ)
θ = f(Y )
2. Probabilistic sensitivity analysis
⇒
θp(θ) ! g(θ)
Estimation & PSA (one stage)
θ
yp(y | θ)
p(θ) p(θ | y)
“Integrated approach”: Spiegelhalter, Abrams & Myles (2004)Baio, Berardi & Heath (2017)
∆e = fe(θ)
∆c = fc(θ)
. . .
ICER = g(∆e,∆c)
EIB = h(∆e,∆c; k)
. . .
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 4 / 21
Knowledge is power?... (A tale of two stupid examples)
• Example 1: Intervention t = 1 is the most cost-effective, givencurrent evidence
– Pr(t = 1 is cost-effective) = 0.51– If we get it wrong: Increase in costs = £3
If we get it wrong: Decrease in effectiveness = 0.000001 QALYs– Large uncertainty/negligible consequences ⇒ can afford
uncertainty– Best course of action: make a decision now (on the basis of
current evidence)
• Example 2: Intervention t = 1 is the most cost-effective, givencurrent evidence
– Pr(t = 1 is cost-effective) = 0.999– If we get it wrong: Increase in costs = £1 000 000 000
If we get it wrong: Decrease in effectiveness = 999999 QALYs– Tiny uncertainty/dire consequences ⇒ probably should think
about it...– Best course of action: gather more evidence before making a
decision
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 5 / 21
Knowledge is power?... (A tale of two stupid examples)
• Example 1: Intervention t = 1 is the most cost-effective, givencurrent evidence
– Pr(t = 1 is cost-effective) = 0.51– If we get it wrong: Increase in costs = £3
If we get it wrong: Decrease in effectiveness = 0.000001 QALYs– Large uncertainty/negligible consequences ⇒ can afford
uncertainty– Best course of action: make a decision now (on the basis of
current evidence)
• Example 2: Intervention t = 1 is the most cost-effective, givencurrent evidence
– Pr(t = 1 is cost-effective) = 0.999– If we get it wrong: Increase in costs = £1 000 000 000
If we get it wrong: Decrease in effectiveness = 999999 QALYs– Tiny uncertainty/dire consequences ⇒ probably should think
about it...– Best course of action: gather more evidence before making a
decision
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 5 / 21
Knowledge is power?... (A tale of two stupid examples)
• Example 1: Intervention t = 1 is the most cost-effective, givencurrent evidence
– Pr(t = 1 is cost-effective) = 0.51– If we get it wrong: Increase in costs = £3
If we get it wrong: Decrease in effectiveness = 0.000001 QALYs– Large uncertainty/negligible consequences ⇒ can afford
uncertainty– Best course of action: make a decision now (on the basis of
current evidence)
• Example 2: Intervention t = 1 is the most cost-effective, givencurrent evidence
– Pr(t = 1 is cost-effective) = 0.999– If we get it wrong: Increase in costs = £1 000 000 000
If we get it wrong: Decrease in effectiveness = 999999 QALYs– Tiny uncertainty/dire consequences ⇒ probably should think
about it...– Best course of action: gather more evidence before making a
decision
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 5 / 21
Value of Information (VoI): Basic idea
• A new study will provide new data– Reducing (ideally, eliminating) uncertainty in a subset of model parameters
• Update the cost-effectiveness model– If the optimal decision changes, gain in monetary net benefit (NB = utility) from
using new optimal treatment– If optimal decision unchanged, no gain in NB
• Expected VOI is the average gain in NB
1 Expected Value of Perfect Information (EVPI)
– Value of completely resolving uncertainty in all input parameters to decision model– Infinite-sized long-term follow-up trial measuring everything!– Gives an upper-bound on the value of new study — if EVPI is low, suggests we can
make our decision based on existing information
2 Expected Value of Partial Perfect Information (EVPPI)
– Value of completely resolving uncertainty in subset of input parameters– Infinite-sized trial measuring relative effects on 1-year survival– Useful to identify which parameters responsible for decision uncertainty
3 Expected Value of Sample Information (EVSI)
– Value of reducing uncertainty by conducting a finite-sized study of given design– Can compare the benefits and costs of different designs– Is the proposed study likely to be a good use of resources? What is the optimal design?
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 6 / 21
Value of Information (VoI): Basic idea and relevant measures
• A new study will provide new data– Reducing (ideally, eliminating) uncertainty in a subset of model parameters
• Update the cost-effectiveness model– If the optimal decision changes, gain in monetary net benefit (NB = utility) from
using new optimal treatment– If optimal decision unchanged, no gain in NB
• Expected VOI is the average gain in NB
1 Expected Value of Perfect Information (EVPI)– Value of completely resolving uncertainty in all input parameters to decision model– Infinite-sized long-term follow-up trial measuring everything!– Gives an upper-bound on the value of new study — if EVPI is low, suggests we can
make our decision based on existing information
2 Expected Value of Partial Perfect Information (EVPPI)
– Value of completely resolving uncertainty in subset of input parameters– Infinite-sized trial measuring relative effects on 1-year survival– Useful to identify which parameters responsible for decision uncertainty
3 Expected Value of Sample Information (EVSI)
– Value of reducing uncertainty by conducting a finite-sized study of given design– Can compare the benefits and costs of different designs– Is the proposed study likely to be a good use of resources? What is the optimal design?
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 6 / 21
Value of Information (VoI): Basic idea and relevant measures
• A new study will provide new data– Reducing (ideally, eliminating) uncertainty in a subset of model parameters
• Update the cost-effectiveness model– If the optimal decision changes, gain in monetary net benefit (NB = utility) from
using new optimal treatment– If optimal decision unchanged, no gain in NB
• Expected VOI is the average gain in NB
1 Expected Value of Perfect Information (EVPI)– Value of completely resolving uncertainty in all input parameters to decision model– Infinite-sized long-term follow-up trial measuring everything!– Gives an upper-bound on the value of new study — if EVPI is low, suggests we can
make our decision based on existing information
2 Expected Value of Partial Perfect Information (EVPPI)– Value of completely resolving uncertainty in subset of input parameters– Infinite-sized trial measuring relative effects on 1-year survival– Useful to identify which parameters responsible for decision uncertainty
3 Expected Value of Sample Information (EVSI)
– Value of reducing uncertainty by conducting a finite-sized study of given design– Can compare the benefits and costs of different designs– Is the proposed study likely to be a good use of resources? What is the optimal design?
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 6 / 21
Value of Information (VoI): Basic idea and relevant measures
• A new study will provide new data– Reducing (ideally, eliminating) uncertainty in a subset of model parameters
• Update the cost-effectiveness model– If the optimal decision changes, gain in monetary net benefit (NB = utility) from
using new optimal treatment– If optimal decision unchanged, no gain in NB
• Expected VOI is the average gain in NB
1 Expected Value of Perfect Information (EVPI)– Value of completely resolving uncertainty in all input parameters to decision model– Infinite-sized long-term follow-up trial measuring everything!– Gives an upper-bound on the value of new study — if EVPI is low, suggests we can
make our decision based on existing information
2 Expected Value of Partial Perfect Information (EVPPI)– Value of completely resolving uncertainty in subset of input parameters– Infinite-sized trial measuring relative effects on 1-year survival– Useful to identify which parameters responsible for decision uncertainty
3 Expected Value of Sample Information (EVSI)– Value of reducing uncertainty by conducting a finite-sized study of given design– Can compare the benefits and costs of different designs– Is the proposed study likely to be a good use of resources? What is the optimal design?
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 6 / 21
Value of Information (VoI): Basic idea and relevant measures
• A new study will provide new data– Reducing (ideally, eliminating) uncertainty in a subset of model parameters
• Update the cost-effectiveness model– If the optimal decision changes, gain in monetary net benefit (NB = utility) from
using new optimal treatment– If optimal decision unchanged, no gain in NB
• Expected VOI is the average gain in NB
1 Expected Value of Perfect Information (EVPI)– Value of completely resolving uncertainty in all input parameters to decision model– Infinite-sized long-term follow-up trial measuring everything!– Gives an upper-bound on the value of new study — if EVPI is low, suggests we can
make our decision based on existing information
2 Expected Value of Partial Perfect Information (EVPPI)– Value of completely resolving uncertainty in subset of input parameters– Infinite-sized trial measuring relative effects on 1-year survival– Useful to identify which parameters responsible for decision uncertainty
3 Expected Value of Sample Information (EVSI)– Value of reducing uncertainty by conducting a finite-sized study of given design– Can compare the benefits and costs of different designs– Is the proposed study likely to be a good use of resources? What is the optimal design?
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 6 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility (EU) if we were able to learn φ but not ψ– If we knew φ perfectly, best decision = the maximum of this EU– Of course we cannot learn φ perfectly, so take the expected value– And compare this with the maximum expected utility overall– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!
– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility (EU) if we were able to learn φ but not ψ
– If we knew φ perfectly, best decision = the maximum of this EU– Of course we cannot learn φ perfectly, so take the expected value– And compare this with the maximum expected utility overall– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!
– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility ( ) if we were able to learn φ but not ψ
– If we knew φ perfectly, best decision = the maximum of this EU– Of course we cannot learn φ perfectly, so take the expected value– And compare this with the maximum expected utility overall– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!
– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility ( ) if we were able to learn φ but not ψ– If we knew φ perfectly, best decision = the maximum of this EU
– Of course we cannot learn φ perfectly, so take the expected value– And compare this with the maximum expected utility overall– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!
– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility ( ) if we were able to learn φ but not ψ– If we knew φ perfectly, best decision = the maximum of this EU– Of course we cannot learn φ perfectly, so take the expected value
– And compare this with the maximum expected utility overall– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!
– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility ( ) if we were able to learn φ but not ψ– If we knew φ perfectly, best decision = the maximum of this EU– Of course we cannot learn φ perfectly, so take the expected value– And compare this with the maximum expected utility overall
– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!
– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
Summarising PSA + Research priority: Expected Value of Partial Perfect Information
• θ = all the model parameters; can be split into two subsets– The “parameters of interest” φ, e.g. prevalence of a disease, HRQL measures, length
of stay in hospital, ...– The “remaining parameters” ψ, e.g. cost of treatment with other established
medications,
• We are interested in quantifying the value of gaining more information on φ, whileleaving the current level of uncertainty on ψ unchanged
• In formulæ:– First, consider the expected utility ( ) if we were able to learn φ but not ψ– If we knew φ perfectly, best decision = the maximum of this EU– Of course we cannot learn φ perfectly, so take the expected value– And compare this with the maximum expected utility overall– This is the EVPPI!
EVPPI = Eφ
[maxt
Eψ|φ [NBt(θ)]
]−max
tEθ [NBt(θ)]
• That’s the difficult part!– Can do nested Monte Carlo, but takes forever to get accurate results– Recent methods based on Gaussian Process regression very efficient & quick!
Strong et al Medical Decision Making. 2014; 34(3): 311-26. http://savi.shef.ac.uk/SAVI/Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280. https://egon.stats.ucl.ac.uk/projects/EVSI/
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 7 / 21
EVPPI Brute force — Nested Monte Carlo
Assuming only two interventions, can consider INB(θ) = NB1(θ)− NB0(θ)
−20
0−
100
010
020
030
0
E(I
NB
|θ)
Parameter of interest, θ
Monte Carlo approach K=250, J=200Nested Monte Carlo (Sφ = 250, Sψ = 200)
Parameter of interest φ
Eψ|φ
[INB(θ
)]
Thanks to Mark Strong (slide stolen from “Summer School in Bayesian methods in health economics”)www.statistica.it/gianluca/teaching/summer-school
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Brute force — Nested Monte Carlo
Assuming only two interventions, can consider INB(θ) = NB1(θ)− NB0(θ)
−20
0−
100
010
020
030
0
E(I
NB
|θ)
Parameter of interest, θ
Monte Carlo approach K=250, J=200
Adopt new treatment Don't adopt
Nested Monte Carlo (Sφ = 250, Sψ = 200)
Parameter of interest φ
Eψ|φ
[INB(θ
)]
Thanks to Mark Strong (slide stolen from “Summer School in Bayesian methods in health economics”)www.statistica.it/gianluca/teaching/summer-school
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Model as a regression problem
• Instead of Nested Monte Carlo, can model as a regression problem
NBt(θ) = Eψ|φ [NBt(θ)] + ε, with ε ∼ Normal(0, σ2ε)
= gt(φ) + ε
“Data”: simulations for NBt(θ) as “response”• “Data”: simulations for φ as “covariates”
• NB: Only need S data points (= PSA simulations), instead of Sφ × Sψ!
π0 ρ β0 . . . σ η γ NB0(θ) NB1(θ)
0.365 0.076 0.243 . . . 0.622 0.001 0.162 19 214 751 19 647 706
0.421 0.024 0.115 . . . 0.519 0.010 0.134 17 165 526 17 163 407
0.125 0.017 0.420 . . . 0.482 0.007 0.149 18 710 928 16 458 433
0.117 0.073 0.419 . . . 0.317 0.003 0.120 16 991 321 18 497 648
0.481 0.008 0.176 . . . 0.497 0.004 0.191 19 772 898 18 662 329
0.163 0.127 0.227 . . . 0.613 0.083 0.004 17 106 136 18 983 331
. . . . . . . . . . . .
0.354 0.067 0.318 . . . 0.519 0.063 0.117 18 043 921 16 470 805
“covariates” “response” “response”
Strong et al Medical Decision Making. 2014; 34(3): 311-26
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Model as a regression problem
−20
0−
100
010
020
030
0
INB
Parameter of interest, θ
Regression approach K=2,000Regression approach S = 2000
Parameter of interest φ
INB(θ
)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Model as a regression problem
−20
0−
100
010
020
030
0
INB
Parameter of interest, θ
Regression approach K=2,000Regression approach S = 2000
Parameter of interest φ
INB(θ
)
−20
0−
100
010
020
030
0
E(I
NB
|θ)
Parameter of interest, θ
Regression approach K=2,000Regression approach S = 2000
Parameter of interest φ
Eψ|φ
[INB(θ
)]g(φ)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Model as a regression problem
• Instead of Nested Monte Carlo, can model as a regression problem
NBt(θ) = Eψ|φ [NBt(θ)] + ε, with ε ∼ Normal(0, σ2ε)
= gt(φ) + ε
“Data”: simulations for NBt(θ) as “response”• “Data”: simulations for φ as “covariates”
• Once the functions gt(φ) are estimated, then can approximate
EVPPI = Eφ[maxt
Eψ|φ [NBt(θ)]]−max
tEθ [NBt(θ)]
≈ 1
S
S∑s=1
maxtgt(φs)−max
t
1
S
S∑s=1
gt(φs)
• NB: gt(φ) can be complex, so need to use flexible regression methods
– GAMs: gt(φ) =
Qφ∑q=1
ht(φsq) ht(·) = smooth functions (cubic polynomials)
very fast, but only work if number of important parameters Qφ ≤ 5 (interactionsincrease model size exponentially!)
– If P > 5, can use Gaussian Process regression
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Model as a regression problem
−20
0−
100
010
020
030
0
INB
Parameter of interest, θ
Regression approach K=2,000Regression approach S = 2000
Parameter of interest φ
INB(θ
)
−20
0−
100
010
020
030
0
E(I
NB
|θ)
Parameter of interest, θ
Regression approach K=2,000 True relationship shown in red
Don't adopt
Adopt new treatment
Regression approach S = 2000 (True relationship in red)
Parameter of interest φ
Eψ|φ
[INB(θ
)]
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI Model as a regression problem
• Instead of Nested Monte Carlo, can model as a regression problem
NBt(θ) = Eψ|φ [NBt(θ)] + ε, with ε ∼ Normal(0, σ2ε)
= gt(φ) + ε
“Data”: simulations for NBt(θ) as “response”• “Data”: simulations for φ as “covariates”
• Once the functions gt(φ) are estimated, then can approximate
EVPPI = Eφ[maxt
Eψ|φ [NBt(θ)]]−max
tEθ [NBt(θ)]
≈ 1
S
S∑s=1
maxtgt(φs)−max
t
1
S
S∑s=1
gt(φs)
• NB: gt(φ) can be complex, so need to use flexible regression methods
– GAMs: gt(φ) =
Qφ∑q=1
ht(φsq) ht(·) = smooth functions (cubic polynomials)
very fast, but only work if number of important parameters Qφ ≤ 5 (interactionsincrease model size exponentially!)
– If P > 5, can use Gaussian Process regression
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 8 / 21
EVPPI via GP regression
Model NBt(θ1)NBt(θ2)
...NBt(θS)
:= NBt ∼ Normal(Hβ,CExp + σ2εI)
H =
1 φ11 · · · φ1P
1 φ21 · · · φ2P
.... . .
1 φS1 · · · φSP
and CExp(r, s) = σ2 exp
[P∑p=1
(φrp − φsp
δp
)2]
• Parameters: β, δ, σ2, σ2ε
• Very flexible structure — good approximation level
• Can use conjugate priors + numerical optimisation
• But: can still be very slow — computational cost in the order of S3 (involvesinversion of a dense covariance matrix)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 9 / 21
EVPPI via GP regression — but faster
1 Build from ideas in spatial statistics and use a Matern covariance function
CM(r, s) =σ2
Γ(ν)2ν−1(κ‖φr − φs‖)νKν(κ‖φr − φs‖)
– Fewer parameters, but still implies a dense covariance matrix– But: can use efficient algorithms to solve Stochastic Partial Differential Equations
(SPDE) to approximate it — with computational cost ∝ S3/2!
2 Re-formulate the model as
NBt ∼ Normal(Hβ,CM + σ2εI)
∼ Normal(Hβ + f(ω), σ2εI)
– f(ω) are a set of “spatially structured” effects, with ω ∼ Normal(0,Q−1(ξ)
)– Q(ξ) is a sparse precision matrix determined by the SPDE solution– Project the P−dimensional information in φ to a d−dimensional space (d << P ,
ideally d = 2) — can use “Principal Fitted Components”
3 Crucially, if we set a sparse Gaussian prior on β, this is a Latent Gaussian model ⇒can estimate using super-fast Integrated Nested Laplace Approximation (INLA)
NB: Both GP-based methods implemented in the R package BCEA (Bayesian Cost-Effectiveness Analysis)http://www.statistica.it/gianluca/BCEA https://github.com/giabaio/BCEA
Heath et al Stats in Med. 2016; 35(23): 4264-4280
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 10 / 21
EVPPI via GP regression — but faster
1 Build from ideas in spatial statistics and use a Matern covariance function
CM(r, s) =σ2
Γ(ν)2ν−1(κ‖φr − φs‖)νKν(κ‖φr − φs‖)
– Fewer parameters, but still implies a dense covariance matrix– But: can use efficient algorithms to solve Stochastic Partial Differential Equations
(SPDE) to approximate it — with computational cost ∝ S3/2!
2 Re-formulate the model as
NBt ∼ Normal(Hβ,CM + σ2εI)
∼ Normal(Hβ + f(ω), σ2εI)
– f(ω) are a set of “spatially structured” effects, with ω ∼ Normal(0,Q−1(ξ)
)– Q(ξ) is a sparse precision matrix determined by the SPDE solution– Project the P−dimensional information in φ to a d−dimensional space (d << P ,
ideally d = 2) — can use “Principal Fitted Components”
3 Crucially, if we set a sparse Gaussian prior on β, this is a Latent Gaussian model ⇒can estimate using super-fast Integrated Nested Laplace Approximation (INLA)
NB: Both GP-based methods implemented in the R package BCEA (Bayesian Cost-Effectiveness Analysis)http://www.statistica.it/gianluca/BCEA https://github.com/giabaio/BCEA
Heath et al Stats in Med. 2016; 35(23): 4264-4280; Lindgren et al JRSS/B. 2011; 73(4): 423-498
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 10 / 21
EVPPI via GP regression — but faster
1 Build from ideas in spatial statistics and use a Matern covariance function
CM(r, s) =σ2
Γ(ν)2ν−1(κ‖φr − φs‖)νKν(κ‖φr − φs‖)
– Fewer parameters, but still implies a dense covariance matrix– But: can use efficient algorithms to solve Stochastic Partial Differential Equations
(SPDE) to approximate it — with computational cost ∝ S3/2!
2 Re-formulate the model as
NBt ∼ Normal(Hβ,CM + σ2εI)
∼ Normal(Hβ + f(ω), σ2εI)
– f(ω) are a set of “spatially structured” effects, with ω ∼ Normal(0,Q−1(ξ)
)– Q(ξ) is a sparse precision matrix determined by the SPDE solution– Project the P−dimensional information in φ to a d−dimensional space (d << P ,
ideally d = 2) — can use “Principal Fitted Components”
3 Crucially, if we set a sparse Gaussian prior on β, this is a Latent Gaussian model ⇒can estimate using super-fast Integrated Nested Laplace Approximation (INLA)
NB: Both GP-based methods implemented in the R package BCEA (Bayesian Cost-Effectiveness Analysis)http://www.statistica.it/gianluca/BCEA https://github.com/giabaio/BCEA
Heath et al Stats in Med. 2016; 35(23): 4264-4280; Lindgren et al JRSS/B. 2011; 73(4): 423-498
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 10 / 21
EVPPI via GP regression — but faster
1 Build from ideas in spatial statistics and use a Matern covariance function
CM(r, s) =σ2
Γ(ν)2ν−1(κ‖φr − φs‖)νKν(κ‖φr − φs‖)
– Fewer parameters, but still implies a dense covariance matrix– But: can use efficient algorithms to solve Stochastic Partial Differential Equations
(SPDE) to approximate it — with computational cost ∝ S3/2!
2 Re-formulate the model as
NBt ∼ Normal(Hβ,CM + σ2εI)
∼ Normal(Hβ + f(ω), σ2εI)
– f(ω) are a set of “spatially structured” effects, with ω ∼ Normal(0,Q−1(ξ)
)– Q(ξ) is a sparse precision matrix determined by the SPDE solution– Project the P−dimensional information in φ to a d−dimensional space (d << P ,
ideally d = 2) — can use “Principal Fitted Components”
3 Crucially, if we set a sparse Gaussian prior on β, this is a Latent Gaussian model ⇒can estimate using super-fast Integrated Nested Laplace Approximation (INLA)
NB: Both GP-based methods implemented in the R package BCEA (Bayesian Cost-Effectiveness Analysis)http://www.statistica.it/gianluca/BCEA https://github.com/giabaio/BCEA
Heath et al Stats in Med. 2016; 35(23): 4264-4280; Lindgren et al JRSS/B. 2011; 73(4): 423-498; Rue et al JRSS/B. 2009; 71: 319-392
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 10 / 21
Examples SAVI
●
●
●
●
●
●
●
●
●
●
●●
Running time (secs)
Number of important parameters
Tim
e (s
ecs)
●● ●
●●
● ●●
●● ● ●
17
14
18
21
26
31
37
47
52
66
70 71
6 7 7 6 7 6 68
5 6 6 6
5 6 7 8 9 10 11 12 13 14 15 16
1020
3040
5060
70
●
●
GP regressionINLA−SPDE
● ● ● ● ● ●
●
●
● ● ● ●
Estimated values
Number of important parametersE
VP
PI e
stim
ates
5 6 7 8 9 10 11 12 13 14 15 16
1400
1600
1800
2000
1280 1280 1290 1290 1290
2100 2100 2100 2100
1270
1460
1690
1280
1480
1710
●
●
●
Both methodsGPSPDE−INLA
Running time for a single value of k (willingness to pay)
• Fictional decision tree model with correlated parameters
• 2 treatment options and overall 19 parameters
• Parameters simulated from multivariate Normal distribution, so can compute exact EVPPI
Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 12 / 21
Examples Vaccine
●
● ●●
●●
●●
●
●
●
●
Running time (secs)
Number of important parameters
Tim
e (s
ecs)
● ● ● ● ● ● ● ● ● ● ● ●17
42 4557
7486
7060
84
188
470
121
7 7 7 7 7 6 6 7 6 7 6 7
5 6 7 8 9 10 11 12 13 14 15 16
010
020
030
040
0
●
●
GP regressionINLA−SPDE
● ●
●
● ●
●
● ●●
●● ●
Estimated values
Number of important parametersE
VP
PI e
stim
ates
● ●
●
● ●
● ● ● ●
●
● ●
1.14 1.14
1.22
1.36 1.36
1.55
1.39 1.391.4
1.471.48 1.48
1.11 1.11
1.17
1.32 1.32
1.34 1.34 1.34 1.34
1.4
1.43 1.43
5 6 7 8 9 10 11 12 13 14 15 16
1.2
1.3
1.4
1.5
●
●
GP regressionINLA−SPDE
Running time for a single value of k (willingness to pay)
• Cost-effectiveness model for influenza vaccine based on evidence synthesis
• 2 treatment options and overall 63 parameters
• Model not available in closed form (needs MCMC simulations)
Heath et al Statistics in Medicine. 2016; 35(23): 4264-4280
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 12 / 21
Breaking bad...
Breast cancer screening (Welton et al. 2008. JRSS/A)
• Multi-decision model developed for the UK setting, with 4 interventions
• Complex evidence synthesis for 6 parameters — highly structured!
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
● ●
●
●
●●
●
●
●●●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●●
●
●
● ●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●● ●
●● ●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
● ●●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
● ●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●● ●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
● ●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
−6 −4 −2 0 2
−0.
15−
0.10
−0.
050.
000.
050.
100.
15
Residual plot for costs
Fitted values
Res
idua
ls
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●●
●
●●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
● ●
●●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●● ●
●
●
●●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
● ●
●●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●●
●●
●
●
●
●●
●
●
●
● ●
●
●
●
●●
●
●
●
● ●
●
●
●
●
●●
●●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
● ●
● ●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●● ●
●●●
●
● ●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
● ●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
● ●
●
●
● ●
●
● ●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●●
●
●●
●
●
●
●
●
● ●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●●
●●
●
●
●
●
●
●
●
●
●●
●●
●●●
●
●
●
●
●●
●
●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●●
●
●
●●● ●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●●
●
●
●●
●
●
●
●●
●
●
●
●
●●
●
●
●●
●
●●●
●●
●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
● ●●
●●●
●
●
●
●
● ●
●
●
●
●
●
●
●
● ●
●
●
●
●●
● ●
●●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●●●●
●
●
●
●●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●●
●●
●●
●
●
●
●●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●●●
●
●
●● ●
●
●
●
●
●
●
●●
●
●
●
●
●
●
● ●●
●
●
●
●
●
●
●
● ●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●●
●●
●
●●●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●●
●
●
●
●
●
●
●
● ● ●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●●
●●
●
●●
●
●
●
●●
●
●
●
●●
● ●
●
●
●
●
●
●
●
●
●
−0.0005 0.0000 0.0005
−0.
0000
06−
0.00
0002
0.00
0002
0.00
0006
Residual plot for effects
Fitted values
Res
idua
ls
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 13 / 21
The fix! Skip to conclusions
• Can relatively easily modify the basic structure of the model, e.g. include interactionterms to make Hβ non-linear
β0 + β1φ1s + β2φ2s + β3φ3s + β4φ1sφ2s + β5φ1sφ3s + β6φ2sφ3s
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
● ●
●
●
●●
●
●
●●●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●●
●
●
● ●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●● ●
●● ●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
● ●●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
● ●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●● ●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
● ●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
−6 −4 −2 0 2
−0.
15−
0.10
−0.
050.
000.
050.
100.
15Residual plot for costs
Fitted values
Res
idua
ls
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
● ● ●
●
●
●
●
● ●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
● ●●
●
● ●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
−0.0005 0.0000 0.0005−0.
0000
0008
−0.
0000
0002
0.00
0000
020.
0000
0006
Residual plot for effects
Fitted values
Res
idua
ls
Baio et al (2017). Bayesian Cost-Effectiveness Analysis with the R package BCEA
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 14 / 21
Value of Information (VoI): Basic idea and relevant measures
• A new study will provide new data– Reducing (ideally, eliminating) uncertainty in a subset of model parameters
• Update the cost-effectiveness model– If the optimal decision changes, gain in monetary net benefit (NB = utility) from
using new optimal treatment– If optimal decision unchanged, no gain in NB
• Expected VOI is the average gain in NB
1 Expected Value of Perfect Information (EVPI)– Value of completely resolving uncertainty in all input parameters to decision model– Infinite-sized long-term follow-up trial measuring everything!– Gives an upper-bound on the value of new study — if EVPI is low, suggests we can
make our decision based on existing information
2 Expected Value of Partial Perfect Information (EVPPI)– Value of completely resolving uncertainty in subset of input parameters– Infinite-sized trial measuring relative effects on 1-year survival– Useful to identify which parameters responsible for decision uncertainty
3 Expected Value of Sample Information (EVSI)– Value of reducing uncertainty by conducting a finite-sized study of given design– Can compare the benefits and costs of different designs– Is the proposed study likely to be a good use of resources? What is the optimal design?
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 15 / 21
Research priority: Expected Value of Sample Information
• EVSI measures the value of reducing uncertainty by running a study of a given design
• Can compare the benefits and costs of a study with given design– To see if a proposed study likely to be a good use of resources– To find the optimal study design
Literature Review
Clinical Trials
Expert Opinion
Model Inputs
θ
Value of OptimalTreatment: $C
Current DecisionMaking Process
PotentialStudy
“Posterior”Model Inputs
θ | X
Value of OptimalTreatment: $F
VSI = $F - $C
One Possible FutureDecision Making ProcessPrior
Data
X | θ
θ definemodel for X
EconomicModel
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 16 / 21
Research priority: Expected Value of Sample Information
• EVSI measures the value of reducing uncertainty by running a study of a given design
• Can compare the benefits and costs of a study with given design– To see if a proposed study likely to be a good use of resources– To find the optimal study design
EVSI = EX
maxt
Eθ|X [NBt(θ)]︸ ︷︷ ︸Value of decision based on
sample information(for a given study design)
− max
tEθ [NBt(θ)]︸ ︷︷ ︸
Value of decision based oncurrent information
Prior predictivedistribution
(pre-posterior)
Posterior given data X
• Computationally complex– Requires specific knowledge of the model for (future/hypothetical) data collection– Again, recent methods have improved efficiency (Heath et al. 2019; https://arxiv.org/abs/1905.12013)
• Can be used to drive design of new study (eg sample size calculations)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 16 / 21
Moment matching EVSI = EX[maxt Eθ|X [NBt(θ)]
]−maxt Eθ [NBt(θ)]
Objective: Estimate the distribution p(µX) with µX = Eθ|X [INB(θ)]
• That’s the hard part to estimate the EVSI
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 17 / 21
Moment matching EVSI = EX[maxt Eθ|X [NBt(θ)]
]−maxt Eθ [NBt(θ)]
Objective: Estimate the distribution p(µX) with µX = Eθ|X [INB(θ)]
• That’s the hard part to estimate the EVSI
We know that
1 EX[µX]
= EX[Eθ|X [INB(θ)]
]= Eθ [INB(θ)]
2 VarX[µX]
= Varθ [INB(θ)]︸ ︷︷ ︸PSA variance for INB(θ)
− EX[Varθ|X [INB(θ)]
]︸ ︷︷ ︸Posterior variance for INB(θ)
3 As n→∞, p(µX) is “similar” to the PSA distribution of INB(θ)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 17 / 21
Moment matching EVSI = EX[maxt Eθ|X [NBt(θ)]
]−maxt Eθ [NBt(θ)]
Objective: Estimate the distribution p(µX) with µX = Eθ|X [INB(θ)]
• That’s the hard part to estimate the EVSI
We know that
1 EX[µX]
= EX[Eθ|X [INB(θ)]
]= Eθ [INB(θ)]
2 VarX[µX]
= Varθ [INB(θ)]︸ ︷︷ ︸PSA variance for INB(θ)
− EX[Varθ|X [INB(θ)]
]︸ ︷︷ ︸Posterior variance for INB(θ)
3 As n→∞, p(µX) is “similar” to the PSA distribution of INB(θ)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 17 / 21
Moment matching A counterintuitive relationship...
−1000 −500 0 500 1000 1500
0.00
000.
0005
0.00
100.
0015
Potential means
Den
sity
N=2
N=3
N=4
N=6
N=10
N=25N=100
INB
• If n→∞, we will have learned the value of θ perfectly– But we haven’t seen the new (infinite sized) data yet — so current uncertainty
described by the prior
• If n→ 0, there are fewer values we can learn from future data– So pre-posterior more concentrated away from the prior
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 17 / 21
Moment matching
Objective: Estimate the distribution p(µX) with µX = Eθ|X [INB(θ)]
• That’s the hard part to estimate the EVSI
We know that
1 EX[µX]
= EX[Eθ|X [INB(θ)]
]= Eθ [INB(θ)]
2 VarX[µX]
= Varθ [INB(θ)]︸ ︷︷ ︸PSA variance for INB(θ)
− EX[Varθ|X [INB(θ)]
]︸ ︷︷ ︸Posterior variance for INB(θ)
3 As n→∞, p(µX) is “similar” to the PSA distribution of INB(θ)
Idea: can approximate the unknown distribution p(µX) by rescaling the PSA distributionfor INB(θ), moment-matching it to the mean and variance defined above
• All we need is to estimate the expected posterior variance...
• Can do this efficiently by only using Q ≈ 30 to 50 << S PSA simulations!
Heath et al. 2017. Medical Decision Making. 38(2): 163-173
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 17 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
4 Estimate the sample varianceσ2q = Varθ|Xq [INB(θ)] σ2
q
Varθ [INB(θ)]
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
4 Estimate the sample varianceσ2q = Varθ|Xq [INB(θ)]
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
4 Estimate the sample varianceσ2q = Varθ|Xq [INB(θ)]
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Estimating the Expected Posterior Variance
1 Select q = 1, . . . , Q values out ofthe S PSA simulations for θ
2 Simulate data Xq from p(X |θ)
3 Run the model to estimatep(θ | X) and simulate values forINB(θ |Xq)
4 Estimate the sample varianceσ2q = Varθ|Xq [INB(θ)]
5 Use the Q estimates for σ2q
to estimate the expected pos-terior variance
VarX
[Eθ|X [INB(θ)]
]︸ ︷︷ ︸
σ2X
= Varθ [INB(θ)]︸ ︷︷ ︸σ2
− EX
[Varθ|X [INB(θ)]
]︸ ︷︷ ︸
1Q
∑Qq=1 σ
2q
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Rescaling and computing the EVSI
• Can now rescale the original PSA samples for INB(θ) to ensure that mean andvariance now match the computed values
ηX = f(µX) = INB(θ(s)
)√σ2X
σ2+ µ
(1−
√σ2X
σ2
)
– INB(θ(s)
)= s−th PSA simulation for the INB
– µ = Eθ [INB(θ)] = PSA average INB– σ2 = PSA variance of the INB
and finally estimate the EVSI as
EVSI =1
S
S∑s=1
max{
0, ηX}−max {0, µ}
• Can also compute conditional version for φ ∈ θ. “Simply” substitute
– σ2 with σ2φ = PSA variance for conditional INB (obtained using analysis of EVPPI)
– INB(θ(s)
)with INB
(φ(s)
)= Eψ|φ
[INB
(θ(s)
)]
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Rescaling and computing the EVSI
A Small Technicality...
• Only the focal parameters φ will be informed by the future study
• The distribution of µX is similar to that induced by the EVPPI analysis!
Den
sity
−40000 0 20000 40000 60000
0e+
001e
−05
2e−
053e
−05
Den
sity
2000 4000 6000 8000 10000
0e+
001e
−04
2e−
043e
−04
INB(θ) INB(φ) = Eψ|φ [INB(θ)]
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Moment matching Rescaling and computing the EVSI
• Can now rescale the original PSA samples for INB(θ) to ensure that mean andvariance now match the computed values
ηX = f(µX) = INB(θ(s)
)√σ2X
σ2+ µ
(1−
√σ2X
σ2
)
– INB(θ(s)
)= s−th PSA simulation for the INB
– µ = Eθ [INB(θ)] = PSA average INB– σ2 = PSA variance of the INB
and finally estimate the EVSI as
EVSI =1
S
S∑s=1
max{
0, ηX}−max {0, µ}
• Can also compute conditional version for φ ∈ θ. “Simply” substitute– σ2 with σ2
φ = PSA variance for conditional INB (obtained using analysis of EVPPI)
– INB(θ(s)
)with INB
(φ(s)
)= Eψ|φ
[INB
(θ(s)
)]Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 18 / 21
Research priority: Expected Value of Sample Information
●
0 500 1000 1500
0.0
0.5
1.0
1.5
Sample Size
Per
Per
son
EV
SI
0.0250.250.50.750.975EVPPI
5 10 15 20 250
2000
4000
6000
8000
1000
0
Probability of Cost−Effective Trial
Time Horizon
Inci
denc
e P
opul
atio
n
Prob=0
Prob=.5
Prob=1
https://github.com/giabaio/EVSIhttps://egon.stats.ucl.ac.uk/projects/EVSIHeath et al Value in Health, 2018; 21(11): 1299-1304Heath et al Medical Decision Making, 2017; 38(2): 163-173
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 19 / 21
Research priority: Expected Value of Sample Information
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 19 / 21
Conclusions
• VoI can be very valuable in driving the whole economic evaluation process– Summarising PSA (in addition to standard tools, eg CEAC)– Research priority (in place of standard tools, eg sample size calculations?)
• Historically limited use — also for computational complexity– Computation still a crucial component — but this is the price to pay for increasingly
realistic and complex models?– Things can only get better(?) — recent research has improved this massively!
• Need standardised softward to enable practitioners to use the new tools– And to move from Excel-based modelling to using fully proper statistical software
(eg R): http://www.statistica.it/gianluca/teaching/r-hta-workshop/2019/
– Packages and web-applications exist to do this: SAVI, BCEA, BCEAweb, ...
• Can we improve in any way by being more objective in our priors?– EVPPI with full GP based on conjugacy, while INLA/SPDE can get away with
relatively flat priors (but has to fit LGM structure)– EVSI with Moment Matching has a bit more scope for subjective knowledge &
informative priors — but (so far!) we’ve not experienced unwanted impact of priors...
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 20 / 21
Conclusions
• VoI can be very valuable in driving the whole economic evaluation process– Summarising PSA (in addition to standard tools, eg CEAC)– Research priority (in place of standard tools, eg sample size calculations?)
• Historically limited use — also for computational complexity– Computation still a crucial component — but this is the price to pay for increasingly
realistic and complex models?– Things can only get better(?) — recent research has improved this massively!
• Need standardised softward to enable practitioners to use the new tools– And to move from Excel-based modelling to using fully proper statistical software
(eg R): http://www.statistica.it/gianluca/teaching/r-hta-workshop/2019/
– Packages and web-applications exist to do this: SAVI, BCEA, BCEAweb, ...
• Can we improve in any way by being more objective in our priors?– EVPPI with full GP based on conjugacy, while INLA/SPDE can get away with
relatively flat priors (but has to fit LGM structure)– EVSI with Moment Matching has a bit more scope for subjective knowledge &
informative priors — but (so far!) we’ve not experienced unwanted impact of priors...
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 20 / 21
Conclusions
• VoI can be very valuable in driving the whole economic evaluation process– Summarising PSA (in addition to standard tools, eg CEAC)– Research priority (in place of standard tools, eg sample size calculations?)
• Historically limited use — also for computational complexity– Computation still a crucial component — but this is the price to pay for increasingly
realistic and complex models?– Things can only get better(?) — recent research has improved this massively!
• Need standardised softward to enable practitioners to use the new tools– And to move from Excel-based modelling to using fully proper statistical software
(eg R): http://www.statistica.it/gianluca/teaching/r-hta-workshop/2019/
– Packages and web-applications exist to do this: SAVI, BCEA, BCEAweb, ...
• Can we improve in any way by being more objective in our priors?– EVPPI with full GP based on conjugacy, while INLA/SPDE can get away with
relatively flat priors (but has to fit LGM structure)– EVSI with Moment Matching has a bit more scope for subjective knowledge &
informative priors — but (so far!) we’ve not experienced unwanted impact of priors...
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 20 / 21
Conclusions
• VoI can be very valuable in driving the whole economic evaluation process– Summarising PSA (in addition to standard tools, eg CEAC)– Research priority (in place of standard tools, eg sample size calculations?)
• Historically limited use — also for computational complexity– Computation still a crucial component — but this is the price to pay for increasingly
realistic and complex models?– Things can only get better(?) — recent research has improved this massively!
• Need standardised softward to enable practitioners to use the new tools– And to move from Excel-based modelling to using fully proper statistical software
(eg R): http://www.statistica.it/gianluca/teaching/r-hta-workshop/2019/
– Packages and web-applications exist to do this: SAVI, BCEA, BCEAweb, ...
• Can we improve in any way by being more objective in our priors?– EVPPI with full GP based on conjugacy, while INLA/SPDE can get away with
relatively flat priors (but has to fit LGM structure)– EVSI with Moment Matching has a bit more scope for subjective knowledge &
informative priors — but (so far!) we’ve not experienced unwanted impact of priors...
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 20 / 21
Go raibh maith agat!
(Thank you!)
Gianluca Baio (UCL) Quick & clean VoI O’Bayes 2019, 1 Jul 2019 21 / 21