Home > Research > Publications & Outputs > Improving power by conditioning on less in post...

Links

Text available via DOI:

View graph of relations

Improving power by conditioning on less in post-selection inference for changepoints

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Improving power by conditioning on less in post-selection inference for changepoints. / Carrington, R.; Fearnhead, P.
In: Statistics and Computing, Vol. 35, No. 1, 8, 31.01.2025.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

APA

Vancouver

Carrington R, Fearnhead P. Improving power by conditioning on less in post-selection inference for changepoints. Statistics and Computing. 2025 Jan 31;35(1):8. Epub 2024 Dec 4. doi: 10.1007/s11222-024-10542-1

Author

Bibtex

@article{a6c95173838b4a22806d9f783f339a71,
title = "Improving power by conditioning on less in post-selection inference for changepoints",
abstract = "Post-selection inference has recently been proposed as a way of quantifying uncertainty about detected changepoints. The idea is to run a changepoint detection algorithm, and then re-use the same data to perform a test for a change near each of the detected changes. By defining the p-value for the test appropriately, so that it is conditional on the information used to choose the test, this approach will produce valid p-values. We show how to improve the power of these procedures by conditioning on less information. This gives rise to an ideal post-selection p-value that is intractable but can be approximated by Monte Carlo. We show that for any Monte Carlo sample size, this procedure produces valid p-values, and empirically that noticeable increase in power is possible with only very modest Monte Carlo sample sizes. Our procedure is easy to implement given existing post-selection inference methods, as we just need to generate perturbations of the data set and re-apply the post-selection method to each of these. On genomic data consisting of human GC content, our procedure increases the number of significant changepoints that are detected when compared to the method of Jewell et al. (J R Stat Soc Ser B 84(4):1082-1104, 2022). ",
keywords = "Binary segmentation, Breakpoint, Fused lasso, Penalised likelihood, Post-selection p-value",
author = "R. Carrington and P. Fearnhead",
year = "2025",
month = jan,
day = "31",
doi = "10.1007/s11222-024-10542-1",
language = "English",
volume = "35",
journal = "Statistics and Computing",
issn = "0960-3174",
publisher = "Springer",
number = "1",

}

RIS

TY - JOUR

T1 - Improving power by conditioning on less in post-selection inference for changepoints

AU - Carrington, R.

AU - Fearnhead, P.

PY - 2025/1/31

Y1 - 2025/1/31

N2 - Post-selection inference has recently been proposed as a way of quantifying uncertainty about detected changepoints. The idea is to run a changepoint detection algorithm, and then re-use the same data to perform a test for a change near each of the detected changes. By defining the p-value for the test appropriately, so that it is conditional on the information used to choose the test, this approach will produce valid p-values. We show how to improve the power of these procedures by conditioning on less information. This gives rise to an ideal post-selection p-value that is intractable but can be approximated by Monte Carlo. We show that for any Monte Carlo sample size, this procedure produces valid p-values, and empirically that noticeable increase in power is possible with only very modest Monte Carlo sample sizes. Our procedure is easy to implement given existing post-selection inference methods, as we just need to generate perturbations of the data set and re-apply the post-selection method to each of these. On genomic data consisting of human GC content, our procedure increases the number of significant changepoints that are detected when compared to the method of Jewell et al. (J R Stat Soc Ser B 84(4):1082-1104, 2022).

AB - Post-selection inference has recently been proposed as a way of quantifying uncertainty about detected changepoints. The idea is to run a changepoint detection algorithm, and then re-use the same data to perform a test for a change near each of the detected changes. By defining the p-value for the test appropriately, so that it is conditional on the information used to choose the test, this approach will produce valid p-values. We show how to improve the power of these procedures by conditioning on less information. This gives rise to an ideal post-selection p-value that is intractable but can be approximated by Monte Carlo. We show that for any Monte Carlo sample size, this procedure produces valid p-values, and empirically that noticeable increase in power is possible with only very modest Monte Carlo sample sizes. Our procedure is easy to implement given existing post-selection inference methods, as we just need to generate perturbations of the data set and re-apply the post-selection method to each of these. On genomic data consisting of human GC content, our procedure increases the number of significant changepoints that are detected when compared to the method of Jewell et al. (J R Stat Soc Ser B 84(4):1082-1104, 2022).

KW - Binary segmentation

KW - Breakpoint

KW - Fused lasso

KW - Penalised likelihood

KW - Post-selection p-value

U2 - 10.1007/s11222-024-10542-1

DO - 10.1007/s11222-024-10542-1

M3 - Journal article

VL - 35

JO - Statistics and Computing

JF - Statistics and Computing

SN - 0960-3174

IS - 1

M1 - 8

ER -