hildensia/bayesian_changepoint_detection

Scaling of Data

stefan37 opened this issue · 3 comments

Hi,
I've noticed is the scaling of the data can have an effect on the result, but I am not sure why it would and can't find any reason for it in the code or references. Below I have the CP probabilities for the same data with or without a constant factor, which are somewhat different.

Are there some assumptions about the input data I am missing?
Thanks

image
image

The student T likelihood scales with the squared mean distance, which is non-linear w.r.t. to data scaling.

Also intuitively that makes sense, because the difference of your generative models is now different, and thus the probability of them being the same/different should be different

Thanks for quick reply. Confusion for me is that often the scale is arbitrary, such as if there are multiple ways to make some data dimensionless, yet that could yield vastly different results; my assumption before was that I should always just always normalize over the entire time series. Is there some prior used here in calculating the student T likelihood that I should keep in mind with how I scale my data, or any other way to decide the scale?

Good question. I would believe that mean centering your data is probably a good idea. But w.r.t. scaling I have to think a bit more. It has probably to do with an implicit prior somewhere, but I cannot pinpoint it right now.