Observations are generated according to a regression with normal error as a function of time,when the process is in control. The process potentially changes at some unknown point oftime and then the ensuing observations are normal with the same mean function plus an arbitrary function under suitable regularity conditions. The problem is to obtain a stopping rule that is optimal in the sense that the rule minimizes the expected delay in detecting a change subject to a constraint on the average run length to a false alarm. A bound on the expected delay is first obtained. It is then shown that the cusum and Shiryayev–Roberts procedures achieve this bound to first order.
"Detecting a change in regression: first-order optimality." Ann. Statist. 27 (6) 1896 - 1913, December 1999. https://doi.org/10.1214/aos/1017939243