Unveiling low-dimensional patterns induced by convex non-differentiable regularizers

التفاصيل البيبلوغرافية
العنوان: Unveiling low-dimensional patterns induced by convex non-differentiable regularizers
المؤلفون: Hejný, Ivan, Wallin, Jonas, Bogdan, Małgorzata, Kos, Michał
سنة النشر: 2024
المجموعة: Mathematics
Statistics
مصطلحات موضوعية: Mathematics - Statistics Theory
الوصف: Popular regularizers with non-differentiable penalties, such as Lasso, Elastic Net, Generalized Lasso, or SLOPE, reduce the dimension of the parameter space by inducing sparsity or clustering in the estimators' coordinates. In this paper, we focus on linear regression and explore the asymptotic distributions of the resulting low-dimensional patterns when the number of regressors $p$ is fixed, the number of observations $n$ goes to infinity, and the penalty function increases at the rate of $\sqrt{n}$. While the asymptotic distribution of the rescaled estimation error can be derived by relatively standard arguments, the convergence of the pattern does not simply follow from the convergence in distribution, and requires a careful and separate treatment. For this purpose, we use the Hausdorff distance as a suitable mode of convergence for subdifferentials, resulting in the desired pattern convergence. Furthermore, we derive the exact limiting probability of recovering the true model pattern. This probability goes to 1 if and only if the penalty scaling constant diverges to infinity and the regularizer-specific asymptotic irrepresentability condition is satisfied. We then propose simple two-step procedures that asymptotically recover the model patterns, irrespective whether the irrepresentability condition holds. Interestingly, our theory shows that Fused Lasso cannot reliably recover its own clustering pattern, even for independent regressors. It also demonstrates how this problem can be resolved by ``concavifying'' the Fused Lasso penalty coefficients. Additionally, sampling from the asymptotic error distribution facilitates comparisons between different regularizers. We provide short simulation studies showcasing an illustrative comparison between the asymptotic properties of Lasso, Fused Lasso, and SLOPE.
Comment: 30 pages, 5 figures, for associated code, see https://github.com/IvanHejny/pyslope.git
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/2405.07677
رقم الأكسشن: edsarx.2405.07677
قاعدة البيانات: arXiv