You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Optional weights to put on each term in the penalty.
22
+
23
+
groups: None, list of ints
24
+
Optional groups of variables. If groups is provided then each element in the list should be a list of feature indices. Variables not in a group are not penalized.
Tunes the lasso penalty parameter via cross-validation.
143
+
""")
144
+
106
145
def_check_base_estimator(self, estimator):
107
146
check_estimator_type(estimator, GlmLasso)
108
147
109
148
149
+
_glm_lasso_params=dedent("""
150
+
pen_val: float
151
+
The penalty strength (corresponds to lambda in glmnet)
152
+
153
+
l1_ratio: float
154
+
The ElasticNet mixing parameter, with ``0 <= l1_ratio <= 1``. For
155
+
``l1_ratio = 0`` the penalty is an L2 penalty. ``For l1_ratio = 1`` it
156
+
is an L1 penalty. For ``0 < l1_ratio < 1``, the penalty is a
157
+
combination of L1 and L2.
158
+
159
+
lasso_weights: None, array-like
160
+
Optional weights to put on each term in the penalty.
161
+
162
+
groups: None, list of ints
163
+
Optional groups of variables. If groups is provided then each element in the list should be a list of feature indices. Variables not in a group are not penalized.
@@ -192,6 +268,10 @@ def _get_pen_val_max_from_pro(self, X, y, sample_weight=None):
192
268
classGlmENetCVPath(ENetCVPathMixin, GlmCVENet):
193
269
solve_glm_path=None
194
270
271
+
desrc=dedent("""
272
+
Tunes the ElasticNet penalty parameter and or the l1_ratio via cross-validation. Makes use of a path algorithm for computing the penalty value tuning path.
0 commit comments