tick.solver.SCPG

class tick.solver.SCPG(step: float = None, tol: float = 0.0, max_iter: int = 100, verbose: bool = True, print_every: int = 10, record_every: int = 1, linesearch_step_increase: float = 2.0, linesearch_step_decrease: float = 0.5, modified=False)[source]

Self-Concordant Proximal Gradient descent

For the minimization of objectives of the form

\[f(w) + g(w),\]

where \(f\) is self-concordant and \(g\) is prox-capable. Function \(f\) corresponds to the model.loss method of the model (passed with set_model to the solver) and \(g\) corresponds to the prox.value method of the prox (passed with the set_prox method). One iteration of SCPG is as follows:

\[\begin{split}y^k &\gets \mathrm{prox}_{g / L_k} \big( w^k - \tfrac{1}{L_k} \nabla f(w^k) \big) \\ d^k &\gets y^k - w^k \\ \beta_k &\gets \sqrt{L_k} \| d^k \|_2 \\ \lambda_k &\gets \sqrt{{d^k}^\top \nabla^2 f(w^k) d^k} \\ \alpha_k &\gets \frac{\beta_k}{\lambda_k (\lambda_k + \beta_k^2)} \\ w^{k + 1} &\gets w^{k} + \alpha_k d^k \\\end{split}\]

where \(\nabla f(w)\) is the gradient of \(f\) given by the model.grad method and \(\mathrm{prox}_{g / L_k}\) is given by the prox.call method. The first step size \(1 / L_k\) can be tuned with step it will then be updated with Barzilai-Borwein steps and linesearch. The iterations stop whenever tolerance tol is achieved, or after max_iter iterations. The obtained solution \(w\) is returned by the solve method, and is also stored in the solution attribute of the solver.

Parameters

step : float, default=None

Step-size parameter that will be used at the first iteration. Then the linesearch algorithm will compute the next steps.

tol : float, default=1e-10

The tolerance of the solver (iterations stop when the stopping criterion is below it)

max_iter : int, default=100

Maximum number of iterations of the solver.

verbose : bool, default=True

If True, solver verboses history, otherwise nothing is displayed, but history is recorded anyway

print_every : int, default=10

Print history information every time the iteration number is a multiple of print_every. Used only is verbose is True

record_every : int, default=1

Save history information every time the iteration number is a multiple of record_every

linesearch_step_increase : float, default=2.

Factor of step increase when using linesearch

linesearch_step_decrease : float, default=0.5

Factor of step decrease when using linesearch

modified : bool, default=False

Enables modified verison. The modified version of this algorithm consists in evaluating \(f\) at \(y_k\) and \(w^{k+1}\) and keeping the iterate that minimizes the best \(f\).

dtype : {'float64', 'float32'}, default=’float64’

Type of the arrays used. This value is set from model and prox dtypes.

Attributes

model : Model

The model used by the solver, passed with the set_model method

prox : Prox

Proximal operator used by the solver, passed with the set_prox method

solution : numpy.array, shape=(n_coeffs,)

Minimizer found by the solver

history : dict-like

A dict-type of object that contains history of the solver along iterations. It should be accessed using the get_history method

time_start : str

Start date of the call to solve()

time_elapsed : float

Duration of the call to solve(), in seconds

time_end : str

End date of the call to solve()

Notes

This algorithm is designed to work properly with self-concordant losses such as Poisson regression with linear link ModelPoisReg or Hawkes processes likelihood ModelHawkesSumExpKernLogLik

References

Tran-Dinh Q, Kyrillidis A, Cevher V. Composite self-concordant minimization. The Journal of Machine Learning Research (2015)

__init__(step: float = None, tol: float = 0.0, max_iter: int = 100, verbose: bool = True, print_every: int = 10, record_every: int = 1, linesearch_step_increase: float = 2.0, linesearch_step_decrease: float = 0.5, modified=False)[source]

Initialize self. See help(type(self)) for accurate signature.

get_history(key=None)

Returns history of the solver

Parameters

key : str, default=None

  • If None all history is returned as a dict

  • If str, name of the history element to retrieve

Returns

output : list or dict

  • If key is None or key is not in history then output is a dict containing history of all keys

  • If key is the name of an element in the history, output is a list containing the history of this element

objective(coeffs, loss: float = None)

Compute the objective function

Parameters

coeffs : np.array, shape=(n_coeffs,)

Point where the objective is computed

loss : float, default=`None`

Gives the value of the loss if already known (allows to avoid its computation in some cases)

Returns

output : float

Value of the objective at given coeffs

set_model(model: tick.base_model.model_second_order.ModelSecondOrder)[source]

Set model in the solver

Parameters

model : ModelSecondOrder and ModelSelfConcordant

Sets the model in the solver. The model gives information about the model (loss, gradient, among other things)

Returns

output : Solver

The same instance with given model

set_prox(prox: tick.prox.base.prox.Prox)[source]

Set proximal operator in the solver.

Parameters

prox : Prox

The proximal operator of the penalization function

Returns

output : Solver

The solver with given prox

solve(x0=None, step=None)

Launch the solver

Parameters

x0 : np.array, shape=(n_coeffs,), default=`None`

Starting point of the solver

step : float, default=`None`

Step-size or learning rate for the solver. This can be tuned also using the step attribute

Returns

output : np.array, shape=(n_coeffs,)

Obtained minimizer for the problem, same as solution attribute