Poblano is a Matlab toolbox of large-scale algorithms for unconstrained nonlinear optimization problems. The algorithms in Poblano require only first-order derivative information (e.g., gradients for scalar-valued objective functions), and therefore can scale to very large problems. The driving application for Poblano development has been tensor decompositions in data analysis applications (bibliometric analysis, social network analysis, chemometrics, etc.). Poblano optimizers find local minimizers of scalar-valued objective functions taking vector inputs. The gradient (i.e., first derivative) of the objective function is required for all Poblano optimizers. The optimizers converge to a stationary point where the gradient is approximately zero. A line search satisfying the strong Wolfe conditions is used to guarantee global convergence of the Poblano optimizers. The optimization methods in Poblano include several nonlinear conjugate gradient methods (Fletcher-Reeves, Polak-Ribiere, Hestenes-Stiefel), a limited-memory quasi-Newton method using BFGS updates to approximate second-order derivative information, and a truncated Newton method using finite differences to approximate second-order derivative information.
Keywords for this software
References in zbMATH (referenced in 4 articles )
Showing results 1 to 4 of 4.
- De Sterck, Hans; Howse, Alexander: Nonlinearly preconditioned optimization on Grassmann manifolds for computing approximate Tucker tensor decompositions (2016)
- Carlberg, Kevin; Tuminaro, Ray; Boggs, Paul: Preserving Lagrangian structure in nonlinear model reduction with application to structural dynamics (2015)
- De Sterck, Hans; Winlaw, Manda: A nonlinearly preconditioned conjugate gradient algorithm for rank-$R$ canonical tensor approximation. (2015)
- De Sterck, H.: A nonlinear GMRES optimization algorithm for canonical tensor decomposition (2012)