Computes an approx. of the Hessian w/ finite differences of the gradient. function hessfd = getHessianFD(problem, x, d) function hessfd = getHessianFD(problem, x, d, storedb) function hessfd = getHessianFD(problem, x, d, storedb, key) Returns a finite difference approximation of the Hessian at x along d of the cost function described in the problem structure. The finite difference is based on computations of the gradient. storedb is a StoreDB object, key is the StoreDB key to point x. If the gradient cannot be computed, an exception is thrown. See also: approxhessianFD

- StoreDB
- getGradient Computes the gradient of the cost function at x.

- getApproxHessian Computes an approximation of the Hessian of the cost fun. at x along d.

0001 function hessfd = getHessianFD(problem, x, d, storedb, key) 0002 % Computes an approx. of the Hessian w/ finite differences of the gradient. 0003 % 0004 % function hessfd = getHessianFD(problem, x, d) 0005 % function hessfd = getHessianFD(problem, x, d, storedb) 0006 % function hessfd = getHessianFD(problem, x, d, storedb, key) 0007 % 0008 % Returns a finite difference approximation of the Hessian at x along d of 0009 % the cost function described in the problem structure. The finite 0010 % difference is based on computations of the gradient. 0011 % 0012 % storedb is a StoreDB object, key is the StoreDB key to point x. 0013 % 0014 % If the gradient cannot be computed, an exception is thrown. 0015 % 0016 % See also: approxhessianFD 0017 0018 % This file is part of Manopt: www.manopt.org. 0019 % Original author: Nicolas Boumal, Dec. 30, 2012. 0020 % Contributors: 0021 % Change log: 0022 % 0023 % Feb. 19, 2015 (NB): 0024 % It is sufficient to ensure positive radial linearity to guarantee 0025 % (together with other assumptions) that this approximation of the 0026 % Hessian will confer global convergence to the trust-regions method. 0027 % Formerly, in-code comments referred to the necessity of having 0028 % complete radial linearity, and that this was harder to achieve. 0029 % This appears not to be necessary after all, which simplifies the 0030 % code. 0031 % 0032 % April 3, 2015 (NB): 0033 % Works with the new StoreDB class system. 0034 % 0035 % Nov. 1, 2016 (NB): 0036 % Removed exception in case of unavailable gradient, as getGradient 0037 % now knows to fall back to an approximate gradient if need be. 0038 0039 % Allow omission of the key, and even of storedb. 0040 if ~exist('key', 'var') 0041 if ~exist('storedb', 'var') 0042 storedb = StoreDB(); 0043 end 0044 key = storedb.getNewKey(); 0045 end 0046 0047 % Step size 0048 norm_d = problem.M.norm(x, d); 0049 0050 % First, check whether the step d is not too small 0051 if norm_d < eps 0052 hessfd = problem.M.zerovec(x); 0053 return; 0054 end 0055 0056 % Parameter: how far do we look? 0057 % If you need to change this parameter, use approxhessianFD explicitly. 0058 % A power of 2 is chosen so that scaling by epsilon does not incur any 0059 % round-off error in IEEE arithmetic. 0060 epsilon = 2^-14; 0061 0062 c = epsilon/norm_d; 0063 0064 % Compute the gradient at the current point. 0065 grad = getGradient(problem, x, storedb, key); 0066 0067 % Compute a point a little further along d and the gradient there. 0068 % Since this is a new point, we need a new key for it, for the storedb. 0069 x1 = problem.M.retr(x, d, c); 0070 key1 = storedb.getNewKey(); 0071 grad1 = getGradient(problem, x1, storedb, key1); 0072 0073 % Transport grad1 back from x1 to x. 0074 grad1 = problem.M.transp(x1, x, grad1); 0075 0076 % Return the finite difference of them. 0077 hessfd = problem.M.lincomb(x, 1/c, grad1, -1/c, grad); 0078 0079 end

Generated on Fri 08-Sep-2017 12:43:19 by