| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.commons.math3.optimization.general; |
| |
| import org.apache.commons.math3.analysis.DifferentiableMultivariateVectorFunction; |
| import org.apache.commons.math3.analysis.FunctionUtils; |
| import org.apache.commons.math3.analysis.differentiation.DerivativeStructure; |
| import org.apache.commons.math3.analysis.differentiation.MultivariateDifferentiableVectorFunction; |
| import org.apache.commons.math3.exception.DimensionMismatchException; |
| import org.apache.commons.math3.exception.NumberIsTooSmallException; |
| import org.apache.commons.math3.exception.util.LocalizedFormats; |
| import org.apache.commons.math3.linear.ArrayRealVector; |
| import org.apache.commons.math3.linear.RealMatrix; |
| import org.apache.commons.math3.linear.DiagonalMatrix; |
| import org.apache.commons.math3.linear.DecompositionSolver; |
| import org.apache.commons.math3.linear.MatrixUtils; |
| import org.apache.commons.math3.linear.QRDecomposition; |
| import org.apache.commons.math3.linear.EigenDecomposition; |
| import org.apache.commons.math3.optimization.OptimizationData; |
| import org.apache.commons.math3.optimization.InitialGuess; |
| import org.apache.commons.math3.optimization.Target; |
| import org.apache.commons.math3.optimization.Weight; |
| import org.apache.commons.math3.optimization.ConvergenceChecker; |
| import org.apache.commons.math3.optimization.DifferentiableMultivariateVectorOptimizer; |
| import org.apache.commons.math3.optimization.PointVectorValuePair; |
| import org.apache.commons.math3.optimization.direct.BaseAbstractMultivariateVectorOptimizer; |
| import org.apache.commons.math3.util.FastMath; |
| |
| /** |
| * Base class for implementing least squares optimizers. |
| * It handles the boilerplate methods associated to thresholds settings, |
| * Jacobian and error estimation. |
| * <br/> |
| * This class constructs the Jacobian matrix of the function argument in method |
| * {@link BaseAbstractMultivariateVectorOptimizer#optimize(int, |
| * org.apache.commons.math3.analysis.MultivariateVectorFunction,OptimizationData[]) |
| * optimize} and assumes that the rows of that matrix iterate on the model |
| * functions while the columns iterate on the parameters; thus, the numbers |
| * of rows is equal to the dimension of the |
| * {@link org.apache.commons.math3.optimization.Target Target} while |
| * the number of columns is equal to the dimension of the |
| * {@link org.apache.commons.math3.optimization.InitialGuess InitialGuess}. |
| * |
| * @deprecated As of 3.1 (to be removed in 4.0). |
| * @since 1.2 |
| */ |
| @Deprecated |
| public abstract class AbstractLeastSquaresOptimizer |
| extends BaseAbstractMultivariateVectorOptimizer<DifferentiableMultivariateVectorFunction> |
| implements DifferentiableMultivariateVectorOptimizer { |
| /** |
| * Singularity threshold (cf. {@link #getCovariances(double)}). |
| * @deprecated As of 3.1. |
| */ |
| @Deprecated |
| private static final double DEFAULT_SINGULARITY_THRESHOLD = 1e-14; |
| /** |
| * Jacobian matrix of the weighted residuals. |
| * This matrix is in canonical form just after the calls to |
| * {@link #updateJacobian()}, but may be modified by the solver |
| * in the derived class (the {@link LevenbergMarquardtOptimizer |
| * Levenberg-Marquardt optimizer} does this). |
| * @deprecated As of 3.1. To be removed in 4.0. Please use |
| * {@link #computeWeightedJacobian(double[])} instead. |
| */ |
| @Deprecated |
| protected double[][] weightedResidualJacobian; |
| /** Number of columns of the jacobian matrix. |
| * @deprecated As of 3.1. |
| */ |
| @Deprecated |
| protected int cols; |
| /** Number of rows of the jacobian matrix. |
| * @deprecated As of 3.1. |
| */ |
| @Deprecated |
| protected int rows; |
| /** Current point. |
| * @deprecated As of 3.1. |
| */ |
| @Deprecated |
| protected double[] point; |
| /** Current objective function value. |
| * @deprecated As of 3.1. |
| */ |
| @Deprecated |
| protected double[] objective; |
| /** Weighted residuals |
| * @deprecated As of 3.1. |
| */ |
| @Deprecated |
| protected double[] weightedResiduals; |
| /** Cost value (square root of the sum of the residuals). |
| * @deprecated As of 3.1. Field to become "private" in 4.0. |
| * Please use {@link #setCost(double)}. |
| */ |
| @Deprecated |
| protected double cost; |
| /** Objective function derivatives. */ |
| private MultivariateDifferentiableVectorFunction jF; |
| /** Number of evaluations of the Jacobian. */ |
| private int jacobianEvaluations; |
| /** Square-root of the weight matrix. */ |
| private RealMatrix weightMatrixSqrt; |
| |
| /** |
| * Simple constructor with default settings. |
| * The convergence check is set to a {@link |
| * org.apache.commons.math3.optimization.SimpleVectorValueChecker}. |
| * @deprecated See {@link org.apache.commons.math3.optimization.SimpleValueChecker#SimpleValueChecker()} |
| */ |
| @Deprecated |
| protected AbstractLeastSquaresOptimizer() {} |
| |
| /** |
| * @param checker Convergence checker. |
| */ |
| protected AbstractLeastSquaresOptimizer(ConvergenceChecker<PointVectorValuePair> checker) { |
| super(checker); |
| } |
| |
| /** |
| * @return the number of evaluations of the Jacobian function. |
| */ |
| public int getJacobianEvaluations() { |
| return jacobianEvaluations; |
| } |
| |
| /** |
| * Update the jacobian matrix. |
| * |
| * @throws DimensionMismatchException if the Jacobian dimension does not |
| * match problem dimension. |
| * @deprecated As of 3.1. Please use {@link #computeWeightedJacobian(double[])} |
| * instead. |
| */ |
| @Deprecated |
| protected void updateJacobian() { |
| final RealMatrix weightedJacobian = computeWeightedJacobian(point); |
| weightedResidualJacobian = weightedJacobian.scalarMultiply(-1).getData(); |
| } |
| |
| /** |
| * Computes the Jacobian matrix. |
| * |
| * @param params Model parameters at which to compute the Jacobian. |
| * @return the weighted Jacobian: W<sup>1/2</sup> J. |
| * @throws DimensionMismatchException if the Jacobian dimension does not |
| * match problem dimension. |
| * @since 3.1 |
| */ |
| protected RealMatrix computeWeightedJacobian(double[] params) { |
| ++jacobianEvaluations; |
| |
| final DerivativeStructure[] dsPoint = new DerivativeStructure[params.length]; |
| final int nC = params.length; |
| for (int i = 0; i < nC; ++i) { |
| dsPoint[i] = new DerivativeStructure(nC, 1, i, params[i]); |
| } |
| final DerivativeStructure[] dsValue = jF.value(dsPoint); |
| final int nR = getTarget().length; |
| if (dsValue.length != nR) { |
| throw new DimensionMismatchException(dsValue.length, nR); |
| } |
| final double[][] jacobianData = new double[nR][nC]; |
| for (int i = 0; i < nR; ++i) { |
| int[] orders = new int[nC]; |
| for (int j = 0; j < nC; ++j) { |
| orders[j] = 1; |
| jacobianData[i][j] = dsValue[i].getPartialDerivative(orders); |
| orders[j] = 0; |
| } |
| } |
| |
| return weightMatrixSqrt.multiply(MatrixUtils.createRealMatrix(jacobianData)); |
| } |
| |
| /** |
| * Update the residuals array and cost function value. |
| * @throws DimensionMismatchException if the dimension does not match the |
| * problem dimension. |
| * @throws org.apache.commons.math3.exception.TooManyEvaluationsException |
| * if the maximal number of evaluations is exceeded. |
| * @deprecated As of 3.1. Please use {@link #computeResiduals(double[])}, |
| * {@link #computeObjectiveValue(double[])}, {@link #computeCost(double[])} |
| * and {@link #setCost(double)} instead. |
| */ |
| @Deprecated |
| protected void updateResidualsAndCost() { |
| objective = computeObjectiveValue(point); |
| final double[] res = computeResiduals(objective); |
| |
| // Compute cost. |
| cost = computeCost(res); |
| |
| // Compute weighted residuals. |
| final ArrayRealVector residuals = new ArrayRealVector(res); |
| weightedResiduals = weightMatrixSqrt.operate(residuals).toArray(); |
| } |
| |
| /** |
| * Computes the cost. |
| * |
| * @param residuals Residuals. |
| * @return the cost. |
| * @see #computeResiduals(double[]) |
| * @since 3.1 |
| */ |
| protected double computeCost(double[] residuals) { |
| final ArrayRealVector r = new ArrayRealVector(residuals); |
| return FastMath.sqrt(r.dotProduct(getWeight().operate(r))); |
| } |
| |
| /** |
| * Get the Root Mean Square value. |
| * Get the Root Mean Square value, i.e. the root of the arithmetic |
| * mean of the square of all weighted residuals. This is related to the |
| * criterion that is minimized by the optimizer as follows: if |
| * <em>c</em> if the criterion, and <em>n</em> is the number of |
| * measurements, then the RMS is <em>sqrt (c/n)</em>. |
| * |
| * @return RMS value |
| */ |
| public double getRMS() { |
| return FastMath.sqrt(getChiSquare() / rows); |
| } |
| |
| /** |
| * Get a Chi-Square-like value assuming the N residuals follow N |
| * distinct normal distributions centered on 0 and whose variances are |
| * the reciprocal of the weights. |
| * @return chi-square value |
| */ |
| public double getChiSquare() { |
| return cost * cost; |
| } |
| |
| /** |
| * Gets the square-root of the weight matrix. |
| * |
| * @return the square-root of the weight matrix. |
| * @since 3.1 |
| */ |
| public RealMatrix getWeightSquareRoot() { |
| return weightMatrixSqrt.copy(); |
| } |
| |
| /** |
| * Sets the cost. |
| * |
| * @param cost Cost value. |
| * @since 3.1 |
| */ |
| protected void setCost(double cost) { |
| this.cost = cost; |
| } |
| |
| /** |
| * Get the covariance matrix of the optimized parameters. |
| * |
| * @return the covariance matrix. |
| * @throws org.apache.commons.math3.linear.SingularMatrixException |
| * if the covariance matrix cannot be computed (singular problem). |
| * @see #getCovariances(double) |
| * @deprecated As of 3.1. Please use {@link #computeCovariances(double[],double)} |
| * instead. |
| */ |
| @Deprecated |
| public double[][] getCovariances() { |
| return getCovariances(DEFAULT_SINGULARITY_THRESHOLD); |
| } |
| |
| /** |
| * Get the covariance matrix of the optimized parameters. |
| * <br/> |
| * Note that this operation involves the inversion of the |
| * <code>J<sup>T</sup>J</code> matrix, where {@code J} is the |
| * Jacobian matrix. |
| * The {@code threshold} parameter is a way for the caller to specify |
| * that the result of this computation should be considered meaningless, |
| * and thus trigger an exception. |
| * |
| * @param threshold Singularity threshold. |
| * @return the covariance matrix. |
| * @throws org.apache.commons.math3.linear.SingularMatrixException |
| * if the covariance matrix cannot be computed (singular problem). |
| * @deprecated As of 3.1. Please use {@link #computeCovariances(double[],double)} |
| * instead. |
| */ |
| @Deprecated |
| public double[][] getCovariances(double threshold) { |
| return computeCovariances(point, threshold); |
| } |
| |
| /** |
| * Get the covariance matrix of the optimized parameters. |
| * <br/> |
| * Note that this operation involves the inversion of the |
| * <code>J<sup>T</sup>J</code> matrix, where {@code J} is the |
| * Jacobian matrix. |
| * The {@code threshold} parameter is a way for the caller to specify |
| * that the result of this computation should be considered meaningless, |
| * and thus trigger an exception. |
| * |
| * @param params Model parameters. |
| * @param threshold Singularity threshold. |
| * @return the covariance matrix. |
| * @throws org.apache.commons.math3.linear.SingularMatrixException |
| * if the covariance matrix cannot be computed (singular problem). |
| * @since 3.1 |
| */ |
| public double[][] computeCovariances(double[] params, |
| double threshold) { |
| // Set up the Jacobian. |
| final RealMatrix j = computeWeightedJacobian(params); |
| |
| // Compute transpose(J)J. |
| final RealMatrix jTj = j.transpose().multiply(j); |
| |
| // Compute the covariances matrix. |
| final DecompositionSolver solver |
| = new QRDecomposition(jTj, threshold).getSolver(); |
| return solver.getInverse().getData(); |
| } |
| |
| /** |
| * <p> |
| * Returns an estimate of the standard deviation of each parameter. The |
| * returned values are the so-called (asymptotic) standard errors on the |
| * parameters, defined as {@code sd(a[i]) = sqrt(S / (n - m) * C[i][i])}, |
| * where {@code a[i]} is the optimized value of the {@code i}-th parameter, |
| * {@code S} is the minimized value of the sum of squares objective function |
| * (as returned by {@link #getChiSquare()}), {@code n} is the number of |
| * observations, {@code m} is the number of parameters and {@code C} is the |
| * covariance matrix. |
| * </p> |
| * <p> |
| * See also |
| * <a href="http://en.wikipedia.org/wiki/Least_squares">Wikipedia</a>, |
| * or |
| * <a href="http://mathworld.wolfram.com/LeastSquaresFitting.html">MathWorld</a>, |
| * equations (34) and (35) for a particular case. |
| * </p> |
| * |
| * @return an estimate of the standard deviation of the optimized parameters |
| * @throws org.apache.commons.math3.linear.SingularMatrixException |
| * if the covariance matrix cannot be computed. |
| * @throws NumberIsTooSmallException if the number of degrees of freedom is not |
| * positive, i.e. the number of measurements is less or equal to the number of |
| * parameters. |
| * @deprecated as of version 3.1, {@link #computeSigma(double[],double)} should be used |
| * instead. It should be emphasized that {@code guessParametersErrors} and |
| * {@code computeSigma} are <em>not</em> strictly equivalent. |
| */ |
| @Deprecated |
| public double[] guessParametersErrors() { |
| if (rows <= cols) { |
| throw new NumberIsTooSmallException(LocalizedFormats.NO_DEGREES_OF_FREEDOM, |
| rows, cols, false); |
| } |
| double[] errors = new double[cols]; |
| final double c = FastMath.sqrt(getChiSquare() / (rows - cols)); |
| double[][] covar = computeCovariances(point, 1e-14); |
| for (int i = 0; i < errors.length; ++i) { |
| errors[i] = FastMath.sqrt(covar[i][i]) * c; |
| } |
| return errors; |
| } |
| |
| /** |
| * Computes an estimate of the standard deviation of the parameters. The |
| * returned values are the square root of the diagonal coefficients of the |
| * covariance matrix, {@code sd(a[i]) ~= sqrt(C[i][i])}, where {@code a[i]} |
| * is the optimized value of the {@code i}-th parameter, and {@code C} is |
| * the covariance matrix. |
| * |
| * @param params Model parameters. |
| * @param covarianceSingularityThreshold Singularity threshold (see |
| * {@link #computeCovariances(double[],double) computeCovariances}). |
| * @return an estimate of the standard deviation of the optimized parameters |
| * @throws org.apache.commons.math3.linear.SingularMatrixException |
| * if the covariance matrix cannot be computed. |
| * @since 3.1 |
| */ |
| public double[] computeSigma(double[] params, |
| double covarianceSingularityThreshold) { |
| final int nC = params.length; |
| final double[] sig = new double[nC]; |
| final double[][] cov = computeCovariances(params, covarianceSingularityThreshold); |
| for (int i = 0; i < nC; ++i) { |
| sig[i] = FastMath.sqrt(cov[i][i]); |
| } |
| return sig; |
| } |
| |
| /** {@inheritDoc} |
| * @deprecated As of 3.1. Please use |
| * {@link BaseAbstractMultivariateVectorOptimizer#optimize(int, |
| * org.apache.commons.math3.analysis.MultivariateVectorFunction,OptimizationData[]) |
| * optimize(int,MultivariateDifferentiableVectorFunction,OptimizationData...)} |
| * instead. |
| */ |
| @Override |
| @Deprecated |
| public PointVectorValuePair optimize(int maxEval, |
| final DifferentiableMultivariateVectorFunction f, |
| final double[] target, final double[] weights, |
| final double[] startPoint) { |
| return optimizeInternal(maxEval, |
| FunctionUtils.toMultivariateDifferentiableVectorFunction(f), |
| new Target(target), |
| new Weight(weights), |
| new InitialGuess(startPoint)); |
| } |
| |
| /** |
| * Optimize an objective function. |
| * Optimization is considered to be a weighted least-squares minimization. |
| * The cost function to be minimized is |
| * <code>∑weight<sub>i</sub>(objective<sub>i</sub> - target<sub>i</sub>)<sup>2</sup></code> |
| * |
| * @param f Objective function. |
| * @param target Target value for the objective functions at optimum. |
| * @param weights Weights for the least squares cost computation. |
| * @param startPoint Start point for optimization. |
| * @return the point/value pair giving the optimal value for objective |
| * function. |
| * @param maxEval Maximum number of function evaluations. |
| * @throws org.apache.commons.math3.exception.DimensionMismatchException |
| * if the start point dimension is wrong. |
| * @throws org.apache.commons.math3.exception.TooManyEvaluationsException |
| * if the maximal number of evaluations is exceeded. |
| * @throws org.apache.commons.math3.exception.NullArgumentException if |
| * any argument is {@code null}. |
| * @deprecated As of 3.1. Please use |
| * {@link BaseAbstractMultivariateVectorOptimizer#optimize(int, |
| * org.apache.commons.math3.analysis.MultivariateVectorFunction,OptimizationData[]) |
| * optimize(int,MultivariateDifferentiableVectorFunction,OptimizationData...)} |
| * instead. |
| */ |
| @Deprecated |
| public PointVectorValuePair optimize(final int maxEval, |
| final MultivariateDifferentiableVectorFunction f, |
| final double[] target, final double[] weights, |
| final double[] startPoint) { |
| return optimizeInternal(maxEval, f, |
| new Target(target), |
| new Weight(weights), |
| new InitialGuess(startPoint)); |
| } |
| |
| /** |
| * Optimize an objective function. |
| * Optimization is considered to be a weighted least-squares minimization. |
| * The cost function to be minimized is |
| * <code>∑weight<sub>i</sub>(objective<sub>i</sub> - target<sub>i</sub>)<sup>2</sup></code> |
| * |
| * @param maxEval Allowed number of evaluations of the objective function. |
| * @param f Objective function. |
| * @param optData Optimization data. The following data will be looked for: |
| * <ul> |
| * <li>{@link Target}</li> |
| * <li>{@link Weight}</li> |
| * <li>{@link InitialGuess}</li> |
| * </ul> |
| * @return the point/value pair giving the optimal value of the objective |
| * function. |
| * @throws org.apache.commons.math3.exception.TooManyEvaluationsException if |
| * the maximal number of evaluations is exceeded. |
| * @throws DimensionMismatchException if the target, and weight arguments |
| * have inconsistent dimensions. |
| * @see BaseAbstractMultivariateVectorOptimizer#optimizeInternal(int, |
| * org.apache.commons.math3.analysis.MultivariateVectorFunction,OptimizationData[]) |
| * @since 3.1 |
| * @deprecated As of 3.1. Override is necessary only until this class's generic |
| * argument is changed to {@code MultivariateDifferentiableVectorFunction}. |
| */ |
| @Deprecated |
| protected PointVectorValuePair optimizeInternal(final int maxEval, |
| final MultivariateDifferentiableVectorFunction f, |
| OptimizationData... optData) { |
| // XXX Conversion will be removed when the generic argument of the |
| // base class becomes "MultivariateDifferentiableVectorFunction". |
| return super.optimizeInternal(maxEval, FunctionUtils.toDifferentiableMultivariateVectorFunction(f), optData); |
| } |
| |
| /** {@inheritDoc} */ |
| @Override |
| protected void setUp() { |
| super.setUp(); |
| |
| // Reset counter. |
| jacobianEvaluations = 0; |
| |
| // Square-root of the weight matrix. |
| weightMatrixSqrt = squareRoot(getWeight()); |
| |
| // Store least squares problem characteristics. |
| // XXX The conversion won't be necessary when the generic argument of |
| // the base class becomes "MultivariateDifferentiableVectorFunction". |
| // XXX "jF" is not strictly necessary anymore but is currently more |
| // efficient than converting the value returned from "getObjectiveFunction()" |
| // every time it is used. |
| jF = FunctionUtils.toMultivariateDifferentiableVectorFunction((DifferentiableMultivariateVectorFunction) getObjectiveFunction()); |
| |
| // Arrays shared with "private" and "protected" methods. |
| point = getStartPoint(); |
| rows = getTarget().length; |
| cols = point.length; |
| } |
| |
| /** |
| * Computes the residuals. |
| * The residual is the difference between the observed (target) |
| * values and the model (objective function) value. |
| * There is one residual for each element of the vector-valued |
| * function. |
| * |
| * @param objectiveValue Value of the the objective function. This is |
| * the value returned from a call to |
| * {@link #computeObjectiveValue(double[]) computeObjectiveValue} |
| * (whose array argument contains the model parameters). |
| * @return the residuals. |
| * @throws DimensionMismatchException if {@code params} has a wrong |
| * length. |
| * @since 3.1 |
| */ |
| protected double[] computeResiduals(double[] objectiveValue) { |
| final double[] target = getTarget(); |
| if (objectiveValue.length != target.length) { |
| throw new DimensionMismatchException(target.length, |
| objectiveValue.length); |
| } |
| |
| final double[] residuals = new double[target.length]; |
| for (int i = 0; i < target.length; i++) { |
| residuals[i] = target[i] - objectiveValue[i]; |
| } |
| |
| return residuals; |
| } |
| |
| /** |
| * Computes the square-root of the weight matrix. |
| * |
| * @param m Symmetric, positive-definite (weight) matrix. |
| * @return the square-root of the weight matrix. |
| */ |
| private RealMatrix squareRoot(RealMatrix m) { |
| if (m instanceof DiagonalMatrix) { |
| final int dim = m.getRowDimension(); |
| final RealMatrix sqrtM = new DiagonalMatrix(dim); |
| for (int i = 0; i < dim; i++) { |
| sqrtM.setEntry(i, i, FastMath.sqrt(m.getEntry(i, i))); |
| } |
| return sqrtM; |
| } else { |
| final EigenDecomposition dec = new EigenDecomposition(m); |
| return dec.getSquareRoot(); |
| } |
| } |
| } |