English | Español

Try our Free Online Math Solver!

Online Math Solver

 

 

 

 

 

 

 

 
 
 
 
 
 
 
 
 

 

 

 
 
 
 
 
 
 
 
 

Please use this form if you would like
to have this math solver on your website,
free of charge.


Nonlinear Programming Final Exam

section A

Indicate whether each statement is true (+) or false (o). If false, briefly state why or give a
counterexample.

_____ 1. A barrier function adds a penalty to the objective when the solution approaches a
boundary of the feasible region.

_____ 2. The number of "dependent" variables of the GRG algorithm is equal to the number of
equality constraints.

_____ 3. The cubic interpolation procedure for one-dimensional minimization requires the
computation of second derivatives.

_____ 4. The function xa is convex for x>0.

_____ 5. The feasible directions algorithm computes a search direction by solving an LP .

_____ 6. The Fibonacci method for one -dimensional minimization requires the computation of
derivatives.

_____ 7. If all posynomials in a (posynomial) GP problem are condensed into single terms,
then it is always possible to rewrite the resulting problem as an LP problem.

_____ 8. A linear function is neither convex nor concave.

_____ 9. The Gradient Projection method always fol lows the boundary of the feasible region,
and never enters the interior.

_____ 10. The logarithm of the dual GP objective v(δ,λ) above is a concave function of δ and λ.

_____ 11. The logarithm of the dual GP objective v(δ,λ) above is not a concave function of
both δ and λ, but becomes so if you eliminate using

_____ 12. A locally optimal solution for a signomial geometric programming problem must be
globally optimal.

_____ 13. The dual ("pseudo-dual") of a signomial GP problem may have multiple local
maxima, but its global maximum will have the same objective value as the global
minimum of the primal signomial GP.

_____ 14. Penalty functions may be used for both equality and inequality constraints , but barrier
functions may be used only for inequality constraints.

_____ 16. If the Lagrangian function L(x,λ) has a saddlepoint , then there is zero duality
gap between primal and dual optimal solutions.

_____ 17. Consider the problem: min f(x) subject to gi(x) ≤ 0, i=1,2...m, x ≥ 0 (where gi satisfy
some constraint qualification) and its Lagrangian function

Then if f and gi are convex, the partial derivative of L with respect to each xj must be
zero at the optimum.

_____ 18. For the problem in the previous statement, must be zero at the optimum.

_____ 19. For the problem in the previous statement, either xj or (or possibly both)
must be zero at the optimum.

_____20. The limit, as w → 0, of the function (c/w)w is ∞.

_____21.

_____22. If we define the function  then

_____23. When the dual problem is solved, the primal variables of a posynomial GP problem
are often computed by ration al-exponents-and-radical.html">exponentiating the Lagrange multipliers of the orthogonality
constraints.

_____24. The function f(x,y) = xy is a concave function of x and y.

_____25. If an algorithm is applied to a minimization problem with optimal value zero, and the
objective value is approximately halved at each iteration, then we would say that the
algorithm has a linear rate of convergence.

_____26. The Hessian matrix of (xtQx + cx) is 2Q.

_____27. In the QC/LD problem, the objective function is as sumed to be quadratic .

_____28. If fi(x) is linear for i=1,2,...n, then the function F(x) defined by F(x)= max{fi(x):
i=1,2,...n} is a convex function.

_____29. If a posynomial GP objective function continues to decrease as xj → 0 for some primal
variable xj, then the dual problem objective is unbounded.

_____30. If a posynomial GP objective function continues to decrease as xj → +∞ for some
primal variable xj, then the dual problem is infeasible.

_____ 31. Every posynomial function is convex.

_____ 32. The dual of a constrained posynomial GP problem has only linear equality and
nonnegativity constraints.

_____ 33. If a primal GP constraint is slack, then all the weights of the terms in that constraint
must be zero.

_____ 34. In the QC/LD problem, the variables are restricted to be integer-valued.

_____ 35. The gradient of the function f(x) = (xtQx + cx) is Qx + c.

_____ 36. Solving the QC/LD problem requires a one-dimensional search at each iteration.

_____ 37. The posynomial constraint gi(x) ≤1 has a convex feasible region.

_____ 38. It is always possible (e.g., by a change of variable ) to re formulate a posynomial GP
problem so as to have a convex objective and convex feasible region.

_____ 39. The objective of the posynomial GP problem, i.e.,

is a concave function of δ and λ.

_____ 40. In the "Golden Section Search" method, more than one-third of the interval of
uncertainty is eliminated at each iteration (assuming that no function values are equal).

_____ 41. If all posynomials in a (posynomial) GP problem are condensed into single terms,
then the resulting problem has a feasible region which includes the original feasible
region.

_____ 42. The Hessian matrix of a quadratic function is constant.

_____ 43. In Wolfe\'s complementary pivoting algorithm for QP, if a single artificial variable is
used, then the tableau does not require an objective row.

_____ 44. The gradient projection method computes Lagrange multipliers at each iteration, and
stops when they have the appropriate sign.

_____ 45. "Quasi-Newton" search methods for unconstrained minimization require the
computation of second partial derivatives.

_____ 46. Newton\'s method for unconstrained minimization requires the computation of second
partial derivatives.

_____ 47. If you guess at the value of some primal GP variable and then fix it at this value, the
dual GP problem becomes more difficult to solve.

_____ 48. In a "generalized linear programming" problem, the column of coefficients of the
variables must be selected as well as the variables.

_____ 49. The function eax is convex only for a ≥ 0.

_____ 50. The Davidon-Fletcher-Powell (DFP) algorithm requires the computation of partial
derivatives.

_____ 51 . If a posynomial function is "condensed" into a monomial, the monomial function (if
not equal) is an overestimate of the posynomial function.

_____ 52 . Wolfe\'s method for quadratic programming requires a one-dimensional search at
every iteration.

_____ 53. If a constrained nonlinear programming problem satisfies a "constraint qualification",
then a point which satisfies the Karush-Kuhn-Tucker conditions must be an optimal
solution.

_____ 54. A barrier function allows a constraint to be violated, but adds a penalty if the
constraint is violated.

_____ 55. The GRG algorithm requires the use of a one-dimensional search method.

_____ 56. The Feasible Directions algorithm requires the use of a one-dimensional search
method.

_____ 57. If a constrained nonlinear programming problem satisfies a "constraint qualification",
then the Karush-Kuhn-Tucker conditions must be satisfied by an optimal solution.

_____ 58. The function eax is convex for all values of a.

_____ 59. In Wolfe\'s complementary pivoting method for quadratic programming, the
complementary slackness conditions are satisfied after each iteration.

_____ 60 . The tableau for Wolfe\'s method for quadratic programming includes columns for
both primal and dual variables.

_____ 61. The function xa is convex for a ≥ 0.

_____ 62. If a nonlinear programming problem has only linear constraints, then any point which
satisfies the Karush-Kuhn-Tucker conditions must be optimal.

_____ 63. The Fletcher-Reeves method is also known as a "Quasi-Newton" method.

_____ 64. The quadratic interpolation procedure for one-dimensional minimization requires the
computation of derivatives.

_____ 65. If the GRG algorithm were applied to a LP problem, it would produce, at each
iteration, the same solution as the simplex algorithm for LP.

_____ 66. A penalty function allows a constraint to be violated, but adds a penalty if the
constraint is violated.

_____ 67. The Lagrangian dual of a convex quadratic programming problem is a quadratic
programming problem with only nonnegativity constraints on the dual variables.

_____ 68. Powell\'s algorithm for unconstrained minimization requires the computation of partial
derivatives.

_____ 69. The Davidon-Fletcher-Powell method is also known as the "Conjugate Gradient"
method.

_____ 70. In the QC/LD problem, the objective function is assumed to be convex.

Prev Next