Optimization Methods- Lecture1

ref: Optimization Methods

Outline:

  • Mathematical Optimization
  • Least-squares
  • Linear Programming
  • Convex Optimization
  • Nonlinear Optimization
  • Summary

Mathematical Optimization

  • Optimization Problem minf0(x)s.t.fi(x)bi,i=1,2,,n

    • Optimization Variable: x=(x1,,xn)
    • Objective Function: f0:Rn\rarrR
    • Constraint Functions: fi:Rn\rarrR
  • x​ is called optimal or a solution

    • fi(x)bi , i=1,,m
    • For any z with fi(z)bi, we have f0(z)f0(x)

  • Linear Program 线性函数的定义 fi(αx+βy)=αfi(x)+βfi(y)

    • For all x,yRn and all α,βR
    • i​ 可以为0​(目标函数), 也可以为1,2,,n​​​​( 约束函数 )。 当优化问题的目标函数和约束函数都是线性函数时, 整个问题称为线性规划(线性问题)
  • Nonlinear Program

    • If the Optimization Problem is not linear
  • Convex Optimization Problem fi(αx+βy)αfi(x)+βfi(y)

    • For all x,yRn​ and all α,βR​ with α+β=1,α0,β0

    • 线性规划是凸优化的一个特例

      • 只要求对α+β=1,α0,β0α,β成立, 实际上把条件放松了​

Applications

minf0(x)s.t.fi(x)bi,i=1,2,,n

  • Abstraction
    • x​ represents the choice made
    • fi(x)bi represent firm requirements that limit the possible choices
    • f0(x)​ represents the cost of choosing
    • A solution corresponds to a choice that has minimum cost, among all choices that meet the requirements

Portfolio Optimization

  • Variables
    • xi represents the investment in the 𝑖th asset
    • xR​ describes the overall portfolio allocation across the set of asset
  • Constraints
    • A limit on the budget the requirement
    • Investments are nonnegative
    • A minimum acceptable value of expected return for the whole portfolio
  • Objective
    • Minimize the variance of the portfolio return 投资的回报最稳定( 也可以把目标换成期望的回报最高 )

Device Sizing

  • Variables
    • xRn​ describes the widths and lengths of the devices
  • Constraints
    • Limits on the device
    • Timing
    • A limit on the total area of the circuit
  • Objective
    • Minimize the total power consumed by the circuit

Data Fitting

  • Variables
    • xRn describes parameters in the model
  • Constraints
    • Prior information
    • Required limits on the parameters (such as nonnegativity)
  • Objective
    • Minimize the prediction error between the observed data and the values predicted by the model

Solving Optimization Problems

  • General Optimization Problem
    • Very difficult to solve
    • Constraints can be very complicated, the number of variables can be very large 约束复杂, 变量多
    • Methods involve some compromise, e.g., computation time, or suboptimal solution
  • Exceptions
    • Least-squares problems
    • Linear programming problems
    • Convex optimization problems

Least-squares

  • The Problem min||Axb||22=i=1k(aiTxbi)2

    • ARk×n​ , aiT​ is the ith​ row of A,bRk
    • ARn​ is the optimization variable​
  • Setting the gradient to be 0 2AT(Axb)=0\rarrATAx=ATb\rarrx=(ATA)1ATb


  • A Set of Linear Equations ATAx=ATb

  • Solving least-squares problems

    • Reliable and efficient algorithms and software
    • Computation time proportional to n2k(ARk×n) ; less if structured
    • A mature technology
    • Challenging for extremely large problems

Using Least-squares

  • Easy to Recognize

  • Weighted least-squares i=1kwi(aiTxbi)2wi移到括号里就是标准最小二乘

    • Different importance
  • Different importance i=1k(aiTxbi)2+ρi=1nxi2 也可化为标准最小二乘

    • More stable,避免过拟合

Linear Programming

  • The Problem mincTxs.t.aiTxbi,i=1,,m

    • c,a1,,amRn,b1,,bmR
  • Solving Linear Programs

    • No analytical formula for solution
    • Reliable and efficient algorithms and software
    • Computation time proportional to 𝑛2𝑚 if m𝑛; less with structure
    • A mature technology
    • Challenging for extremely large problems

Using Linear Programming

  • Not as easy to recognize
  • Chebyshev Approximation Problem

minmaxi=1,,k|aiTxbi|mints.t.t=maxi=1,,k|aiTxbi|mints.t.t|aiTxbi|,i=1,,k()mints.t.t|aiTxbi|,i=1,,kmints.t.taiTxbit,i=1,,k

Convex Optimization

  • Local minimizers are also global minimizers

  • The Problem fi(αx+βy)αfi(x)+βfi(y)

    • For all x,yRn and all α,βR with α+β=1,α0,β0

    • Least-squares and linear programs as special cases


Solving Convex Optimization Problems

  • No analytical solution
  • Reliable and efficient algorithms (e.g., interior-point methods)
  • Computation time (roughly) proportional to max{n3,n2m,F}
    • F is cost of evaluating fi​'s and their first and second derivatives
  • Almost a technology

  • Often difficult to recognize
  • Many tricks for transforming problems into convex form
  • Surprisingly many problems can be solved via convex optimization

Nonlinear Optimization

  • An optimization problem when the objective or constraint functions are not linear, but not known to be convex
  • Sadly, there are no effective methods for solving the general nonlinear programming problem
    • Could be NP-hard
  • We need compromise

Local Optimization Methods

  • Find a point that minimizes f0 among feasible points near it
    • The compromise is to give up seeking the optimal x
  • Fast, can handle large problems
    • Differentiability
  • Require initial guess
  • Provide no information about distance to (global) optimum
  • Local optimization methods are more art than technology
    • 全凭运气

Global Optimization

  • Find the global solution
    • The compromise is efficiency
  • Worst-case complexity grows exponentially with problem size
  • Applications
  • A small number of variables, where computing time is not critical
  • The value of finding the true global solution is very high

  • Worst-case Analysis of a high value or safety-critical system
    • Variables represent uncertain parameters
    • Objective function is a utility function
    • Constraints represent prior knowledge
    • If the worst-case value is acceptable, we can certify the system as safe or reliable 能证明系统可靠
  • Local optimization methods can be tried
    • If finding values that yield unacceptable performance, then the system is not reliable 局部优化已经发现了不可接受的值,那系统就不可靠
    • But it cannot certify the system as reliable 只能证明系统不可靠,不能证明可靠

Role of Convex Optimization in Nonconvex Problems

  • Initialization for local optimization
    • An approximate, but convex, formulation
  • Convex heuristics for nonconvex optimization
    • Sparse solutions (compressive sensing)
  • Bounds for global optimization 通过凸优化来找全局优化问题的下界
    • Relaxation 把条件松弛
    • Lagrangian relaxation