
by Neil E. Cotter
Research
Assistant Professor
(� = missing, * = article)
Optimization
Lagrange Multipliers
Karush-Kuhn-Tucker Theorem
Descent Algorithms
Newton's method
Line search
Gradient (or steepest) descent
Conjugate gradient methods
Polak-Ribiere method
Modified Newton's method
Linear Optimization
Linear programming
Simplex algorithm
Interior-point algorithms
Quadratic Programming
Global Optimization
Downhill simplex algorithm
Genetic algorithms
�Diffusion
algorithms
Random fields
Markov random fields (MRF)
Gibbs random fields (GRF)
Equivalence of MRF and GRF
Convergence
Notation
Kolmogorov-Chapman equation
Forgetting initial conditions
�Approaches
invariant distribution
�Invariant
distribution is Gibbs
�Freezing to
global minima
Metropolis algorithm
Simulated annealing
*Prejudicial search
Discrete Optimization
Dynamic programming
NP-completene problems
Downhill simplex algorithm
Continuation
�Dynamic
programming: Werbos
�Random
optimization
Stochastic approx
CMAC learning
Gradient descent
�K-means squared