You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: vignettes/Optimalgo.Rmd
+21-21Lines changed: 21 additions & 21 deletions
Original file line number
Diff line number
Diff line change
@@ -24,15 +24,15 @@ options(digits = 3)
24
24
```
25
25
26
26
27
-
# Quick overview of main optimization methods
27
+
# 1. Quick overview of main optimization methods
28
28
29
29
We present very quickly the main optimization methods.
30
30
Please refer to **Numerical Optimization (Nocedal \& Wright, 2006)**
31
31
or **Numerical Optimization: theoretical and practical aspects
32
32
(Bonnans, Gilbert, Lemarechal \& Sagastizabal, 2006)** for a good introduction.
33
33
We consider the following problem $\min_x f(x)$ for $x\in\mathbb{R}^n$.
34
34
35
-
## Derivative-free optimization methods
35
+
## 1.1. Derivative-free optimization methods
36
36
The Nelder-Mead method is one of the most well known derivative-free methods
37
37
that use only values of $f$ to search for the minimum.
38
38
It consists in building a simplex of $n+1$ points and moving/shrinking
@@ -67,12 +67,12 @@ this simplex into the good direction.
67
67
The Nelder-Mead method is available in `optim`.
68
68
By default, in `optim`, $\alpha=1$, $\beta=1/2$, $\gamma=2$ and $\sigma=1/2$.
69
69
70
-
## Hessian-free optimization methods
70
+
## 1.2. Hessian-free optimization methods
71
71
72
72
For smooth non-linear function, the following method is generally used:
73
73
a local method combined with line search work on the scheme $x_{k+1} =x_k + t_k d_{k}$, where the local method will specify the direction $d_k$ and the line search will specify the step size $t_k \in \mathbb{R}$.
74
74
75
-
### Computing the direction $d_k$
75
+
### 1.2.1. Computing the direction $d_k$
76
76
A desirable property for $d_k$ is that $d_k$ ensures a descent $f(x_{k+1}) < f(x_{k})$.
77
77
Newton methods are such that $d_k$ minimizes a local quadratic approximation of $f$ based on a Taylor expansion, that is $q_f(d) = f(x_k) + g(x_k)^Td +\frac{1}{2} d^T H(x_k) d$ where $g$ denotes the gradient and $H$ denotes the Hessian.
78
78
@@ -121,7 +121,7 @@ See Yuan (2006) for other well-known schemes such as Hestenses-Stiefel, Dixon or
121
121
The three updates (Fletcher-Reeves, Polak-Ribiere, Beale-Sorenson) of the (non-linear) conjugate gradient are available in `optim`.
122
122
123
123
124
-
### Computing the stepsize $t_k$
124
+
### 1.2.2. Computing the stepsize $t_k$
125
125
126
126
Let $\phi_k(t) = f(x_k + t d_k)$ for a given direction/iterate $(d_k, x_k)$.
127
127
We need to find conditions to find a satisfactory stepsize $t_k$. In literature, we consider the descent condition: $\phi_k'(0) < 0$
0 commit comments