w i =
( 1
ifn r < n p
exp( − γ · dX i 2 )
ifn r ≥ n p
(3.6.1)
where the fator
γ
is determined byγ = − ln(ε dX ¯ res 2 )
, withdX ¯
orresponding to the(n p − n)
th best iterate. This gives the weighting fatorε res
for thispartiular residualelement. Inthisways theweighting fatorsfor the
resid-ual elements orresponding to the best iterates are kept above or equal to
the threshold value
ε res
, and all other points are onsidered low priority.Theweighting funtion for
ε r = 0.10
isseen ingure3.6.2. Ifthere arenotenough rows in the residual to math the number of unknown parameters,
allresidual rows areweighted equallywithafator
1
.IftheParameterExtrationproblemisoverdetermined,theweightingfators
makesure thatthe bestpointsaregiven the highestpriority. Asmentioned
before the weighting fators do not inuene the solution, if the system is
onsistent. The threshold valueis usedto deide,how important arole the
'unneessary'iterates should playinndingthe optimal parameter set.
If theParameter Extration problem is underdetermined, it is usually
pos-sibleto satisfyall equations exatly. Inthis asethe weighting fatorshave
no eet on the solution, sine if
r(p ∗ ) = 0
then alsoW · r(p ∗ ) = 0
. Thestrategy for determining the weighting fators inthease
n r < n p
isthere-forenot important,aslong asthe
w
'sarenot equal to zero.Astrategy for hoosing thenormalization fatorsisto put:
a j = 1
√ ε M + k f i (x j ) k 2
for
j = 1, . . . , k
(3.7.1)d = 1
√ ε M + k J f,i (x 0 ) k 2
(3.7.2)
v = 1
√ ε M + k p k k 2
(3.7.3)
wheretheiterates
x 0 , . . . , x k
arethesortediterates orrespondingto setion 3.6. Thefatorv
in(3.7.3)isusedforsalingtheregularization term,ifthis isinluded inthe residual formulation.Eah of therst
k
residual elements isthedierene between the surrogateand the ne model response in
k
dierent iteration points. We sale theseresidual elements aording to the norm of the ne model response in the
iterate.
The gradient residual is saled with the normalization fator
d
, whih isfound by means of the norm of the ne model gradient. To avoid innite
saling fators, when
k f i (x i ) k 2
ork J f,i (x 0 ) k 2
are very small, we add√ ε M
to thedenominator of all normalization fators.
Finallythe regularization term,whenpresent,issaledaordingtothe
ini-tialparametervetor
p k
,whihholdsthemappingparametersorresponding to thek
th surrogate model.Chapter 4
Test Problems
4.1 Introdution
The Spae Mapping method performed by the three algorithms in setion
2.1has been tested on various test problems. Thetwo test problems TLT2
and TLT7 arefrom theSMIS framework by Frank Pedersen. Another test
problem is the Rosenbrok funtion, whih has been tested in its lassial
formaswell asinan augmented version.
The Spae Mapping method an produe very dierent results, depending
on theimplementation and theformulation of theresidual. In all test runs
theimplementation ofthe threealgorithms orrespondto thedesriptionsin
hapter2.
ThetestshavebeenperformedontheSUNFire3800serverontheIMM
sys-tem withthe following data: 8CPU, 16 GB RAM and thelok frequeny
1200 MHz.
A very important fator for the general performane of the SpaeMapping
methodisonnetedwiththeParameter Extrationproblems. Asdesribed
intheprevious hapter the formulation of theresidual anbevaried bythe
use of normalization fators, weighting fators and regularization. These
threeapproahesan be ombinedwiththeredutionoftheparameter
ve-tor. There are manyoptions to hoose fromand every one of these options
havean inuene ontheresults.
The test investigations presented here are onerned only with the
formu-lation of the Parameter Extration problems. Through the work with the
implementation and the following test runsof thealgorithms, theeets of
a ertain approah has been somewhat laried. On the basis of this the
options have been hosen to provide dierent senarios, whih dene the
algorithms used. All senarios have been used on the three problem types.
Inevery test run we mustdene four tolerane options for usein the
stop-pingriterafor thethree algorithms. Theseoptions are:
ε F
: Stoppingriterion for the objetivefuntion.ε hx
: Stoppingriterion for the step lengthforx
-iterates.ε hp
: Stoppingriterion for the step lengthforp
-iterates.ε K
: Stoppingriterion for the gradient residual math.Inthesetup leitisurrently onlypossibleto settwotolerane parameters
ε 1
andε 2
,where the rstisthedesiredaurayfor themain problem,andthe latter is the desired auray, when solving the Parameter Extration
problem. Weuse
ε 1 = ε F = ε hx
andε 2 = ε hp = ε K
.Furthermorethemaximalnumberoffuntionevaluationsineahofthe
algo-rithmsis needed. These values areset aording to thepartiular problem.
Theoptions inmarquardt mustalso be dened f.p.19.
The tolerane options and marquardt-options in some ases have an eet
on the results. On that ground we use either one of the following sets of
options,dependingon theproblem:
• ε 1 = 10 − 14
,ε 2 = 10 − 4
and opts=
[1e-8 1e-4 1e-4 200 1e-12℄.• ε 1 = 10 − 14
,ε 2 = 10 − 14
and opts=
[1e-8 1e-14 1e-14 200 1e-12℄.Theinitial trust region radius for all test problems is
∆ (0) = 10 − 1 · k x (0) k 2
.The parameter
η
,that denes the step length indiffjaobian, is xed atη = 10 − 5
for both forward dierene approximations wrt.x
andwrt.p
.TheupdatingofthepenaltyfatorisdoneonlybythestrategyinAlgorithm
3. Finally we only onsider minimax optimization orresponding to
p = ∞
inAlgorithm1 and 2,whereas Algorithm 3 orrespondsto minimization in
the
2
-norm.4.1.1 The Test Senarios
Forshowingtheeetsofagivenapproahregardingtheresidualdenition,
we ompare the test runs from dierent residual proles. We here present
anoverviewof theinvestigated test senarios.
Regularization
Toshowtheeetsofinludingtheregularizationtermintheresidualvetor
we ompare the testruns withtheproles:
•
Regularization of theresidual vetor asdesribed in setion3.3. The regularization term isadded to theresidualvetor (1.3.12) whihnowhas
n r = k + n + n p
elements.•
No regularization of the residual vetor. The residual has the formu-lation(1.3.12) andonsistsofn r = k + n
elements.Bothasesareinludingompletenormalizationoftheresidualelements
or-responding to
a j = √ ε M + k 1 f i (x j ) k 2
forj = 1, . . . , k
andd = √ ε M + k J 1 f,i (x 0 ) k 2
.Whentheregularization termispresent,itismultiplied withthe
normaliza-tionfator
v = √ 1
ε M + k p (0) i k 2
.
The Normalization Fators
For investigating the inuene of thenormalization fators we test the
fol-lowing ases:
•
Normalization of all residual elements: The normalization fatorsare given bya j = √ ε M + k 1 f i (x j ) k 2
for
j = 1, . . . , k
andd = √ ε M + k J 1 f,i (x 0 ) k 2
.
•
Partlynormalizationoftheresidualelements: Onlythegradient resid-ualissaled orresponding toa 1 , . . . , a k = 1
andd = √ ε M + k J 1 f,i (x 0 ) k 2
.
•
No normalization ofthe residual elements:a i = 1
fori = 1, . . . , k
andd = 1
.The Weighting Fators
The eet of the weighting fatorsare onsidered and we ompare the
se-narios:
•
No weighting ofthefuntion valueresidual,w i = 1
fori = 1, . . . , k
.•
Weighting of the residual elements orresponding to the Gauss dis-tributed weight funtion, and the strategy in setion 3.6 with theweights
w 1 , . . . , w k
given by(3.6.1) .The two ases are ombined with a omplete normalization of the residual
elements.
The Number of MappingParameters
The last test senario onerns the number of mapping parameters. We
onsidertwo ases:
•
Withafull-sizeparametervetor: Thenumberofmappingparametersis
n p = n 2 + n + 1
.•
With a redued parameter vetor orresponding to using a diagonal inputmappingmatrixA
: Thenumberofmappingparameters isn p =
2n + 1
.In the problem setup-le the size of
p
is ontrolled by the ag diagA. FordiagA
= 0
theinput mapping matrixA
is full,and for diagA= 1
we haveadiagonal
A
.4.1.2 Visualization of The Results
The results of the Matlab runs of the test problems will be presented in
dierent ways. The SMIS framework ontains dierent plotting programs,
whihanbealledaftertheendedSpaeMappingiterationsforatest
prob-lem. Thefollowing plots areusedto showtheresults:
Convergene of the Iteration Sequene
Thetestproblemsallhaveknownoptimizersoftheneandtheoarsemodel,
whih makes it possible to see if the Spae Mapping method onverges or
not. Theoptimizers
x ∗
andz ∗
mustbe provided intheproblem setup-le.Twomeasuresfortheonvergenetotheoptimizerareplotted withdierent
plotsymbolsasa funtionof the iteration number:
k x (k) − x ∗ k 2
♦ F (x (k) ) − F(x ∗ )
Theseplotsareomparablewiththeresultsof[1℄. Theformulationsarestill
valid,when
x ∗
andF (x ∗ )
areequal to zero.Approximation Error
The surrogate model provides an alternative to the ne model, whih is
heaper to evaluate than the ne model. The surrogate model is used as
a loal approximation to the ne model, and the region, where the model
approximationisgood,isofinterest. WiththenewSpaeMapping
formula-tionwithbothinputandoutputmappings,thesurrogateandthenemodel
math exatly intheexpansion point. To evaluate theresults we are
inter-estedinthemodelmathinginthe region aroundtheexpansion point. The
modelmathingofthesurrogatemodelanbeompared withthemathing
ofaTaylormodel, whihisusedfor approximationinlassialoptimization
methods.
Inlassialoptimizationwithrstorderinformationavailable, weuseaT
ay-f (x + h) = f(x) + J f (x)h + O( k h k 2 2 )
andwhenonlyusingtherstorderinformation,wegetthelinearizedmodel:
l(h) = f (x) + J f (x)h
Theapproximationerror isthengiven by
f (x + h) − l(h)
. Ameasureoftheapproximationerror suitable for plottingis:
E l (h) = k f (x + h) − l(h) k 2
and it is obvious, that the error grows quadratially as we move from the
expansionpoint.
In the Spae Mapping Method we use the surrogate model as an
approxi-mation to the objetive funtion. The norm of the dierene between the
surrogateandthenemodelanbeusedasameasureoftheapproximation
error given by:
E s (h) = k f(x + h) − s(x + h) k 2
Beause of the new Spae Mapping formulation with the output mapping
we are ensured an exat funtion value math in theexpansion point. F
ur-thermorethe residualfuntionusedintheParameter Extrationguarantees
asatisfatorygradient mathinthis point.
For two-dimensional problems (
n = 2
) thetwo approximationerrorsE l
andE s
an be visualizedinthree-dimensional plots,withtheerrorsplotted ina region ofthe(x 1 , x 2 )
-plane.Diret Optimization
The SMIS framework also ontains two algorithms, whih an be used for
diret optimization based on rst order derivatives of the ne model. The
rst algorithm diret is an implementation of a rst order method whih
uses approximations to the rst order derivatives from a Broyden update.
Theseondalgorithm diretdexploitsthegradientsreturned diretlyfrom
themodelfuntions.
Theiterationsequenefromthediretoptimizationanbeusedtoompare
theSpaeMappingmethodwiththeinterpolatingsurrogatewithalassial
optimization method.