Research Article Volume 5 Issue 3
Department of Statistics, University of Florida, USA
Correspondence: Andre I. Khuri, Department of Statistics, University of Florida, USA
Received: October 29, 2017 | Published: March 10, 2017
Citation: Khuri AI. A general overview of response surface methodology. Biom Biostat Int J. 2017;5(3):87-93. DOI: 10.15406/bbij.2017.05.00133
Response surface methodology (RSM) is a collection of statistical and mathematical techniques used for the purpose of
Formal work on RSM began with the publication of the article On the Experimental Attainment of Optimum Conditions by Box et al.1
See also the article by Steinberg and Hunter.5
The present article provides a general overview of response surface methodology (RSM).
Let η denote the mean of a response variable y . Let x1,…,xk denote a set of input, or control, variables. Then,
η(x)=ϕ(x1,…,xk).
For example, we have the first-order model
η(x)=β0+k∑i=1βixi
or the second-order model
η(x)=β0+k∑i=1βixi+k∑i=1k∑j=1βijxixj
y=η+∈,
where E(∈)=0 and Var(∈)=σ2∈
Let xui denote the design setting of variable xi at the uth experimental run (u=1,2,…,n;i=1,2,…,k) . Let yu denote the corresponding response value. By definition, the design matrix D is the n×k matrix
D=[x11x12…x1kx21x22…x2k…………xn1xn2…xnk]
We have the linear model
y=Xβ+∈,
where
X
is
n×p
of rank p and
β
is a vector of unknown parameters.
The least-squares estimator of
β
is
ˆβ=(X'X)−1X'y
Var(ˆβ)=(X'X)−1σ2∈
ˆy(x)=f'(x)ˆβ,
where
f'(x)
is a vector of order
1×p
of the same form as a row of
X
, but
is evaluated at the point
x=(x1,…,xk)'
For a first-order model, the predicted response is
ˆy(x)=ˆβ0+k∑i=1ˆβixi,
and for a second-order model we have
ˆy(x)=ˆβ0+k∑i=1ˆβixi,+k∑i=1k∑j=1ˆβijxixj.
The prediction variance is given by
Var[ˆy(x)]=σ2∈f'(x)(X'X)−1f(x).
Some important properties of a response surface design (Box and Draper)8:
|X'X|−12 is propotional to the volume of an ellipsoidal confidence region on β .
Examples:
The fitted model is
y=f'(x)β+∈
The “True" model is
η(x)=f'(x)β+g'(x)γ.
Integrated Mean Squared Error Criterion:10,11 This amounts to the minimization of J, where
J=nΩσ2∈∫RE[ˆy(x)−η(x)]2dx,
Ω
is the reciprocal of the volume of
R
.
J=V+B.
V=nΩσ2∈∫RVar[ˆy(x)] dx
V=nΩ∫Rf'(x)(X'X)−1f(x)dx
B=nΩσ2∈∫R{E[ˆy(x)]−η(x)}2dx
B=nΩσ2∈∫Rγ'[A'f(x)−g(x)][f'(x)A−g'(x)]γdx
where
A=(X'X)−1X'Z
is the alias matrix, and
z
is the matrix corresponding to
g'(x)
The Box and Draper criterion for the minimization of
B
is:
1nX'X=Ω∫Rf(x)f'(x)dx
1nX'Z=Ω∫Rf(x)g'(x)dx
Designs that minimize J have characteristics similar to those of designs which minimize
B
alone (all bias designs).
A design is robust if it helps reduce the impact of nonideal conditions on data analysis. Some of these conditions include
Box12 introduced the word “robust" when examining the effect of departure from normality on the analysis of variance. Some relevant articles are:13,15,16,48
These are designs for the estimation of the partial derivatives of the mean response
η(x)
with respect
x1,…,xk
Myers and Lahoda17 extended the Box-Draper integrated mean squared error criterion to find appropriate designs for the joint estimation of the partial derivatives of
η(x)
.
Hader and Park18 introduced the concept of design rotatability for second-order models. Under this design criterion,
Var[∂ˆy(x)∂xi]
is constant at all points that are equidistant from the design center
(i=1,2,…,k)
Park19 considered “slope rotatability over all directions":
ˆy(x)=f'(x)ˆβ.
∂ˆy∂v=v'grad[ˆy(x)]=v'Gˆβ,
where
v
is a unit vector.
Var[∂ˆy(x)∂v]=σ2∈v'G(X'X)−1G'v
AvgvVar[∂ˆy(x)∂v]=c∫sVar[∂ˆy(x)∂v]dA,
where S denotes the surface of a hypersphere of unit radius, c is the reciprocal of the surface of S .
W(x)=Ave.SlopeVar.=σ2∈ktr[G(X'X)−1G']
A design
D
is slope rotatable over all directions if
W(x)
is constant at all points equidistant from the design center.
Park gave necessary and sufficient conditions for a design to be slope rotatable over all directions for second-order models.
Rotatability⇒SlopeRotatabilityOverallDirections
Related articles
These are designs that induce a certain degree of sensitivity to possible inadequacy of the fitted model.
Construction of such designs is done by the maximization of
λ=γ'Lγ,
L=Z'[I−X(X'X)−1X']Z.
Λ1
optimality: Maximize
Λ1
, the minimum of
λ
over a specified region
ϕ
in the
γ
space.
Λ2
optimality: Maximize
Λ2
, the average value of
λ
over the boundary of
ϕ
.
Estimation of Optimum Conditions.
The Method of Steepest Ascent This is a maximum-region-seeking procedure introduced by Box and Wilson1 for RSM: To maximize
ˆy(x)=ˆβ0+k∑i=1ˆβixi
subject to the constraint ∑ki=1x2i=r2.
xi=rυi,
where υ=(υ1,…,υk)' is a unit vector in the direction of grad[ˆy(x)] A related article is Myers and Khuri.23
Related articles include Hoerl,24 Draper25
To maximize (minimize)
ˆy(x)=ˆβ0+k∑i=1ˆβixi+k∑i=1k∑j=1βijxixj
ˆy(x)=ˆβ0+x'b+x'Bx
subject to the constraint
∑ki=1x2i=r2
h(x)=ˆy(x)−μ(k∑i=1x2i−r2),
where
μ
is a Lagrange multiplier.
For a maximum,
μ
should be larger than
emax(B)
.
For a minimum,
μ
should be smaller than
emin(B)
.
By selecting several values of
μ
as shown above, we obtain plots of
max.
ˆy
vs. r and plots of
xi
vs.
r
for
i=1,2,...,k
Khuri and Myers26 introduced a certain modification to the method of ridge analysis in cases in which the design is not rotatable and may even be ill conditioned. In such cases,
Var[ˆy(x)]
can vary appreciably on a hypersphere
S(r)
centered at the origin inside a region of interest. The proposed modification optimizes
ˆy(x)
on
S(r)
subject to maintaining certain a constraint on the size of
Var[ˆy(x)]:
Var[ˆy(x)]=σ2∈f'(x)(X'X)−1f(x)
Var[ˆy(x)]=σ2∈p∑i=1[f'(x)ωi]2vi
f'(x)ωi=ai+x'τi+x'Tix,
where
p
is the number of parameters in the model,
ω1,...,ωp
are orthonor-mal eigenvectors of
X'X
, and
v1,....,vp
are the corresponding eigenvalues of
X'X
. Let
vmin
be the smallest eigenvalue of
X'X
and let
ωm
be the corresponding eigenvector. To optimize
ˆy(x)
subject to the constraints
k∑i=1x2i=r2
|f'(x)ωm| ≤ q
where
q
is a small positive constant, or equivalently,
|am+x'τm+x'Tmx| ≤ q
The estimated optimum of a mean response and the location of the optimum are obtained by using ˆy(x) . They are therefore random variables. In order to have a better assessment of optimum conditions, confidence intervals on the optimum of the mean response as well as confidence regions on the location of the optimum are needed. Relevant references include
A design is rotattable if and only if the X'X matrix has a particular form, which we denote by X'Xrot . For any given design D , the “closeness" of the corresponding X'X matrix to X'Xrot can be determined. On this basis, a measure ψ(D) , which quantifies the percent rotatability that is inherent in D , can be derived. A design D is rotatable if and only if ψ(D)=100 .
For example, the
32
factorial design is 93.08% rotatable.
Properties of
ψ(D)
:
Draper and Pukelsheim32 introduced a similar measure that is invariant to design rotation, but it applies only to second-order models.
Giovannitti-jensen and myers (1989)
They plotted maximum and minimum values of
1σ2∈Var|ˆy(x)|
, on a hyperspher
S(r)
, against
r
.
Relevant Reference are Khuri33 used the measure of rotatability
ψ(D)
to derive several upper bounds on the range of
1σ2∈Var|ˆy(x)|
. These bounds are easy to compute since they only require determining eigenvalues and traces associated with the matrices
X'X
and
F
, where
F=(X'X)−1− U−1rot
and Urot is the “rotatable" portion of X'X .
Park and Kim34 defined a measure Qk(D) to quantify the degree of slope rotatability over axial directions18 for second-order models. It takes the value zero if and only if the design is slope rotatable. It can also be used to “repair" slope rotatability by design augmentation.
Jang and Park35 considered the maximum, Vmax(r) , and minimum, Vmin(r) , values of 1σ2∈W(x) over a hypersphere S(r) , where
W(x)=σ2∈ktr[G(X'X)−1G']
The quantity
Δ(r)=Vmin(r)−Vmax(r)
is zero if and only if the design is slope rotatable over all directions. Jang and Park plotted
Vmax(r)
and
Vmin(r)
against
r
(slope variance dispersion graphs).
Khuri and vining36
Specification Problem: To determine conditions on the input variables in a process that cause the mean response η(x) to fall within specified bounds, e.g., α<η(x)<b , where a and b are given, with a specified degree of confidence.
Procedure
Start with an initial n-point design
Dn
Calculate
ˆβn=(X'nXn)−1X'nyn
ˆy(x)=f'(x)ˆβn
Compute
y1n(x)=ˆy(x)−STDntα/2,n−p
y2n(x)=ˆy(x)+STDntα/2,n−p,
Where
STDn=[MSEf'(x)(X'nXn)−1f(x)]1/2
If there is a point
x0
such that
α<y1n(x0),y2n(x0)<b
, then
x0
has a probability of at least
1−α
of satisfying
α<η(x0)<b,
x0
is a solution to the specification problem.
Khuri and Vining presented a sequential procedure of adding points to
Dn
, if necessary, until a solution can be found to the inequalities
α<y1n(x),y2n(x)<b
for a sufficiently large n .
An experiment in which a number of responses can be measured for each setting of a group of input variable, x1,...,xk , is called a multiresponse experiment.
When several responses are considered simultaneously, any statistical investigation of the responses should take into consideration the multi- variate nature of the data. The response variables should not be analyzed individually or independently of one another. This is particularly true when the response variables are correlated.
Traditional response surface techniques that apply to single response models are, in general, not adequate to analyze multiresponse models.
In particular, if the individual response models are linear, then
yi=Xiβi+∈i i=1,2,…,r.
Y=F(D,B)+∈
An estimate of
β
is obtained by minimizing the determinant of
S(D,β)=[Y−F(D,β)]'[Y−F(D,β)]
Relevant articles include Box, Hunter, MacGregor, and Erjavec;38 Khuri.39
The r linear response models,
yi=Xiβi+∈i, i=1,2,...,r,
can be represented as a single linear multiresponse model
y=XΘ+δ,
Where
y=[y'1,y'2,…,y'r]'
X is a block-diagonal matrix diag(X1,X2,…,Xr) ,
Θ=[β'1,β'2,…,β'r]' and,
δ=[∈'1,∈'2,…,∈'r]'
The variance-covariance matrix of is given by the direct (Kronecker) product ∑⊗In
ˆΘ=[X'(∑−1⊗In)X]1X'(∑−1⊗In)y
If
∑
is unknown, an estimate
∑=(ˆσij)
can be used, where
ˆσij=1ny'i[In−Pj]yj, i,j=1,2,...,r,
and
Pi=Xi(X'iXi)1X'i
i=1,2,…,r.
Zelner40
In particular, if
X1=X2=⋯=Xr=X0
, then,
ˆβi=(X'0X0)−1X'0yi,
Draper and Hunter,41 Fedorov):42 D-optimal designs.
∑
must be known. Wijesinha and Khuri43 used an estimate of
∑
. Wijesinha and Khuri44 considered designs that maximize the power of the multi variate lack fit test.
Khuri:45 Multiresponse rotatability for linear multirespose models.
A design is multiresponse rotatable if
Var[ˆy(x)]
is constant at all points x that are equidistant from the origin, where
ˆy(x)
is the vector of r predicted responses.
This design property can be achieved if and only if the design is rotatable for a single-response model having the highest degree among all the r response models.
Khuri46
The models
yi=Xiβi+∈i, i=1,2,…,r
can be written as
Y=X*B+∈
where
Y=[y1: y2: ... :yr]
X*=[X1: X2: … : Xr]
B=diag(β1,β2, ... ,βr)
∈=[∈1 : ∈2 : ... : ∈r]
Assume that replicated observations are available on all the responses at some points in a region of interest.
The above model suffers from lack of fit if and only if there exists a linear combination of the responses that suffers from lack of fit.
The multivariate lack of fit test is Roy's largest-root test, emax(G2−1G1)
Where
G1=Y'[In−X(X'X)−1X'−K]Y
G2=Y'KY
K=diag(K1,K2, … ,Km,0)
Ki=Ivi−1viJvi i=1,2,…,m
(
vi
repeated observations are taken at the
i
th design point
(i=1,2, … ,m)
).
Large values of the test statistic are significant.
Note: The test may be significant even though the response models do not individually exhibit a significant lack of fit.
Richert, et al.47 investigated the effects of heating temperature (x1) , pH level (x2) , redox potential (x3) , sodium oxalate (x4) , and sodium lauryl sulfate (x5) on foaming properties of whey protein concentrates. The responses are y1 =whipping time
y2 =maximum overrun
y3 =percent soluble protein.
A second-order model was assumed for each response.
Khuri and Conlon48
The purpose of a multiresponse optimization technique is to find operating conditions on the input variables that lead to optimal, or near optimal, values of several response functions.
yi=X0βi+∈i, i=1,2,…,r.
In this case, an unbiased estimate of ∑ , the variance-covariance matrix of ther response variables, is given by
∑u=Y'[In−X0(X'0X0)−1X'0]Y/(n−p),
r≤n−p.
ˆyi(x)=f'(x)ˆβi, i=1,2,…,r.
Let ˆy(x)=[ˆy1(x),ˆy2(x), … ,ˆyr(x)]'.
Let φi be the optimum value of ˆyi(x) optimized individually over a region of interest R (i=1,2, … ,r) .
φ=(φ1,φ2, … ,φr)'
Var[ˆy(x)]=f'(x)(X'0X0)−1f(x)∑
An “ideal” optimum occurs when ˆyi(x) , for i=1,2,…,r, , attain their individual optima at the same set of conditions.
To find “compromise conditions" that are somewhat favorable to all the responses, we consider the metric (distance function)
ρ[ˆy(x),φ]=[(ˆy(x)−φ)'{Var[ˆy(x)]}−1(ˆy(x)−φ)]1/2
Multiresponse optimization is then reduced to minimizing ρ[ˆy(x),φ] with respect to x over R .
None.
None.
©2017 Khuri. This is an open access article distributed under the terms of the, which permits unrestricted use, distribution, and build upon your work non-commercially.
2 7