SUPPORT THE WORK

GetWiki

separation of variables

ARTICLE SUBJECTS
aesthetics  →
being  →
complexity  →
database  →
enterprise  →
ethics  →
fiction  →
history  →
internet  →
knowledge  →
language  →
licensing  →
linux  →
logic  →
method  →
news  →
perception  →
philosophy  →
policy  →
purpose  →
religion  →
science  →
sociology  →
software  →
truth  →
unix  →
wiki  →
ARTICLE TYPES
essay  →
feed  →
help  →
system  →
wiki  →
ARTICLE ORIGINS
critical  →
discussion  →
forked  →
imported  →
original  →
separation of variables
[ temporary import ]
please note:
- the content below is remote from Wikipedia
- it has been imported raw for GetWiki
{{short description|Technique for solving differential equations}}{{Differential equations}}In mathematics, separation of variables (also known as the Fourier method) is any of several methods for solving ordinary and partial differential equations, in which algebra allows one to rewrite an equation so that each of two variables occurs on a different side of the equation.(File:Sept proportional differential equation.png|thumb|Solve proportional first order differential equation{{cn|date=April 2024}})(File:Sept linear differential equation.png|thumb|Solve linear first order differential equation{{cn|date=April 2024}} by separation of variables.WEB, Separation of Variables,www.mathsisfun.com/calculus/separation-variables.html, 2021-09-18, www.mathsisfun.com, )

Ordinary differential equations (ODE)

A differential equation for the unknown f(x) will be separable if it can be written in the form
frac{d}{dx} f(x) = g(x)h(f(x))
where g and h are given functions. This is perhaps more transparent when written using y = f(x) as:
frac{dy}{dx}=g(x)h(y).
So now as long as h(y) ≠ 0, we can rearrange terms to obtain:
{dy over h(y)} = g(x) , dx,
where the two variables x and y have been separated. Note dx (and dy) can be viewed, at a simple level, as just a convenient notation, which provides a handy mnemonic aid for assisting with manipulations. A formal definition of dx as a differential (infinitesimal) is somewhat advanced.

Alternative notation

Those who dislike Leibniz’s notation may prefer to write this as
frac{1}{h(y)} frac{dy}{dx} = g(x),
but that fails to make it quite as obvious why this is called “separation of variables”. Integrating both sides of the equation with respect to x, we have{{NumBlk|:|int frac{1}{h(y)} frac{dy}{dx} , dx = int g(x) , dx, |{{EquationRef|A1}}}}or equivalently,
int frac{1}{h(y)} , dy = int g(x) , dx
because of the substitution rule for integrals.If one can evaluate the two integrals, one can find a solution to the differential equation. Observe that this process effectively allows us to treat the derivative frac{dy}{dx} as a fraction which can be separated. This allows us to solve separable differential equations more conveniently, as demonstrated in the example below.(Note that we do not need to use two constants of integration, in equation ({{EquationNote|A1}}) as in
int frac{1}{h(y)} , dy + C_1 = int g(x) , dx + C_2,
because a single constant C = C_2 - C_1 is equivalent.)

Example

Population growth is often modeled by the “logistic” differential equation
frac{dP}{dt}=kPleft(1-frac{P}{K}right)
where P is the population with respect to time t, k is the rate of growth, and K is the carrying capacity of the environment.Separation of variables now leads to
begin{align}& intfrac{dP}{Pleft(1-P/K right)}=int k,dtend{align}which is readily integrated using partial fractions on the left side yielding
P(t)=frac{K}{1+Ae^{-kt}}
where A is the constant of integration. We can find A in terms of Pleft(0right)=P_0 at t=0. Noting e^0=1 we get
A=frac{K-P_0}{P_0}.

Generalization of separable ODEs to the nth order

Much like one can speak of a separable first-order ODE, one can speak of a separable second-order, third-order or nth-order ODE. Consider the separable first-order ODE:
frac{dy}{dx}=f(y)g(x)
The derivative can alternatively be written the following way to underscore that it is an operator working on the unknown function, y:
frac{dy}{dx}=frac{d}{dx}(y)
Thus, when one separates variables for first-order equations, one in fact moves the dx denominator of the operator to the side with the x variable, and the d(y) is left on the side with the y variable. The second-derivative operator, by analogy, breaks down as follows:
frac{d^2y}{dx^2} = frac{d}{dx}left(frac{dy}{dx}right) = frac{d}{dx}left(frac{d}{dx}(y)right)
The third-, fourth- and nth-derivative operators break down in the same way. Thus, much like a first-order separable ODE is reducible to the form
frac{dy}{dx}=f(y)g(x)
a separable second-order ODE is reducible to the form
frac{d^2y}{dx^2}=fleft(y’right)g(x)
and an nth-order separable ODE is reducible to
frac{d^ny}{dx^n}=f!left(y^{(n-1)}right)g(x)

Example

Consider the simple nonlinear second-order differential equation:y=(y’)^2.This equation is an equation only of y’ and y, meaning it is reducible to the general form described above and is, therefore, separable. Since it is a second-order separable equation, collect all x variables on one side and all y variables on the other to get:frac{d(y’)}{(y’)^2}=dx.Now, integrate the right side with respect to x and the left with respect to ‘’y:int frac{d(y’)}{(y’)^2}=int dx.This gives-frac{1}{y’}=x+C_1,which simplifies to:y’=-frac{1}{x+C_1}~.This is now a simple integral problem that gives the final answer:y=C_2-ln|x+C_1|.

Partial differential equations{{anchor|pde}}

{{See also|Separable partial differential equation}}The method of separation of variables is also used to solve a wide range of linear partial differential equations with boundary and initial conditions, such as the heat equation, wave equation, Laplace equation, Helmholtz equation and biharmonic equation.The analytical method of separation of variables for solving partial differential equations has also been generalized into a computational method of decomposition in invariant structures that can be used to solve systems of partial differential equations.BOOK,books.google.com/books?id=kxVDDwAAQBAJ, Harmonic Wave Systems: Partial Differential Equations of the Helmholtz Decomposition, 9781618964069, Miroshnikov, Victor A., 15 December 2017,

Example: homogeneous case

Consider the one-dimensional heat equation. The equation is{{NumBlk|:|frac{partial u}{partial t} - alphafrac{partial^{2}u}{partial x^{2}} = 0|{{EqRef|1}}}}The variable u denotes temperature. The boundary condition is homogeneous, that is{{NumBlk|:|ubig|_{x=0}=ubig|_{x=L}=0 |{{EqRef|2}}}}Let us attempt to find a solution which is not identically zero satisfying the boundary conditions but with the following property: u is a product in which the dependence of u on x, t is separated, that is:{{NumBlk|:| u(x,t) = X(x) T(t).|{{EqRef|3}}}}Substituting u back into equation {{EqNote|1}} and using the product rule,{{NumBlk|:|frac{T’(t)}{alpha T(t)} = frac{X’’(x)}{X(x)}.|{{EqRef|4}}}}Since the right hand side depends only on x and the left hand side only on t, both sides are equal to some constant value −λ. Thus:{{NumBlk|:|T’(t) = - lambda alpha T(t),|{{EqRef|5}}}}and{{NumBlk|:|X’’(x) = - lambda X(x).|{{EqRef|6}}}}−λ here is the eigenvalue for both differential operators, and T(t) and X(x) are corresponding eigenfunctions.We will now show that solutions for X(x) for values of λ ≤ 0 cannot occur:Suppose that λ < 0. Then there exist real numbers B, C such that
X(x) = B e^{sqrt{-lambda} , x} + C e^{-sqrt{-lambda} , x}.
From {{EqNote|2}} we get{{NumBlk|:|X(0) = 0 = X(L),|{{EqRef|7}}}}and therefore B = 0 = C which implies u is identically 0.Suppose that λ = 0. Then there exist real numbers B, C such that
X(x) = Bx + C.
From {{EqNote|7}} we conclude in the same manner as in 1 that u is identically 0.Therefore, it must be the case that λ > 0. Then there exist real numbers A, B, C such that
T(t) = A e^{-lambda alpha t},
and
X(x) = B sin(sqrt{lambda} , x) + C cos(sqrt{lambda} , x).
From {{EqNote|7}} we get C = 0 and that for some positive integer n,
sqrt{lambda} = n frac{pi}{L}.
This solves the heat equation in the special case that the dependence of u has the special form of {{EqNote|3}}.In general, the sum of solutions to {{EqNote|1}} which satisfy the boundary conditions {{EqNote|2}} also satisfies {{EqNote|1}} and {{EqNote|3}}. Hence a complete solution can be given as
u(x,t) = sum_{n = 1}^{infty} D_n sin frac{npi x}{L} expleft(-frac{n^2 pi^2 alpha t}{L^2}right),
where Dn are coefficients determined by initial condition.Given the initial condition
ubig|_{t=0}=f(x),
we can get
f(x) = sum_{n = 1}^{infty} D_n sin frac{npi x}{L}.
This is the sine series expansion of f(x) which is amenable to Fourier analysis. Multiplying both sides with sin frac{npi x}{L} and integrating over {{closed-closed|0, L}} results in
D_n = frac{2}{L} int_0^L f(x) sin frac{npi x}{L} , dx.
This method requires that the eigenfunctions X, here left{sin frac{npi x}{L}right}_{n=1}^{infty}, are orthogonal and complete. In general this is guaranteed by Sturm–Liouville theory.

Example: nonhomogeneous case

Suppose the equation is nonhomogeneous,{{NumBlk|:|frac{partial u}{partial t}-alphafrac{partial^{2}u}{partial x^{2}}=h(x,t)|{{EqRef|8}}}}with the boundary condition the same as {{EqNote|2}}.Expand h(x,t), u(x,t) and f(x) into{{NumBlk|:|h(x,t)=sum_{n=1}^{infty}h_{n}(t)sinfrac{npi x}{L},|{{EqRef|9}}}}{{NumBlk|:|u(x,t)=sum_{n=1}^{infty}u_{n}(t)sinfrac{npi x}{L},|{{EqRef|10}}}}{{NumBlk|:|f(x)=sum_{n=1}^{infty}b_{n}sinfrac{npi x}{L},|{{EqRef|11}}}}where h’n(t) and b’n can be calculated by integration, while un(t) is to be determined.Substitute {{EqNote|9}} and {{EqNote|10}} back to {{EqNote|8}} and considering the orthogonality of sine functions we get
u’_{n}(t)+alphafrac{n^{2}pi^{2}}{L^{2}}u_{n}(t)=h_{n}(t),
which are a sequence of linear differential equations that can be readily solved with, for instance, Laplace transform, or Integrating factor. Finally, we can get
u_{n}(t)=e^{-alphafrac{n^{2}pi^{2}}{L^{2}} t} left (b_{n}+int_{0}^h_{n}(s)e^{alphafrac{n^{2}pi^{2}}{L^{2}} s} , ds right).
If the boundary condition is nonhomogeneous, then the expansion of {{EqNote|9}} and {{EqNote|10}} is no longer valid. One has to find a function v that satisfies the boundary condition only, and subtract it from u. The function u-v then satisfies homogeneous boundary condition, and can be solved with the above method.

Example: mixed derivatives

For some equations involving mixed derivatives, the equation does not separate as easily as the heat equation did in the first example above, but nonetheless separation of variables may still be applied. Consider the two-dimensional biharmonic equation
frac{partial^4 u}{partial x^4} + 2frac{partial^4 u}{partial x^2partial y^2} + frac{partial^4 u}{partial y^4} = 0.
Proceeding in the usual manner, we look for solutions of the form
u(x,y) = X(x)Y(y)
and we obtain the equation
frac{X^{(4)}(x)}{X(x)} + 2frac{X(x)}{X(x)}frac{Y(y)}{Y(y)} + frac{Y^{(4)}(y)}{Y(y)} = 0.
Writing this equation in the form
E(x) + F(x)G(y) + H(y) = 0,
Taking the derivative of this expression with respect to x gives E’(x)+F’(x)G(y)=0 which means G(y)=const. or F’(x)=0 and likewise, taking derivative with respect to y leads to F(x)G’(y)+H’(y)=0 and thus F(x)=const. or G’(y)=0 , hence either F(x) or G(y) must be a constant, say −λ. This further implies that either -E(x)=F(x)G(y)+H(y) or -H(y)=E(x)+F(x)G(y) are constant. Returning to the equation for X and Y, we have two cases
begin{align}
X’’(x) &= -lambda_1X(x)
X^{(4)}(x) &= mu_1X(x)
Y^{(4)}(y) - 2lambda_1Y’’(y) &= -mu_1Y(y)
end{align}and
begin{align}
Y’’(y) &= -lambda_2Y(y)
Y^{(4)}(y) &= mu_2Y(y)
X^{(4)}(x) - 2lambda_2X’’(x) &= -mu_2X(x)
end{align}which can each be solved by considering the separate cases for lambda_i0 and noting that mu_i=lambda_i^2.

Curvilinear coordinates

In orthogonal curvilinear coordinates, separation of variables can still be used, but in some details different from that in Cartesian coordinates. For instance, regularity or periodic condition may determine the eigenvalues in place of boundary conditions. See spherical harmonics for example.

Applicability

Partial differential equations

For many PDEs, such as the wave equation, Helmholtz equation and Schrodinger equation, the applicability of separation of variables is a result of the spectral theorem. In some cases, separation of variables may not be possible. Separation of variables may be possible in some coordinate systems but not others,John Renze, Eric W. Weisstein, Separation of variables and which coordinate systems allow for separation depends on the symmetry properties of the equation.Willard Miller(1984) Symmetry and Separation of Variables, Cambridge University Press Below is an outline of an argument demonstrating the applicability of the method to certain linear equations, although the precise method may differ in individual cases (for instance in the biharmonic equation above).Consider an initial boundary value problem for a function u(x,t) on D = {(x,t): x in [0,l], t geq 0 } in two variables:
(Tu)(x,t) = (Su)(x,t)
where T is a differential operator with respect to x and S is a differential operator with respect to t with boundary data:
(Tu)(0,t) = (Tu)(l,t) = 0 for t geq 0 (Su)(x,0)=h(x) for 0 leq x leq l
where h is a known function.We look for solutions of the form u(x,t) = f(x) g(t). Dividing the PDE through by f(x)g(t) gives
frac{Tf}{f} = frac{Sg}{g}
The right hand side depends only on x and the left hand side only on t so both must be equal to a constant K , which gives two ordinary differential equations
Tf = Kf, Sg = Kg
which we can recognize as eigenvalue problems for the operators for T and S. If T is a compact, self-adjoint operator on the space L^2[0,l] along with the relevant boundary conditions, then by the Spectral theorem there exists a basis for L^2[0,l] consisting of eigenfunctions for T. Let the spectrum of T be E and let f_{lambda} be an eigenfunction with eigenvalue lambda in E. Then for any function which at each time t is square-integrable with respect to x, we can write this function as a linear combination of the f_{lambda}. In particular, we know the solution u can be written as
u(x,t) = sum_{lambda in E} c_{lambda}(t)f_{lambda}(x)
For some functions c_{lambda}(t). In the separation of variables, these functions are given by solutions to Sg = Kg Hence, the spectral theorem ensures that the separation of variables will (when it is possible) find all the solutions.For many differential operators, such as frac{d^2}{dx^2}, we can show that they are self-adjoint by integration by parts. While these operators may not be compact, their inverses (when they exist) may be, as in the case of the wave equation, and these inverses have the same eigenfunctions and eigenvalues as the original operator (with the possible exception of zero).David Benson (2007) Music: A Mathematical Offering, Cambridge University Press, Appendix W

Matrices

The matrix form of the separation of variables is the Kronecker sum.As an example we consider the 2D discrete Laplacian on a regular grid:
L = mathbf{D_{xx}}oplusmathbf{D_{yy}}=mathbf{D_{xx}}otimesmathbf{I}+mathbf{I}otimesmathbf{D_{yy}}, ,
where mathbf{D_{xx}} and mathbf{D_{yy}} are 1D discrete Laplacians in the x- and y-directions, correspondingly, and mathbf{I} are the identities of appropriate sizes. See the main article Kronecker sum of discrete Laplacians for details.

Software

Some mathematical programs are able to do separation of variables: XcasWEB,www-fourier.ujf-grenoble.fr/~parisse/giac/cascmd_en.pdf, Symbolic algebra and Mathematics with Xcas, among others.

See also

Notes

{{Reflist}}

References

[ this remote article is provided by Wikipedia ]
LATEST EDITS [ see all ]
GETWIKI 21 MAY 2024
GETWIKI 09 JUL 2019
Eastern Philosophy
History of Philosophy
GETWIKI 09 MAY 2016
GETWIKI 18 OCT 2015
M.R.M. Parrott
Biographies
GETWIKI 20 AUG 2014
CONNECT