Ordinary differential equation

Ordinary differential equation
In mathematics, an ordinary differential equation (or ODE) is a relation that contains functions of only one independent variable, and one or more of its derivatives with respect to that variable.

A simple example is Newton's second law of motion, which leads to the differential equation

$$ m \frac{d^2 x(t)}{dt^2} = F(x(t)),\, $$

for the motion of a particle of mass m. In general, the force F depends upon the position of the particle x(t) at time t, and thus the unknown function x(t) appears on both sides of the differential equation, as is indicated in the notation F(x(t)).

Ordinary differential equations are distinguished from partial differential equations, which involve partial derivatives of several variables.

Ordinary differential equations arise in many different contexts including geometry, mechanics, astronomy and population modelling. Many famous mathematicians have studied differential equations and contributed to the field, including Newton, Leibniz, the Bernoulli family, Riccati, Clairaut, d'Alembert and Euler.

Much study has been devoted to the solution of ordinary differential equations. In the case where the equation is linear, it can be solved by analytical methods. Unfortunately, most of the interesting differential equations are non-linear and, with a few exceptions, cannot be solved exactly. Approximate solutions are arrived at using computer approximations (see numerical ordinary differential equations). The trajectory of a projectile launched from a cannon follows a curve determined by an ordinary differential equation that is derived from Newton's second law. Contents

* 1 Definitions o 1.1 Ordinary differential equation o 1.2 Solutions * 2 Examples * 3 Reduction to a first order system * 4 Linear ordinary differential equations o 4.1 Homogeneous equations o 4.2 Nonhomogeneous equations o 4.3 Fundamental systems for homogeneous equations with constant coefficients * 5 Theories of ODEs o 5.1 Singular solutions o 5.2 Reduction to quadratures o 5.3 Fuchsian theory o 5.4 Lie's theory o 5.5 Sturm-Liouville theory * 6 See also * 7 Bibliography * 8 External links

[edit] Definitions

[edit] Ordinary differential equation

Let y be an unknown function

$$y: \mathbb{R} \to \mathbb{R}$$

in x with y(n) the nth derivative of y, then an equation of the form

$$F(x,y,y',\ \dots,\ y^{(n-1)})=y^{(n)}$$

is called an ordinary differential equation (ODE) of order n; for vector valued functions,

$$y: \mathbb{R} \to \mathbb{R}^m$$,

it is called a system of ordinary differential equations of dimension m.

When a differential equation of order n has the form

$$ F\left(x, y, y', y'',\ \dots,\ y^{(n)}\right) = 0$$

it is called an implicit differential equation whereas the form

$$ F\left(x, y, y', y'',\ \dots,\ y^{(n-1)}\right) = y^{(n)}$$

is called an explicit differential equation.

A differential equation not depending on x is called autonomous.

A differential equation is said to be linear if F can be written as a linear combination of the derivatives of y

$$ y^{(n)} = \sum_{i=0}^{n-1} a_i(x) y^{(i)} + r(x)$$

with ai(x) and r(x) continuous functions in x. The function r(x) is called the source term; if r(x)=0 then the linear differential equation is called homogeneous, otherwise it is called non-homogeneous or inhomogeneous.

[edit] Solutions

Given a differential equation

$$F(x,y,y',\ \dots,\ y^{(n)})=0$$

a function

$$ u: I \subset \mathbb{R} \to \mathbb{R} $$ is called the solution or integral curve for F, if u is n-times differentiable on I, F is defined for all

$$ (x,u,u',\ \dots,\ u^{(n)}) \quad x \in I$$

and

$$ F(x,u,u',\ \dots,\ u^{(n)})=0 \quad x \in I.$$

Given two solutions

$$u: J \subset \mathbb{R} \to \mathbb{R}$$

and

$$v: I \subset \mathbb{R} \to \mathbb{R}$$

u is called an extension of v if I ⊂ J and

$$ u(x) = v(x) \quad x \in I.\,$$

A solution which has no extension is called a global solution.

A general solution of an n-th order equation is a solution containing n arbitrary variables, corresponding to n constants of integration. A particular solution is derived from the general solution by setting the constants to particular values, often chosen to fulfill set 'initial conditions or boundary conditions'. A singular solution is a solution that can't be derived from the general solution.

[edit] Examples

Main article: Examples of differential equations

[edit] Reduction to a first order system

Any differential equation of order n can be written as a system of n first-order differential equations. Given an explicit ordinary differential equation of order n and dimension 1,

$$F\left(x, y, y', y'',\ \dots,\ y^{(n-1)}\right) = y^{(n)}$$

we define a new family of unknown functions

$$ y_n := y^{(n-1)}.\!$$

We can then rewrite the original differential equation as a system of differential equations with order 1 and dimension n.

$$y_1^' = y_2$$

$$\vdots$$

$$y_n^' = F(y_n, \dots, y_1, x).$$

which can be written concisely in vector notation as

$$ \mathbf{y}^'=\mathbf{F}(\mathbf{y}, x)$$

with

$$ \mathbf{y}:=(y,\ldots,y_n).$$

[edit] Linear ordinary differential equations

Main article: Linear differential equation

A well understood particular class of differential equations is linear differential equations. We can always reduce an explicit linear differential equation of any order to a system of differential equation of order 1

$$y_i'(x) = \sum_{j=1}^n a_{i,j}(x) y_j + b_i(x) \, \mathrm{,} \quad i = 1,\ldots,n$$

which we can write concisely using matrix and vector notation as

$$\mathbf{y}^'(x) = \mathbf{A}(x) \mathbf{y}(x) + \mathbf{b}(x)$$

with

$$\mathbf{y}(x):=(y_1(x),\ldots,y_n(x)) \mathbf{b}(x):=(b_1(x),\ldots,b_n(x)) \mathbf{A}(x):=(a_{i,j}(x)) \, \mathrm{,} \quad i,j = 1,\ldots,n.$$

[edit] Homogeneous equations

The set of solutions for a system of homogeneous linear differential equations of order 1 and dimension n

\mathbf{y}^'(x) = \mathbf{A}(x) \mathbf{y}(x)

forms an n-dimensional vector space. Given a basis for this vector space \mathbf{z}_1(x), \ldots, \mathbf{z}_n(x), which is called a fundamental system, every solution \mathbf{s}(x) can be written as

\mathbf{s}(x) = \sum_{i=1}^{n} c_i \mathbf{z}_i(x).

The n × n matrix

\mathbf{Z}(x) := (\mathbf{z}_1(x), \ldots, \mathbf{z}_n(x))

is called fundamental matrix. In general there is no method to explicitly construct a fundamental system, but if one solution is known d'Alembert reduction can be used to reduce the dimension of the differential equation by one.

[edit] Nonhomogeneous equations

The set of solutions for a system of inhomogeneous linear differential equations of order 1 and dimension n

\mathbf{y}^'(x) = \mathbf{A}(x) \mathbf{y}(x) + \mathbf{b}(x)

can be constructed by finding the fundamental system \mathbf{z}_1(x), \ldots, \mathbf{z}_n(x) to the corresponding homogeneous equation and one particular solution \mathbf{p}(x) to the inhomogeneous equation. Every solution \mathbf{s}(x) to nonhomogeneous equation can then be written as

\mathbf{s}(x) = \sum_{i=1}^{n} c_i \mathbf{z}_i(x) + \mathbf{p}(x).

A particular solution to the nonhomogeneous equation can be found by the method of undetermined coefficients or the method of variation of parameters.

[edit] Fundamental systems for homogeneous equations with constant coefficients

If a system of homogeneous linear differential equations has constant coefficients

\mathbf{y}^'(x) = \mathbf{A} \mathbf{y}(x)

then we can explicitly construct a fundamental system. The fundamental system can be written as a matrix differential equation

\mathbf{Y}^' = \mathbf{A} \mathbf{Y}

with solution as a matrix exponential

e^{x \mathbf{A}}

which is a fundamental matrix for the original differential equation. To explicitly calculate this expression we first transform A into Jordan normal form

e^{x \mathbf{A}} = e^{x \mathbf{C}^{-1} \mathbf{J} \mathbf{C}^{1}} = \mathbf{C}^{-1} e^{x \mathbf{J}} \mathbf{C}^{1}

and then evaluate the Jordan blocks

J_i = \begin{bmatrix} \lambda_i & 1 & \; & \; \\ \; & \ddots & \ddots & \; \\ \; & \; & \ddots & 1 \\ \; & \; & \; & \lambda_i \end{bmatrix}

of J separately as

e^{x \mathbf{J_i}} = e^{\lambda_i x} \begin{bmatrix} 1 & x & \frac{x^2}{2} & \dots & \frac{x^{n-1}}{(n-1)!} \\ \; & \ddots & \ddots & \ddots & \vdots \\ \; & \; & \ddots & \ddots & \frac{x^2}{2} \\ \; & \; & \; & \ddots & x \\ \; & \; & \; & \; & 1 \end{bmatrix}.

[edit] Theories of ODEs

[edit] Singular solutions

The theory of singular solutions of ordinary and partial differential equations was a subject of research from the time of Leibniz, but only since the middle of the nineteenth century did it receive special attention. A valuable but little-known work on the subject is that of Houtain (1854). Darboux (starting in 1873) was a leader in the theory, and in the geometric interpretation of these solutions he opened a field which was worked by various writers, notably Casorati and Cayley. To the latter is due (1872) the theory of singular solutions of differential equations of the first order as accepted circa 1900.

[edit] Reduction to quadratures

The primitive attempt in dealing with differential equations had in view a reduction to quadratures. As it had been the hope of eighteenth-century algebraists to find a method for solving the general equation of the nth degree, so it was the hope of analysts to find a general method for integrating any differential equation. Gauss (1799) showed, however, that the differential equation meets its limitations very soon unless complex numbers are introduced. Hence analysts began to substitute the study of functions, thus opening a new and fertile field. Cauchy was the first to appreciate the importance of this view. Thereafter the real question was to be, not whether a solution is possible by means of known functions or their integrals, but whether a given differential equation suffices for the definition of a function of the independent variable or variables, and if so, what are the characteristic properties of this function.

[edit] Fuchsian theory

Two memoirs by Fuchs (Crelle, 1866, 1868), inspired a novel approach, subsequently elaborated by Thomé and Frobenius. Collet was a prominent contributor beginning in 1869, although his method for integrating a non-linear system was communicated to Bertrand in 1868. Clebsch (1873) attacked the theory along lines parallel to those followed in his theory of Abelian integrals. As the latter can be classified according to the properties of the fundamental curve which remains unchanged under a rational transformation, so Clebsch proposed to classify the transcendent functions defined by the differential equations according to the invariant properties of the corresponding surfaces f = 0 under rational one-to-one transformations.

[edit] Lie's theory

From 1870 Lie's work put the theory of differential equations on a more satisfactory foundation. He showed that the integration theories of the older mathematicians can, by the introduction of what are now called Lie groups, be referred to a common source; and that ordinary differential equations which admit the same infinitesimal transformations present comparable difficulties of integration. He also emphasized the subject of transformations of contact.

Sturm-Liouville theory
Sturm-Liouville theory is a general method for resolution of second order linear equations with variable coefficients.

1st ordered ordinary differential equation
1st ordered ordinary differential equation of general form
 * $$A \frac{d}{dt} f(x) + B f(t) = 0$$

Rearrange equation above,
 * $$\frac{d}{dt} f(t) = - s f(t)$$ . With, $$s=\frac{B}{A}$$

Root of equation
 * $$f(t) = Ae^{-st} = Ae^{ - \frac{B}{A} t}$$

2nd ordered ordinary differential equation

 * $$A \frac{d^2}{dx^2} f(x) + B \frac{d}{dx} f(x) + C f(x) = 0$$


 * $$\frac{d^2}{dx^2} f(x) + \frac{B}{A} \frac{d}{dx} f(x) + \frac{C}{A} f(x) = 0$$
 * $$\frac{d^2}{dx^2} f(x) = - \frac{B}{A} \frac{d}{dx} f(x) - \frac{C}{A} f(x) $$
 * $$\frac{d^2}{dx^2} f(x) = - 2 \alpha \frac{d}{dx} f(x) - \beta f(x) $$
 * $$\beta=\frac{C}{A}$$
 * $$\alpha=\beta \gamma= \frac{C}{A} \gamma = \frac{B}{2A}$$
 * $$\gamma=\frac{B}{2C}$$

Roots of differential equations
 * 1 real root . $$f(x)=Ae^{-\alpha t}$$ . $$\alpha=\beta$$
 * 2 real roots . $$f(x)=Ae^{(-\alpha \pm \sqrt{\alpha-\beta}) t}$$ . $$\alpha>\beta$$
 * 2 complex roots . $$f(x)=Ae^{(-\alpha \pm \sqrt{\beta-\alpha}) t}$$ . $$\alpha<\beta$$


 * $$f(x)=Ae^{(-\alpha \pm \sqrt{\beta-\alpha}) t} = A(\alpha)Sin \omega t$$

With
 * $$A(\alpha)=Ae^{-\alpha t}$$
 * $$\omega=\sqrt{\beta-\alpha}$$