Genesis of an Ordinary Differential Equation

Consider an equation

$ f\left(x, y, c_{1}, c_{2}, \ldots, c_{n}\right)=0\tag{A} $

in which $x$ and $y$ are variables and $c_{1}, c_{2}, \ldots, c_{n}$ are arbitrary and independent constants. This equation serves to determine $y$ as a function of $x$; strictly speaking, an $n$-fold infinity of functions is so determined, each function corresponding to a particular set of values attributed to $c_{1}, c_{2}, \ldots c_{n}$. Now an ordinary differential equation can be formed which is satisfied by every one of these functions, as follows.

Let the given equation be differentiated $n$ times in succession, with respect to $x$, then $n$ new equations are obtained, namely,

$ \frac{\partial f}{\partial x}+\frac{\partial f}{\partial y} y^{\prime}=0 $$ \frac{\partial^2 f}{\partial x^2}+2 \frac{\partial^2 f}{\partial x \partial y}y^{\prime}+\frac{\partial^2 f}{\partial y^2} y^{\prime 2}+\frac{\partial f}{\partial y} y^{\prime \prime}=0 $$ \vdots $$ \frac{\partial^n f}{\partial x^n}+\cdots+\frac{\partial f}{\partial y} y^{(n)}=0 $

where

$ y^{\prime}=\frac{d y}{d x}, y^{\prime \prime}=\frac{d^2 y}{d x^2}, \ldots, y^{(n)}=\frac{d^n y}{d x^n}. $

Each equation is manifestly distinct from those which precede it; 1 from the aggregate of $n+1$ equations the $n$ arbitrary constants $c_{1}, c_{2}, \ldots, c_{n}$ can be eliminated by algebraical processes, and the eliminant is the differential equation of order $n$ :

$ F\left(x, y, y^{\prime}, y^{\prime \prime}, \ldots, y^{(n)}\right)=0 $

It is clear from the very manner in which this differential equation was formed that it is satisfied by every function $y=\phi(x)$ defined by the relation (A). This relation is termed the primitive of the differential equation, and every function $y=\phi(x)$ which satisfies the differential equation is known as a solution. 2 A solution which involves a number of essentially distinct arbitrary constants equal to the order of the equation is known as the general solution.3 That this terminology is justified, will be seen when in Chapter III. it is proved that one solution of an equation of order $n$ and one only can always be found to satisfy, for a specified value of $x, n$ distinct conditions of a particular type. The possibility of satisfying these $n$ conditions depends upon the existence of a solution containing $n$ arbitrary constants. The general solution is thus essentially the same as the primitive of the differential equation.

It has been assumed that the primitive actually contains $n$ distinct constants $c_{1}, c_{2}, \ldots, c_{n}$. If there are only apparently $n$ constants, that is to say if two or more constants can be replaced by a single constant without essentially modifying the primitive, then the order of the resulting differential equation will be less than $n$. For instance, suppose that the primitive is given in the form

$ f\{x, y, \phi(a, b)\}=0, $

then it apparently depends upon two constants $a$ and $b$, but in reality upon one constant only, namely $c=\phi(a, b)$. In this case the resulting differential equation is of the first and not of the second order.

Again, if the primitive is reducible, that is to say if $f\left(x, y, c_{1}, \ldots, c_{n}\right)$ breaks up into two factors, each of which contains $y$, the order of the resulting differential equation may be less than $n$. For if neither factor contains all the $n$ constants, then each factor will give rise to a differential equation of order less than $n$, and it may occur that these two differential equations are identical, or that one of them admits of all the solutions of the other, and therefore is satisfied by the primitive itself. Thus let the primitive be

$ y^{2}-(a+b) x y+a b x^{2}=0 ; $

it is reducible and equivalent to the two equations

$ y-a x=0, \quad y-b x=0, $

each of which, and therefore the primitive itself, satisfies the differential equation

$ y-x y^{\prime}=0. $

0.1 The Differential Equation of a Family of Confocal Conies

Consider the equation

$ \frac{x^{2}}{a^{2}+\lambda}+\frac{y^{2}}{b^{2}+\lambda}=1, $

where $a$ and $b$ are definite constants, and $\lambda$ an arbitrary parameter which can assume all real values. This equation represents a family of confocal conics. The differential equation of which it is the primitive is obtained by eliminating $\lambda$ between it and the derived equation

$ \frac{2 x}{a^{2}+\lambda}+\frac{2 y y^{\prime}}{b^{2}+\lambda}=0. $

From the primitive and the derived equation it is found that

$a^{2}+\lambda=\frac{x^{2} y^{\prime}-x y}{y^{\prime}}, \quad b^{2}+\lambda=y^{2}-x y y^{\prime},$

and, eliminating $\lambda$,

$a^{2}-b^{2}=\frac{x^{2} y^{\prime}-x y}{y^{\prime}}-y^{2}+x y y^{\prime},$

and therefore the required differential equation is

$x y y^{\prime 2}+\left(x^{2}-y^{2}-a^{2}+b^{2}\right) y^{\prime}-x y=0 ;$

it is of the first order and the second degree.

When an equation is of the first order it is customary to represent the derivative $y^{\prime}$ by the symbol $p$. Thus the differential equation of the family of confocal conics may be written:

$x y\left(p^{2}-1\right)+\left(x^{2}-y^{2}-a^{2}+b^{2}\right) p=0.$

1. Formation of Partial Differential Equations through the Elimination of Arbitrary Constants

Let $x_{1}, x_{2}, \ldots, x_{m}$ be independent variables, and let $z$, the dependent variable, be defined by the equation

$f\left(x_{1}, x_{2}, \ldots, x_{m} ; z ; c_{1}, c_{2}, \ldots, c_{n}\right)=0,$

where $c_{1}, c_{2}, \ldots, c_{n}$ are $n$ arbitrary constants. To this equation may be adjoined the $m$ equations obtained by differentiating partially with respect to each of the variables $x_{1}$, $x_{2}, \ldots, x_{m}$ in succession, namely,

$ \begin{aligned} \frac{\partial f}{\partial x_{1}}+\frac{\partial f}{\partial z} \cdot \frac{\partial z}{\partial x_{1}}&=0,\\ \vdots\qquad &\\ \frac{\partial f}{\partial x_{m}}+\frac{\partial f}{\partial z} \cdot \frac{\partial z}{\partial x_{m}}&=0 \end{aligned} $

If $m \geqslant n$, sufficient equations are now available to eliminate the constants $c_{1}, c_{2}, \ldots, c_{n}$. If $m $ \frac{\partial^2 f}{\partial x_r^2}+2 \frac{\partial^2 f}{\partial x_r \partial z} \cdot \frac{\partial z}{\partial x_r}+\frac{\partial^2 f}{\partial z^2} \cdot\left(\frac{\partial z}{\partial x_r}\right)^2+\frac{\partial f}{\partial z} \cdot \frac{\partial^2 z}{\partial x_r^2}=0 \quad(r=1,2, \ldots, m) $$ \begin{aligned} \frac{\partial^2 f}{\partial x_r \partial x_s}+\frac{\partial^2 f}{\partial x_r \partial z} \cdot \frac{\partial z}{\partial x_s}+\frac{\partial^2 f}{\partial x_s \partial z} \cdot \frac{\partial z}{\partial x_r}+\frac{\partial^2 f}{\partial z^2} \cdot \frac{\partial z}{\partial x_r} \cdot &\frac{\partial z}{\partial x_s}+\frac{\partial f}{\partial z} \cdot \frac{\partial^2 z}{\partial x_r \partial x_s}=0 \\ &(r, s=1,2, \ldots, m ; r \neq s) . \end{aligned} $

This process is continued until enough equations have been obtained to enable the elimination to be carried out. In general, when this stage has been reached, there will be more equations available than there are constants to eliminate and therefore the primitive may lead not to one partial differential equation but to a system of simultaneous partial differential equations.

1.1 The Partial Differential Equations of all Planes and of all Spheres.

As a first example let the primitive be the equation

$z=a x+b y+c,$

in which $a, b, c$ are arbitrary constants. By a proper choice of these constants, the equation can be made to represent any plane in space except a plane parallel to the $z$ axis. The first derived equations are:

$\frac{\partial z}{\partial x}=a, \quad \frac{\partial z}{\partial y}=b.$

These are not sufficient to eliminate $a, b$, and $c$, and therefore the second derived equations are taken, namely,

$\frac{\partial^{2} z}{\partial x^{2}}=0, \quad \frac{\partial^{2} z}{\partial x \partial y}=0, \quad \frac{\partial^{2} z}{\partial y^{2}}=0.$

They are free of arbitrary constants, and are therefore the differential equations required. It is customary to write

$p=\frac{\partial z}{\partial x}, \quad q=\frac{\partial z}{\partial y}, \quad r=\frac{\partial^{2} z}{\partial x^{2}}, \quad s=\frac{\partial^{2} z}{\partial x \partial y}, \quad t=\frac{\partial^{2} z}{\partial y^{2}}.$

Thus any plane in space which is not parallel to the $z$-axis satisfies simultaneously the three equations

$r=0, \quad s=0, \quad t=0.$

In the second place, consider the equation satisfied by the most general sphere; it is

$(x-a)^{2}+(y-b)^{2}+(z-c)^{2}=r^{2},$

where $a, b, c$ and $r$ are arbitrary constants. The first derived equations are

$(x-a)+(z-c) p=0, \quad(y-b)+(z-c) q=0,$

and the second derived equations are

$ \begin{aligned} 1+p^{2}+(z-c) r & =0 \\ p q+(z-c) s & =0 \\ 1+q^{2}+(z-c) t & =0 \end{aligned} $

When $z-c$ is eliminated, the required equations are obtained, namely,

$\frac{1+p^{2}}{r}=\frac{p q}{s}=\frac{1+q^{2}}{t}.$

Thus there are two distinct equations. Let $\lambda$ be the value of each of the members of the equations, then

$\lambda^{2}\left(r t-s^{2}\right)=1+p^{2}+q^{2}>0.$

Consequently, if the spheres considered are real, the additional condition

$r t>s^{2}$

must be satisfied.

2. A Property of Jacobians

It will now be shown that the natural primitive of a single partial differential equation is a relation into which enter arbitrary functions of the variables. The investigation which leads up to this result depends upon a property of functional determinants or Jacobians.

Let $u_{1}, u_{2}, \ldots, u_{m}$ be functions of the independent variables $x_{1}, x_{2}, \ldots, x_{n}$, and consider the set of partial differential coefficients arranged in order thus:

$ \begin{aligned} &\frac{\partial u_1}{\partial x_1}, &&\frac{\partial u_1}{\partial x_2}, &&&\ldots, &&&&\frac{\partial u_1}{\partial x_n} \\ &\frac{\partial u_2}{\partial x_1}, &&\frac{\partial u_2}{\partial x_2}, &&& \ldots, &&&&\frac{\partial u_2}{\partial x_n} \\ &\quad\vdots &&\quad \vdots &&&\vdots\quad &&&&\vdots\quad\\ &\frac{\partial u_m}{\partial x_1}, && \frac{\partial u_m}{\partial x_2}, &&& \ldots, &&&& \frac{\partial u_m}{\partial x_n} \end{aligned} $

Then the determinant of order $p$ whose elements are the elements common to $p$ rows and $p$ columns of the above scheme is known as a Jacobian.4 Let all the different possible Jacobians be constructed, then if a Jacobian of order $p$, say

$ \begin{vmatrix} \dfrac{\partial u_1}{\partial x_1} & \cdots &\dfrac{\partial u_1} {\partial x_p}\\ \vdots & \ddots &\vdots\\ \dfrac{\partial u_p}{\partial x_1} & \cdots &\dfrac{\partial u_p} {\partial x_p} \end{vmatrix} $

is not zero for a chosen set of values $x_{1}=\xi_{1}, \ldots, x_{n}=\xi_{n}$, but if every Jacobian of order $p+1$ is identically zero, then the functions $u_{1}, u_{2}, \ldots, u_{p}$ are independent, but the remaining functions $u_{p+1}, \ldots, u_{m}$ are expressible in terms of $u_{1}, \ldots, u_{p}$.

Suppose that, for values of $x_{1}, \ldots, x_{n}$ in the neighbourhood of $\xi_{1}, \ldots, \xi_{n}$, the functions $u_{1}$, ..., $u_{p}$ are not independent, but that there exists an identical relationship,

$\phi\left(u_{1}, \ldots, u_{p}\right)=0.$

Then the equations

$ \begin{aligned} \frac{\partial \phi}{\partial u_1} \cdot \frac{\partial u_1}{\partial x_1}&+\cdots+\frac{\partial \phi}{\partial u_p} \cdot \frac{\partial u_p}{\partial x_1}=0 \\ &\qquad\vdots\\ \frac{\partial \phi}{\partial u_1} \cdot \frac{\partial u_1}{\partial x_p}&+\cdots+\frac{\partial \phi}{\partial u_p} \cdot \frac{\partial u_p}{\partial x_p}=0 \end{aligned} $

identically in the neighbourhood of $\xi_{1}, \ldots, \xi_{n}$, which is contrary to the hypothesis. Consequently, the first part of the theorem, namely, that $u_{1}, \ldots, u_{p}$ are independent, is true.

In $u_{p+1}, \ldots, u_{m}$ let the variables $x_{1}, \ldots, x_{p}, x_{p+1}, \ldots, x_{n}$ be replaced by the new set of independent variables $u_{1}, \ldots, u_{p}, x_{p+1}, \ldots, x_{n}$. It will now be shown that if $u_{r}$ is any of the functions $u_{p+1}, \ldots, u_{m}$, and $x_{s}$ any one of the variables $x_{p+1}, \ldots, x_{n}$, then $u_{r}$ is explicitly independent of $x_{s}$ that is

$\frac{\partial u_{r}}{\partial x_{s}}=0.$

Let

$u_{1}=f_{1}\left(x_{1}, \ldots, x_{n}\right), \ldots, u_{m}=f_{m}\left(x_{1}, \ldots, x_{n}\right),$

and let $x_{1}, \ldots, x_{p}$ be replaced by their expressions in terms of the new independent variables $u_{1}, \ldots, u_{p}, x_{p+1}, \ldots, x_{n}$, then differentiating both sides of each equation with respect to $x_{s}$,

$ \begin{aligned} 0&=\frac{\partial f_{1}}{\partial x_{1}} \cdot \frac{\partial x_{1}}{\partial x_{s}}+\ldots+\frac{\partial f_{1}}{\partial x_{p}} \cdot \frac{\partial x_{p}}{\partial x_{s}}+\frac{\partial f_{1}}{\partial x_{s}}, \\ & ~~\vdots\\ 0&=\frac{\partial f_{p}}{\partial x_{1}} \cdot \frac{\partial x_{1}}{\partial x_{s}}+\cdots+\frac{\partial f_{p}}{\partial x_{p}} \cdot \frac{\partial x_{p}}{\partial x_{s}}+\frac{\partial f_{p}}{\partial x_{s}} \\ \frac{\partial u_{r}}{\partial x_{s}}&= \frac{\partial f_{r}}{\partial x_{1}} \cdot \frac{\partial x_{1}}{\partial x_{s}}+\cdots+\frac{\partial f_{r}}{\partial x_{p}} \cdot \frac{\partial x_{p}}{\partial x_{s}}+\frac{\partial f_{r}}{\partial x_{s}}, \\ &(r=p+1, \ldots, m) . \end{aligned} $

The eliminant of $\dfrac{\partial x_1}{\partial x_s}$, $\ldots$, $\dfrac{\partial x_p}{\partial x_s}$ is

$ \begin{vmatrix} \dfrac{\partial f_1}{\partial x_1}& \ldots & \dfrac{\partial f_1}{\partial x_p}& \dfrac{\partial f_1}{\partial x_s} \\ \vdots & \vdots & \vdots & \vdots \\ \dfrac{\partial f_p}{\partial x_1}& \ldots& \dfrac{\partial f_r}{\partial x_p}& \dfrac{\partial f_p}{\partial x_s} \\ \dfrac{\partial f_1}{\partial x_r}& \dots & \dfrac{\partial f_r}{\partial x_p} &\dfrac{\partial u_r}{\partial x_s}-\dfrac{\partial u_r}{\partial x_s} \end{vmatrix} =0 $

But since, by hypothesis,

$\frac{\partial\left(f_{1}, \ldots, f_{p}, f_{r}\right)}{\partial\left(x_{1}, \ldots, x_{p}, x_{s}\right)}=0, \quad \frac{\partial\left(f_{1}, \ldots, f_{p}\right)}{\partial\left(x_{1}, \ldots, x_{p}\right)} \neq 0 $

it follows that

$\frac{\partial u_{r}}{\partial x_{s}}=0 \quad(r=p+1, \ldots, m ; s=p+1, \ldots, n).$

Consequently each of the functions $u_{p+1}, \ldots, u_{m}$ is expressible in terms of the functions $u_{1}, \ldots, u_{p}$ alone, as was to be proved.

3. Formation of a Partial Differential Equation through the Elimination of an Arbitrary Function

Let the dependent variable $z$ be related to the independent variables $x_{1}, \ldots, x_{n}$ by an equation of the form

$F\left(u_{1}, u_{2}, \dots, u_n\right)=0,$

where $F$ is an arbitrary function of its arguments $u_{1}, u_{2}, \ldots, u_{n}$ which, in turn, are given functions of $x_{1}, \ldots, x_{n}$ and $z$. When for $z$ is substituted its value in terms of $x_{1}, \ldots, x_{n}$, the equation becomes an identity. If therefore $D_{r} u_{s}$ represents the partial derivative of $u_{s}$ with respect to $x_{r}$ when $z$ has been replaced by its value, then

$ \begin{vmatrix} D_1 u_1 & \dots & D_n u_1\\ \vdots & \vdots & \vdots\\ D_1 u_n & \dots & D_nu_n \end{vmatrix}=0 $

But

$D_{r} u_{s}=\frac{\partial u_{s}}{\partial x_{r}}+\frac{\partial u_{s}}{\partial z} \cdot \frac{\partial z}{\partial x_{r}},$

and therefore the partial differential equation satisfied by $z$ is

$ \begin{vmatrix} \dfrac{\partial u_{1}}{\partial x_{1}}+\dfrac{\partial u_{1}}{\partial z} \cdot \dfrac{\partial z}{\partial x_{1}}& \cdots & \dfrac{\partial u_{1}}{\partial x_{n}}+\dfrac{\partial u_{1}}{\partial z} \cdot \dfrac{\partial z}{\partial x_{n}} \\ \vdots & \vdots & \vdots \\ \dfrac{\partial u_{n}}{\partial x_{1}}+\dfrac{\partial u_{n}}{\partial z} \cdot \dfrac{\partial z}{\partial x_{1}} & \vdots & \dfrac{\partial u_{n}}{\partial x_{n}}+\dfrac{\partial u_{n}}{\partial z} \cdot \dfrac{\partial z}{\partial x_{n}} \end{vmatrix}=0. $

3.1. The Differential Equation of a Surface of Revolution

The equation

$F\left(z, x^{2}+y^{2}\right)=0$

represents a surface of revolution whose axis coincides with the $z$-axis. In the notation of the preceding section,

$x_{1}=x, \quad x_{2}=y, \quad u_{1}=z, \quad u_{2}=x^{2}+y^{2},$

and therefore $z$ satisfies the partial differential equation:

$ \begin{vmatrix} \dfrac{\partial z}{\partial x} & \dfrac{\partial z}{\partial y}\\ 2x & 2y \end{vmatrix}=0 $

or

$y \frac{\partial z}{\partial x}-x \frac{\partial z}{\partial y}=0.$

Conversely, this equation is satisfied by

$z=\phi\left(x^{2}+y^{2}\right),$

where $\phi$ is an arbitrary function of its argument, and is therefore the differential equation of all surfaces of revolution which have the common axis $x=0, y=0$.

3.2. Euler's Theorem on Homogeneous Functions

Let

$ z=\phi(x, y), $

where $\phi(x, y)$ is a homogeneous function of $x$ and $y$ of degree $n$. Then, since $\phi(x, y)$ can be written in the form

$x^{n} \psi\left(\frac{y}{x}\right),$

it follows that

$ x^{-n}{z}=\psi\left(\frac{y}{x}\right). $

In the notation of 2.3,

$ x_{1}=x, \quad x_{2}=y, \quad u_{1}=x^{-n}z, \quad u_{2}=\psi\left(\frac{y}{x}\right), $$ F\left(u_{1}, u_{2}\right)=u_{1}-u_{2} $

and therefore $z$ satisfies the partial differential equation:

$ \begin{vmatrix} -n x^{-n-1}z+x^{-n}\dfrac{\partial z}{\partial x} & x^{-n}\dfrac{\partial z}{\partial y}\\ -y x^{-2}\psi^\prime & x^{-1}\psi^\prime \end{vmatrix}=0 $

and this equation reduces to

$x \frac{\partial z}{\partial x}+y \frac{\partial z}{\partial y}=n z.$

Similarly, if $u$ is a homogeneous function of the three variables $x, y$ and $z$, of degree $n$,

$x \frac{\partial u}{\partial x}+y \frac{\partial u}{\partial y}+z \frac{\partial u}{\partial z}=n u.$

This theorem can be extended to any number of variables.

4. Formation of a Total Differential Equation in Three Variables

The equation

$\phi(x, y, z)=c$

represents a family of surfaces, and it will be supposed that to each value of $c$ corresponds one, and only one, surface of the family. Now let $(x, y, z)$ be a point on a particular surface and $(x+\delta x, y+\delta y, z+\delta z)$ a neighbouring point on the same surface, then

$\phi(x+\delta x, y+\delta y, z+\delta z)-\phi(x, y, z)=0.$

Assuming that the partial derivatives

$\frac{\partial \phi}{\partial x}, \frac{\partial \phi}{\partial y}, \frac{\partial \phi}{\partial z}$

exist and are continuous, this equation may be written in the form

$\left\{\frac{\partial \phi(x, y, z)}{\partial x}+\epsilon_{1}\right\} \delta x+\left\{\frac{\partial \phi(x, y, z)}{\partial y}+\epsilon_{2}\right\} \delta y+\left\{\frac{\partial \phi(x, y, z)}{\partial z}+\epsilon_{3}\right\} \delta z=0,$

where $\epsilon_{1}, \epsilon_{2}, \epsilon_{3} \rightarrow 0$, as $\delta x, \delta y, \delta z \rightarrow 0$.

Now let $\epsilon_{1}, \epsilon_{2}$ and $\epsilon_{3}$ be made zero and let $d x$, $d y$ and $d z$ be written for $\delta x, \delta y$ and $\delta z$ respectively. Then there results the total differential equation

$ \frac{\partial \phi}{\partial x} d x+\frac{\partial \phi}{\partial y} d y+\frac{\partial \phi}{\partial z} d z=0, $

which has been derived from the primitive by a consistent and logical process.
If the three partial derivatives have a common factor $\mu$, and if

$ \frac{\partial \phi}{\partial x}=\mu P, \quad \frac{\partial \phi}{\partial y}=\mu Q, \quad \frac{\partial \phi}{\partial z}=\mu R, $

then if the factor $\mu$ is removed, the equation takes the form

$P d x+Q d y+R d z=0.$

That there is no inconsistency in the above use of the differentials $d x$, etc., may be verified by considering a particular equation in two variables, namely,

$y-f(x)=c.$

The above process gives rise to the total differential equation

$d y-f(x) d x=0,$

and thus the quotient of the differentials $d y, d x$ is in fact the differential coefficient $d y / d x$.

The primitive $\frac{(x+z)(y+z)}{x+y}=c$ gives rise to the total differential equation $\frac{y^{2}-z^{2}}{(x+y)^{2}} d x+\frac{x^{2}-z^{2}}{(x+y)^{2}} d y+\frac{2 z+x+y}{x+y} d z=0,$ which, after multiplication by $(x+y)^{2}$, becomes $\left(y^{2}-z^{2}\right) d x+\left(x^{2}-z^{2}\right) d y+(2 z+x+y)(x+y) d z=0.$

Footnotes

  1. Needless to say, it is assumed that all the partial differential coefficients of $f$ exist, and that $\dfrac{\partial f}{\partial y}$ is not identically zero.

  2. Originally the terms integral (James Bernoulli, 1689) and particular integral (Euler, Inst. Calc. Int. 1768) were used. The use of the word solution dates back to Lagrange (1774), and, mainly through the influence of Poincaré, it has become established. The term particular integral is now used only in a very restricted sense, cf. Chap. VI. infra.

  3. Formerly known as the complete integral or complete integral equation (œquatio integralis completa, Euler). The term integral equation has now an utterly different meaning (cf. $\S$ 3.2, infra), and its use in any other connection should be abandoned.

  4. Scott and Mathews, Theory of Determinants, Chap. XIII.