152. Differentiation of functions of several variables.

So far we have been concerned exclusively with functions of a single variable x, but there is nothing to prevent us applying the notion of differentiation to functions of several variables x, y, ….

Suppose then that f(x,y) is a function of two1 real variables x and y, and that the limits limh0f(x+h,y)f(x,y)h,limk0f(x,y+k)f(x,y)k exist for all values of x and y in question, that is to say that f(x,y) possesses a derivative df/dx or Dxf(x,y) with respect to x and a derivative df/dy or Dyf(x,y) with respect to y. It is usual to call these derivatives the partial differential coefficients of f, and to denote them by fx,fy or fx(x,y),fy(x,y) or simply fxfy or fxfy. The reader must not suppose, however, that these new notations imply any essential novelty of idea: ‘partial differentiation’ with respect to x is exactly the same process as ordinary differentiation, the only novelty lying in the presence in f of a second variable y independent of x.

In what precedes we have supposed x and y to be two real variables entirely independent of one another. If x and y were connected by a relation the state of affairs would be very different. In this case our definition of fx would fail entirely, as we could not change x into x+h without at the same time changing y. But then f(x,y) would not really be a function of two variables at all. A function of two variables, as we defined it in Ch. II, is essentially a function of two independent variables. If y depends on x, y is a function of x, say y=ϕ(x); and then f(x,y)=f{x,ϕ(x)} is really a function of the single variable x. Of course we may also represent it as a function of the single variable y. Or, as is often most convenient, we may regard x and y as functions of a third variable t, and then f(x,y), which is of the form f{ϕ(t),ψ(t)}, is a function of the single variable t.

Example LX

1. Prove that if x=rcosθ, y=rsinθ, so that r=x2+y2, θ=arctan(y/x), then rx=xx2+y2,ry=yx2+y2,θx=yx2+y2,θy=xx2+y2,xr=cosθ,yr=sinθ,xθ=rsinθ,yθ=rcosθ.

2. Account for the fact that rx1/(xr) and θx1/(xθ). [When we were considering a function y of one variable x it followed from the definitions that dy/dx and dx/dy were reciprocals. This is no longer the case when we are dealing with functions of two variables. Let P (Fig. 46) be the point (x,y) or (r,θ). To find r/x we must increase x, say by an increment MM1=δx, while keeping y constant. This brings P to P1. If along OP1 we take OP=OP, the increment of r is PP1=δr, say; and r/x=lim(δr/δx). If on the other hand we want to calculate x/r, x and y being now regarded as functions of r and θ, we must increase r by Δr, say, keeping θ constant. This brings P to P2, where PP2=Δr: the corresponding increment of x is MM1=Δx, say; and x/r=lim(Δx/Δr). Now Δx=δx:2 but Δrδr. Indeed it is easy to see from the figure that lim(δr/δx)=lim(PP1/PP1)=cosθ, but lim(Δr/Δx)=lim(PP2/PP1)=secθ, so that lim(δr/Δr)=cos2θ.

The fact is of course that x/r and r/x are not formed upon the same hypothesis as to the variation of P.]

3. Prove that if z=f(ax+by) then b(z/x)=a(z/y).

4. Find X/x, X/y, … when X+Y=x, Y=xy. Express xy as functions of XY and find x/X, x/Y, ….

5. Find X/x, … when X+Y+Z=x, Y+Z=xy, Z=xyz; express xyz in terms of XYZ and find x/X, ….

[There is of course no difficulty in extending the ideas of the last section to functions of any number of variables. But the reader must be careful to impress on his mind that the notion of the partial derivative of a function of several variables is only determinate when all the independent variables are specified. Thus if u=x+y+z, xy, and z being the independent variables, then u/x=1. But if we regard u as a function of the variables x, x+y=η, and x+y+z=ζ, so that u=ζ, then u/x=0.]

 

153. Differentiation of a function of two functions.

There is a theorem concerning the differentiation of a function of one variable, known generally as the Theorem of the Total Differential Coefficient, which is of very great importance and depends on the notions explained in the preceding section regarding functions of two variables. This theorem gives us a rule for differentiating f{ϕ(t),ψ(t)}, with respect to t.

Let us suppose, in the first instance, that f(x,y) is a function of the two variables x and y, and that fxfy are continuous functions of both variables (§ 107) for all of their values which come in question. And now let us suppose that the variation of x and y is restricted in that (x,y) lies on a curve x=ϕ(t),y=ψ(t), where ϕ and ψ are functions of t with continuous differential coefficients ϕ(t)ψ(t). Then f(x,y) reduces to a function of the single variable t, say F(t). The problem is to determine F(t).

Suppose that, when t changes to t+τ, x and y change to x+ξ and y+η. Then by definition dF(t)dt=limτ01τ[f{ϕ(t+τ),ψ(t+τ)}f{ϕ(t),ψ(t)}]=lim1τ{f(x+ξ,y+η)f(x,y)}=lim[f(x+ξ,y+η)f(x,y+η)ξξτ+f(x,y+η)f(x,y)ηητ].

But, by the Mean Value Theorem, {f(x+ξ,y+η)f(x,y+η)}/ξ=fx(x+θξ,y+η),{f(x,y+η)f(x,y)}/η=fy(x,y+θη), where θ and θ each lie between 0 and 1. As τ0, ξ0 and η0, and ξ/τϕ(t), η/τψ(t): also fx(x+θξ,y+η)fx(x,y),fy(x,y+θη)fy(x,y). Hence F(t)=Dtf{ϕ(t),ψ(t)}=fx(x,y)ϕ(t)+fy(x,y)ψ(t), where we are to put x=ϕ(t), y=ψ(t) after carrying out the differentiations with respect to x and y. This result may also be expressed in the form dfdt=fxdxdt+fydydt.

Example LXI

1. Suppose ϕ(t)=(1t2)/(1+t2), ψ(t)=2t/(1+t2), so that the locus of (x,y) is the circle x2+y2=1. Then ϕ(t)=4t/(1+t2)2,ψ(t)=2(1t2)/(1+t2)2,F(t)={4t/(1+t2)2}fx+{2(1t2)/(1+t2)2}fy, where x and y are to be put equal to (1t2)/(1+t2) and 2t/(1+t2) after carrying out the differentiations.

We can easily verify this formula in particular cases. Suppose, e.g., that f(x,y)=x2+y2. Then fx=2x, fy=2y, and it is easily verified that F(t)=2xϕ(t)+2yψ(t)=0, which is obviously correct, since F(t)=1.

2. Verify the theorem in the same way when (a) x=tm, y=1tm, f(x,y)=x+y; (b) x=acost, y=asint, f(x,y)=x2+y2.

3. One of the most important cases is that in which t is x itself. We then obtain Dxf{x,ψ(x)}=Dxf(x,y)+Dyf(x,y)ψ(x). where y is to be replaced by ψ(x) after differentiation.

It was this case which led to the introduction of the notation f/x, f/y. For it would seem natural to use the notation df/dx for either of the functions Dxf{x,ψ(x)} and Dxf(x,y), in one of which y is put equal to ψ(x) before and in the other after differentiation. Suppose for example that y=1x and f(x,y)=x+y. Then Dxf(x,1x)=Dx1=0, but Dxf(x,y)=1.

The distinction between the two functions is adequately shown by denoting the first by df/dx and the second by f/x, in which case the theorem takes the form dfdx=fx+fydydx; though this notation is also open to objection, in that it is a little misleading to denote the functions f{x,ψ(x)} and f(x,y), whose forms as functions of x are quite different from one another, by the same letter f in df/dx and f/x.

4. If the result of eliminating t between x=ϕ(t), y=ψ(t) is f(x,y)=0, then fxdxdt+fydydt=0.

5. If x and y are functions of t, and r and θ are the polar coordinates of (x,y), then r=(xx+yy)/r, θ=(xyyx)/r2, dashes denoting differentiations with respect to t.

 

154. The Mean Value Theorem for functions of two variables.

Many of the results of the last chapter depended upon the Mean Value Theorem, expressed by the equation ϕ(x+h)ϕ(x)=hf(x+θh), or as it may be written, if y=ϕ(x), δy=f(x+θδx)δx.

Now suppose that z=f(x,y) is a function of the two independent variables x and y, and that x and y receive increments hk or δxδy respectively: and let us attempt to express the corresponding increment of z, viz. δz=f(x+h,y+k)f(x,y), in terms of hk and the derivatives of z with respect to x and y.

Let f(x+ht,y+kt)=F(t). Then f(x+h,y+k)f(x,y)=F(1)F(0)=F(θ), where 0<θ<1. But, by § 153, F(t)=Dtf(x+ht,y+kt)=hfx(x+ht,y+kt)+kfy(x+ht,y+kt). Hence finally δz=f(x+h,y+k)f(x,y)=hfx(x+θh,y+θk)+kfy(x+θh,y+θk), which is the formula desired. Since fxfy are supposed to be continuous functions of x and y, we have fx(x+θh,y+θk)=fx(x,y)+ϵh,k,fy(x+θh,y+θk)=fy(x,y)+ηh,k, where ϵh,k and ηh,k tend to zero as h and k tend to zero. Hence the theorem may be written in the form (1)δz=(fx+ϵ)δx+(fy+η)δy, where ϵ and η are small when δx and δy are small.

The result embodied in (1) may be expressed by saying that the equation δz=fxδx+fyδy is approximately true; i.e. that the difference between the two sides of the equation is small in comparison with the larger of δx and δy.3 We must say ‘the larger of δx and δy’ because one of them might be small in comparison with the other; we might indeed have δx=0 or δy=0.

It should be observed that if any equation of the form δz=λδx+μδy is ‘approximately true’ in this sense, we must have λ=fx, μ=fy. For we have δzfxδxfyδy=ϵδx+ηδy,δzλδxμδy=ϵδx+ηδy where ϵη, ϵη all tend to zero as δx and δy tend to zero; and so (λfx)δx+(μfy)δy=ρδx+ρδy where ρ and ρ tend to zero. Hence, if ζ is any assigned positive number, we can choose σ so that |(λfx)δx+(μfy)δy|<ζ(|δx|+|δy|) for all values of δx and δy numerically less than σ. Taking δy=0 we obtain |(λfx)δx|<ζ|δx|, or |λfx|<ζ, and, as ζ may be as small as we please, this can only be the case if λ=fx. Similarly μ=fy.


  1. The new points which arise when we consider functions of several variables are illustrated sufficiently when there are two variables only. The generalisations of our theorems for three or more variables are in general of an obvious character.↩︎
  2. Of course the fact that Δx=δx is due merely to the particular value of Δr that we have chosen (viz. PP2). Any other choice would give us values of ΔxΔr proportional to those used here.↩︎
  3. Or with |δx|+|δy| or δx2+δy2.↩︎

151. The contact of plane curves Main Page 155. Differentials