- The derivative gives the slope of a tangent to the graph of the function at a point. In this way, derivatives can be used to determine many geometrical properties of the graph, such as concavity or convexity.
- The derivative provides a mathematical formulation of rate of change; it measures the rate at which the function's value changes as the function's argument changes.
This derivative is the kind usually encountered in a first course on calculus, and historically was the first to be discovered. However, there are also many generalizations of the derivative.
The remainder of this article discusses only the simplest case (real-valued functions of real numbers).
- 1 Differentiation and differentiability
- 2 Newton's difference quotient
- 3 Notations for differentiation
- 4 Critical points
- 5 Physics
- 6 Algebraic manipulation
- 7 Using derivatives to graph functions
- 8 Generalizations
- 9 See also
- 10 External links
- 11 References
Differentiation and differentiability
In physical terms, differentiation expresses the rate at which a quantity, y, changes with respect to the change in another quantity, x, on which it has a functional relationship. Using the symbol Δ to refer to change in a quantity, this rate is defined as a limit of difference quotients
as Δx approaches 0. In Leibniz's notation for derivatives, the derivative of y with respect to x is written
suggesting the ratio of two infinitesimal quantities. The above expression is pronounced in various ways such as "dy by dx" or "dy over dx". The form "dy dx" is also used conversationally, although it may be confused with the notation for element of area.
Modern mathematicians do not bother with "dependent quantities", but simply state that differentiation is a mathematical operation on functions. The precise definition of this operation (which therefore need not deal with infinitesimal quantities) is given as:
A function is differentiable at a point x if its derivative exists at that point; a function is differentiable on an interval if it is differentiable at every x within the interval. If a function is not continuous at x, then there is no tangent line and the function is therefore not differentiable at x; however, even if a function is continuous at x, it may not be differentiable there. In other words, differentiability implies continuity, but not vice versa. One famous example of a function that is continuous everywhere but differentiable nowhere is the Weierstrass function.
The derivative of a differentiable function can itself be differentiable. The derivative of a derivative is called a second derivative. Similarly, the derivative of a second derivative is a third derivative, and so on.
Newton's difference quotient
The derivative of a function f at x is geometrically the slope of the tangent line to the graph of f at x. Without the concept which we are about to define, it is impossible to directly find the slope of the tangent line to a given function, because we only know one point on the tangent line, namely (x, f(x)). Instead, we will approximate the tangent line with multiple secant lines that have progressively shorter distances between the two intersecting points. When we take the limit of the slopes of the nearby secant lines in this progression, we will get the slope of the tangent line. The derivative is then defined by taking the limit of the slope of secant lines as they approach the tangent line.
To find the slopes of the nearby secant lines, choose a small number h. h represents a small change in x, and it can be either positive or negative. The slope of the line through the points (x,f(x)) and (x+h,f(x+h)) is
This expression is Newton's difference quotient. The derivative of f at x is the limit of the value of the difference quotient as the secant lines get closer and closer to being a tangent line:
If the derivative of f exists at every point x in the domain, we can define the derivative of f to be the function whose value at a point x is the derivative of f at x.
Since immediately substituting 0 for h results in division by zero, calculating the derivative directly can be unintuitive. One technique is to simplify the numerator so that the h in the denominator can be cancelled. This happens easily for polynomials; see calculus with polynomials. For almost all functions however, the result is a mess. Fortunately, many guidelines exist.
Notations for differentiation
|for the first derivative,|
|for the second derivative,|
|for the third derivative, and|
|for the nth derivative, provided n > 3|
We can write the derivative of f at the point a in two different ways:
If the output of f(x) is another variable, for example, if y=f(x), we can write the derivative as:
Higher derivatives are expressed as
for the n-th derivative of f(x) or y respectively. Historically, this came from the fact that, for example, the 3rd derivative is:
which we can loosely write as:
Dropping brackets gives the notation above.
Leibniz's notation allows one to specify the variable for differentiation (in the denominator). This is especially relevant for partial differentiation. It also makes the chain rule easy to remember, because the "du" terms appear symbolically to cancel:
(In the popular formulation of calculus in terms of limits, the "du" terms cannot literally cancel, because on their own they are undefined; they are only defined when used together to express a derivative. In nonstandard analysis, however, they can be viewed as infinitesimal numbers that cancel.)
Newton's notation for differentiation (also called the dot notation for differentiation) requires placing a dot over the function name:
and so on.
|for the first derivative,|
|for the second derivative, and|
|for the nth derivative, provided n > 1|
This notation can also be abbreviated when taking derivatives of expressions that contain a single variable. The subscript to the operator is dropped and is assumed to be the only variable present in the expression. In the following examples, u represents any expression of a single variable:
|for the first derivative,|
|for the second derivative, and|
|for the nth derivative, provided n > 1|
Euler's notation is useful for stating and solving linear differential equations.
Points on the graph of a function where the derivative is undefined or equals zero are called critical points or sometimes stationary points (in the case where the derivative equals zero). If the second derivative is positive at a critical point, that point is a local minimum; if negative, it is a local maximum; if zero, it may or may not be a local minimum or local maximum. Taking derivatives and solving for critical points is often a simple way to find local minima or maxima, which can be useful in optimization. In fact, local minima and maxima can only occur at critical points. This is related to the extreme value theorem.
Arguably the most important application of calculus to physics is the concept of the "time derivative"—the rate of change over time—which is required for the precise definition of several important concepts. In particular, the time derivatives of an object's position are significant in Newtonian physics:
- Velocity (instantaneous velocity; the concept of average velocity predates calculus) is the derivative (with respect to time) of an object's position.
- Acceleration is the derivative (with respect to time) of an object's velocity.
- Jerk is the derivative (with respect to time) of an object's acceleration.
For example, if an object's position ; then, the object's velocity is ; the object's acceleration is ; and the object's jerk is
Messy limit calculations can be avoided, in certain cases, because of differentiation rules which allow one to find derivatives via algebraic manipulation; rather than by direct application of Newton's difference quotient. One should not infer that the definition of derivatives, in terms of limits, is unnecessary. Rather, that definition is the means of proving the following "powerful differentiation rules"; these rules are derived from the difference quotient.
- Constant rule: The derivative of any constant is zero.
- Linearity: (af + bg)' = af ' + bg' for all functions f and g and all real numbers a and b.
- General power rule (Polynomial rule): If , for some real number r;
- Product rule: for all functions f and g.
- Quotient rule: unless g is zero.
- Chain rule: If , then .
- Inverse functions and differentiation: If , , and f(x) and its inverse are differentiable, with non-zero, then
- Derivative of one variable with respect to another when both are functions of a third variable: Let and . Now
- Implicit differentiation: If is an implicit function, we have: dy/dx = - (∂f / ∂x) / (∂f / ∂y).
In addition, the derivatives of some common functions are useful to know. See the table of derivatives.
As an example, the derivative of
Using derivatives to graph functions
Derivatives are a useful tool for examining the graphs of functions. In particular, the points in the interior of the domain of a real-valued function which take that function to local extrema will all have a first derivative of zero. However, not all critical points are local extrema; for example, f(x)=x3 has a critical point at x=0, but it has neither a maximum nor a minimum there. The first derivative test and the second derivative test provide ways to determine if the critical points are maxima, minima or neither.
In the case of multidimensional domains, the function will have a partial derivative of zero with respect to each dimension at local extrema. In this case, the Second Derivative Test can still be used to characterize critical points, by considering the eigenvalues of the Hessian matrix of second partial derivatives of the function at the critical point. If all of the eigenvalues are positive, then the point is a local minimum; if all are negative, it is a local maximum. If there are some positive and some negative eigenvalues, then the critical point is a saddle point, and if none of these cases hold then the test is inconclusive (e.g., eigenvalues of 0 and 3).
Once the local extrema have been found, it is usually rather easy to get a rough idea of the general graph of the function, since (in the single-dimensional domain case) it will be uniformly increasing or decreasing except at critical points, and hence (assuming it is continuous) will have values in between its values at the critical points on either side.
Where a function depends on more than one variable, the concept of a partial derivative is used. Partial derivatives can be thought of informally as taking the derivative of the function with all but one variable held temporarily constant near a point. Partial derivatives are represented as ∂/∂x (where ∂ is a rounded 'd' known as the 'partial derivative symbol'). Some people pronounce the partial derivative symbol as 'der' rather than the 'dee' used for the standard derivative symbol, 'd'.
The concept of derivative can be extended to more general settings. The common thread is that the derivative at a point serves as a linear approximation of the function at that point. Perhaps the most natural situation is that of functions between differentiable manifolds; the derivative at a certain point then becomes a linear transformation between the corresponding tangent spaces and the derivative function becomes a map between the tangent bundles.
For complex functions of a complex variable differentiability is a much stronger condition than that the real and imaginary part of the function are differentiable with respect to the real and imaginary part of the argument. For example, the function f(x + iy) = x + 2iy satisfies the latter, but not the first. See also Holomorphic function.
- Derivative (generalizations)
- table of derivatives
- derivative (examples)
- smooth function
- Automatic differentiation
- WIMS Function Calculator makes online calculation of derivatives; this software enables also interactive exercises.
- Spivak, Michael; Calculus (3rd edition, 1994) Publish or Perish Press. ISBN 0914098896. Explains why all this works.
- Thompson, Silvanus Phillips, Calculus made easy : being a very-simplest introduction to those beautiful methods of reckoning which are generally called by the terrifying names of the differential calculus and the integral calculus New York : St. Martin's Press, 1998 ISBN 0312185480. Introduced by Martin Gardner. "What one fool can do, another can."
- Larson, Ron; Hostetler, Robert P.; and Edwards, Bruce H. (2003). Calculus of a Single Variable: Early Transcendental Functions (3rd edition). Houghton Mifflin Company. ISBN 061822307X.
- Anton, Howard (1980). Calculus with analytical geometry.. New York:John Wiley and Sons. ISBN 0-471-03248-4.
af:Afgeleide cs:Derivace da:Differentialregning de:Differentialrechnung es:Función derivada eo:Derivaĵo (matematiko) fr:Dérivée ko:미분 io:Derivado id:Turunan it:Derivata he:נגזרת lt:Išvestinė nl:Afgeleide ja:微分 no:Derivasjon pl:Pochodna funkcji pt:Derivada ro:Derivată ru:Производная функции simple:Derivative fi:Derivaatta sv:Derivata th:อนุพันธ์ tr:Türev zh:导数