answersLogoWhite

0


Best Answer

Intuitive derivation

Finite-difference methods approximate the solutions to differential equations by replacing derivative expressions with approximately equivalent difference quotients. That is, because the first derivative of a function f is, by definition,

then a reasonable approximation for that derivative would be to take

for some small value of h. In fact, this is the forward difference equation for the first derivative. Using this and similar formulae to replace derivative expressions in differential equations, one can approximate their solutions without the need for calculus.

[edit] Derivation from Taylor's polynomial

Assuming the function whose derivatives are to be approximated is properly-behaved, by Taylor's theorem,

where n! denotes the factorial of n, and Rn(x) is a remainder term, denoting the difference between the Taylor polynomial of degree n and the original function. Again using the first derivative of the function f as an example, by Taylor's theorem,

f(x0 + h) = f(x0) + f'(x0)h + R1(x),

which, with some minor algebraic manipulation, is equivalent to

so that for R1(x) sufficiently small,

[edit] Accuracy and order

The error in a method's solution is defined as the difference between its approximation and the exact analytical solution. The two sources of error in finite difference methods are round-off error, the loss of precision due to computer rounding of decimal quantities, and truncation error or discretization error, the difference between the exact solution of the finite difference equation and the exact quantity assuming perfect arithmetic (that is, assuming no round-off).

The finite difference method relies on discretizing a function on a grid.

To use a finite difference method to attempt to solve (or, more generally, approximate the solution to) a problem, one must first discretize the problem's domain. This is usually done by dividing the domain into a uniform grid (see image to the right). Note that this means that finite-difference methods produce sets of discrete numerical approximations to the derivative, often in a "time-stepping" manner.

An expression of general interest is the local truncation error of a method. Typically expressed using Big-O notation, local truncation error refers to the error from a single application of a method. That is, it is the quantity f'(xi) − f'i if f'(xi) refers to the exact value and f'i to the numerical approximation. The remainder term of a Taylor polynomial is convenient for analyzing the local truncation error. Using the Lagrange form of the remainder from the Taylor polynomial for f(x0 + h), which is

, where x0 < ξ < x0 + h,

the dominant term of the local truncation error can be discovered. For example, again using the forward-difference formula for the first derivative, knowing that f(xi) = f(x0 + ih),

and with some algebraic manipulation, this leads to

and further noting that the quantity on the left is the approximation from the finite difference method and that the quantity on the right is the exact quantity of interest plus a remainder, clearly that remainder is the local truncation error. A final expression of this example and its order is:

This means that, in this case, the local truncation error is proportional to the step size.

[edit] Example: ordinary differential equation

For example, consider the ordinary differential equation

The Euler method for solving this equation uses the finite difference quotient

to approximate the differential equation by first substituting in for u'(x) and applying a little algebra to get

The last equation is a finite-difference equation, and solving this equation gives an approximate solution to the differential equation.

[edit] Example: The heat equation

Consider the normalized heat equation in one dimension, with homogeneous Dirichlet boundary conditions

(boundary condition) (initial condition)

One way to numerically solve this equation is to approximate all the derivatives by finite differences. We partition the domain in space using a mesh x0,...,xJ and in time using a mesh t0,....,tN. We assume a uniform partition both in space and in time, so the difference between two consecutive space points will be h and between two consecutive time points will be k. The points

will represent the numerical approximation of u(xj,tn).

[edit] Explicit methodThe stencil for the most common explicit method for the heat equation.

Using a forward difference at time tn and a second-order central difference for the space derivative at position xj ("FTCS") we get the recurrence equation:

This is an explicit method for solving the one-dimensional heat equation.

We can obtain from the other values this way:

where r = k / h2.

So, knowing the values at time n you can obtain the corresponding ones at time n+1 using this recurrence relation. and must be replaced by the boundary conditions, in this example they are both 0.

This explicit method is known to be numerically stable and convergent whenever . The numerical errors are proportional to the time step and the square of the space step:

[edit] Implicit methodThe implicit method stencil.

If we use the backward difference at time ti + 1 and a second-order central difference for the space derivative at position xj ("BTCS") we get the recurrence equation:

This is an implicit method for solving the one-dimensional heat equation.

We can obtain from solving a system of linear equations:

The scheme is always numerically stable and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step.

[edit] Crank-Nicolson method

Finally if we use the central difference at time tn + 1 / 2 and a second-order central difference for the space derivative at position xj ("CTCS") we get the recurrence equation:

This formula is known as the Crank-Nicolson method.

The Crank-Nicolson stencil.

We can obtain from solving a system of linear equations:

The scheme is always numerically stable and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over the time step and formally are of the fourth degree regarding the space step:

However, near the boundaries, the error is often O(h2) instead of O(h4).

Usually the Crank-Nicolson scheme is the most accurate scheme for small time steps. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive. The implicit scheme works the best for large time steps.

User Avatar

Wiki User

15y ago
This answer is:
User Avatar

Add your answer:

Earn +20 pts
Q: What is the method of finite differences for linear equations?
Write your answer...
Submit
Still have questions?
magnify glass
imp
Related questions

Give the example for finite difference method in fluid dynamics?

In fluid dynamics, a common example of using finite difference method is the discretization of the Navier-Stokes equations to solve for fluid flow equations. This entails approximating spatial derivatives with finite differences on a grid, which allows for numerical simulation of the fluid behavior in a computational domain.


What is the theory of finite differential method?

Finite Differential Methods (FDM) are numerical methods for approximating the solutions to differential equations using finite difference equations to approximate derivatives.


Compare the symbolic method for solving linear equations to the methods of using a table or graph?

Equations = the method


What is the disadvantage of using the substitution method in solving linear equations rather than the graph method?

There are no disadvantages. There are three main ways to solve linear equations which are: substitution, graphing, and elimination. The method that is most appropriate can be found by looking at the equation.


Why is it the graphing method is the least reliable method in solving system of linear equations?

putang ina nyu


How do you find the solution set of pair linear equations?

By the substitution method By the elimination method By plotting them on a graph


Can you patent math work?

Higher-level mathematical concepts, such as entirely new methods of calculation, may be patented. For example, this is the abstract of a patent application being reviewed at the moment: A method for obtaining an estimate of a solution to a first system of linear equations. The method comprises obtaining a second system of linear equations, obtaining an estimate of a solution to said second system of linear equations, determining differences between said first and second systems of linear equations, and determining an estimate of a solution to said first system of linear equations based upon said differences and said estimate of said solution to said second system of linear equations. In the language of the various laws, this would be called a "process," which the statute degines as "a process, act, or method." Also according to the law, the process must be useful and novel. It's worth noting, though, that case law has defined that "laws of nature, physical phenomena, and abstract ideas are not patentable subject matter."


What has the author Vivette Girault written?

Vivette Girault has written: 'Finite element approximation of the Navier-Stokes equations' -- subject(s): Finite element method, Navier-Stokes equations, Numerical solutions, Viscous flow, Instrumentation, Airway (Medicine), Methods, Respiratory Therapy, Cardiopulmonary Resuscitation, Trachea, Airway Obstruction, Intubation, Therapy, Airway Management 'Finite element methods for Navier-Stokes equations' -- subject(s): Finite element method, Navier-Stokes equations, Numerical solutions, Viscous flow


What has the author Eric B Becker written?

Eric B. Becker has written: 'Development of non-linear finite element computer code' -- subject(s): Finite element method, Strains and stresses 'Finite elements' -- subject(s): Finite element method


What is a method for solving a system of linear equations in which you multiply one or both equations by a number to get rid of a variable term?

It is called solving by elimination.


What has the author Thomas Kerkhoven written?

Thomas Kerkhoven has written: 'L [infinity] stability of finite element approximations to elliptic gradient equations' -- subject(s): Boundary value problems, Elliptic Differential equations, Finite element method, Stability


How does the bisection method work when solving nonlinear equations?

it works exactly the same as it does with linear equations, you don't need to do any differentiation or anything fancy with this method, just have to plug in values of x, so it shouldn't make a difference if the equation is linear or nonlinear.