List of numerical analysis topics
From Wikipedia, the free encyclopedia
This is a list of numerical analysis topics, by Wikipedia page.
[edit] General
- Iterative method
- Sequence transformations — methods to accelerate the speed of convergence
- Level set method
- Level set (data structures) — data structures for representing level sets
- Abramowitz and Stegun
- Curse of dimensionality
- Superconvergence
- Termination
- Discretization
- Collocation method — discretizes a continuous equation by requiring it only to hold at certain points
- Difference quotient
- Computational complexity of mathematical operations
[edit] Error
- Approximation
- Approximation error
- Arithmetic precision
- Condition number
- Discretization error
- Floating point number
- Guard digit — extra precision introduced during a computation to reduce round-off error
- Arbitrary-precision arithmetic
- Loss of significance
- Numerical stability
- Error propagation:
- Residual (mathematics)
- Round-off error
- Significant figures
- Truncation
- Well-posed problem
- Affine arithmetic
[edit] Elementary and special functions
- Summation:
- Multiplication:
- Multiplication algorithm — general discussion, simple methods
- Karatsuba algorithm — the first algorithm which is faster than straightforward multiplication
- Toom-Cook multiplication — generalization of Karatsuba multiplication
- Schönhage-Strassen algorithm — based on Fourier transform, asymptotically fastest
- Exponentiation:
- Polynomials:
- Square roots and other roots:
- Elementary functions (exponential, logarithm, trigonometric functions):
- Generating trigonometric tables
- CORDIC — shift-and-add algorithm using a table of arc tangents
- BKM algorithm — shift-and-add algorithm using a table of logarithms and complex numbers
- Gamma function:
- Lanczos approximation
- Spouge's approximation — modification of Stirling's approximation; easier to apply than Lanczos
- Computation of π:
- Gauss-Legendre algorithm — iteration which converges quadratically to π, based on arithmetic-geometric mean
- Bailey-Borwein-Plouffe formula — can be used to compute individual hexadecimal digits of π
- Borwein's algorithm — iteration which converges quartically to 1/π, and other algorithms
[edit] Numerical linear algebra
Numerical linear algebra — study of numerical algorithms for linear algebra problems
[edit] Basic concepts
- Types of matrices appearing in numerical analysis:
- Sparse matrix
- Banded matrix
- Tridiagonal matrix
- Pentadiagonal matrix
- Skyline matrix
- Circulant matrix
- Triangular matrix
- Diagonally dominant matrix
- Hilbert matrix — example of a matrix which is extremely ill-conditioned (and thus difficult to handle)
- Sparse matrix
- Algorithms for matrix multiplication:
- Strassen algorithm
- Coppersmith–Winograd algorithm
- Freivald's algorithm — a randomized algorithm for checking the result of a multiplication
[edit] Solving systems of linear equations
- Gaussian elimination
- Row echelon form — matrix in which all entries below a nonzero entry are zero
- Reduced row echelon form — matrix in row echelon form, in which the leading coefficients are one and have only zeros above them
- Gauss–Jordan elimination — variant in which the entries below the pivot are also zeroed
- Montante's method — variant which ensures that all entries remain integers if the initial matrix has integer entries
- Tridiagonal matrix algorithm — simplified form of Gaussian elimination for tridiagonal matrices
- LU decomposition
- Block LU decomposition
- Cholesky decomposition — for solving a system with a positive definite matrix
- Frontal solver — for sparse matrices; used in finite element methods
- Levinson recursion — for Toeplitz matrices
- Iterative methods:
- Jacobi method
- Gauss-Seidel method
- Successive over-relaxation (SOR) — a technique to accelerate the Gauss-Seidel method
- Modified Richardson iteration
- Conjugate gradient method — assumes that the matrix is positive definite
- Biconjugate gradient method (BiCG)
- Generalized minimal residual method (GMRES) — based on the Arnoldi iteration
- Stone method — uses an incomplete LU decomposition
- Preconditioner
[edit] Eigenvalue algorithms
Eigenvalue algorithm — an numerical algorithm for locating the eigenvalues of a matrix
- Power iteration
- Inverse iteration
- Rayleigh quotient iteration
- Arnoldi iteration — based on Krylov subspaces
- Lanczos algorithm — Arnoldi, specialized for positive-definite matrices
- QR algorithm
- Jacobi eigenvalue algorithm — select a small submatrix which can be diagonalized exactly, and repeat
- Jacobi rotation — the building block, almost a Givens rotation
- Divide-and-conquer eigenvalue algorithm
[edit] Other concepts and algorithms
- Orthogonalization algorithms:
- QR decomposition
- Krylov subspace
- Block matrix pseudoinverse
- Bidiagonalization
- Pivot element — entry in a matrix on which the algorithm concentrates
[edit] Interpolation
- Nearest neighbor interpolation — takes the value of the nearest neighbor
[edit] Polynomial interpolation
Polynomial interpolation — interpolation by polynomials
- Linear interpolation
- Runge's phenomenon
- Vandermonde matrix
- Chebyshev polynomials
- Chebyshev nodes
- Lebesgue constant (interpolation)
- Different forms for the interpolant:
- Newton polynomial
- Divided differences
- Neville's algorithm — for evaluating the interpolant; based on the Newton form
- Lagrange polynomial
- Bernstein polynomial
- Newton polynomial
- Extensions to multiple dimensions:
- Hermite interpolation
- Birkhoff interpolation
[edit] Spline interpolation
Spline interpolation — interpolation by piecewise polynomials
- Spline (mathematics) — the piecewise polynomials used as interpolants
- Perfect spline — polynomial spline of degree m whose mth derivate is ±1
- Cubic Hermite spline
- Monotone cubic interpolation
- Hermite spline
- Cardinal spline
- Kochanek-Bartels spline
- Catmull-Rom spline
- Bézier spline
- B-spline
- Truncated power function
- De Boor's algorithm — generalizes De Casteljau's algorithm
- Nonuniform rational B-spline (NURBS)
- Blossom (mathematics) — a unique, affine, symmetric map associated to a polynomial or spline
- See also: List of numerical computational geometry topics
[edit] Trigonometric interpolation
Trigonometric interpolation — interpolation by trigonometric polynomials
- Fast Fourier transform
- Bluestein's FFT algorithm
- Bruun's FFT algorithm
- Cooley-Tukey FFT algorithm
- Split-radix FFT algorithm — variant of Cooley-Tukey that uses a blend of radices 2 and 4
- Goertzel algorithm
- Prime-factor FFT algorithm
- Rader's FFT algorithm
- Butterfly diagram
- Twiddle factor — the trigonometric constant coefficients that are multiplied by the data
- Sigma approximation
- Dirichlet kernel — convolving any function with the Dirichlet kernel yields its trigonometric interpolant
[edit] Other interpolants
- Simple rational approximation
- Wavelet
- Inverse distance weighting
- Radial basis function
- Polyharmonic spline — a commonly used radial basis function
- Thin plate spline — a specific polyharmonic spline: r2 log r
- Radial basis function network — neural network using radial basis functions as activation functions
- Subdivision surface — constructed by recursively subdividing a piecewise linear interpolant
- Slerp
- Irrational base discrete weighted transform
- Pick matrix — appears when interpolating by analytic functions which must be bounded in the unit disc
- Pareto interpolation
- Multivariate interpolation — the function being interpolated depends on more than one variable
- Lanczos resampling — based on convolution with a sinc function
- Natural neighbor interpolation
- PDE surface
- Method based on polynomials are listed under Polynomial interpolation
[edit] Approximation theory
- Orders of approximation
- Lebesgue's lemma
- Curve fitting
- Modulus of continuity — measures smoothness of a function
- Remez algorithm — for constructing the best polynomial approximation
- Jackson's inequality — upper bound for best approximation by a trigonometric polynomial
- Surrogate model — application: replacing a function that is hard to evaluate by a simpler function
- Different approximations:
[edit] Miscellaneous
- Extrapolation
- Linear predictive analysis — linear extrapolation
- Regression analysis
- Curve-fitting compaction
[edit] Finding roots of nonlinear equations
- See #Numerical linear algebra for linear equations
Root-finding algorithm — algorithms for solving the equation f(x) = 0
- General methods:
- Bisection method — simple and robust; linear convergence
- Lehmer-Schur algorithm — variant for complex functions
- Fixed point iteration
- Newton's method — based on linear approximation around the current iterate; quadratic convergence
- Secant method — based on linear interpolation at last two iterates
- False position method — secant method with ideas from the bisection method
- Müller's method — based on quadratic interpolation at last three iterates
- Inverse quadratic interpolation — similar to Müller's method, but interpolates the inverse
- Halley's method — uses f, f' and f''; achieves the cubic convergence
- Brent's method — combines bisection method, secant method and inverse quadratic interpolation
- Bisection method — simple and robust; linear convergence
- Methods for polynomials:
- Methods for other special cases:
- Analysis:
- Numerical continuation — tracking a root as one parameters in the equation changes
[edit] Optimization
Optimization (mathematics) — algorithm for finding maxima or minima of a given function
[edit] Basic concepts
- Active set
- Candidate solution
- Constraint (mathematics)
- Corner solution
- Global optimum and Local optimum
- Maxima and minima
- Slack variable
- Surplus variable
- Continuous optimization
- Discrete optimization is not a part of numerical analysis
[edit] Linear programming
Linear programming (also treats integer programming) — objective function and contraints are linear
- Algorithms for linear programming:
- Algorithms for integer programming:
- Linear complementarity problem
- Fourier–Motzkin elimination
[edit] Nonlinear programming
Nonlinear programming — the most general optimization problem in the usual framework
- Special cases of nonlinear programming:
- Quadratic programming
- Convex optimization
- Linear matrix inequality
- Conic optimization
- Semidefinite programming
- Second order cone programming
- Quadratic programming (see above)
- Subgradient method — extension of steepest descent for problems with a nondifferentiable objective function
- Geometric programming — problems involving posynomials
- Posynomial — similar to polynomials, but coefficients have to be positive while exponents need not be integers
- Quadratically constrained quadratic program
- General algorithms:
- Concepts:
- Gradient descent
- Successive linear programming (SLP) — replace problem by a linear programming problem, solve that, and repeat
- Newton's method in optimization
- Quasi-Newton method — uses an approximation to the Hessian instead of the true Hessian
- BFGS method — a quasi-Newton method
- Nelder-Mead method
- Golden section search
- Tabu search
- Least absolute deviations
- Nonlinear least squares
- Gauss-Newton algorithm
- Levenberg-Marquardt algorithm
- Expectation-maximization algorithm
- Osem (mathematics) — ordered subset expectation maximization
- Nearest neighbor search
- Mixed complementarity problem
[edit] Uncertainty and randomness
- Approaches to deal with uncertainty:
- Random optimization algorithms:
- Simulated annealing
- Adaptive simulated annealing — variant in which the algorithm parameters are adjusted during the computation.
- Great Deluge algorithm
- Evolutionary algorithm, Evolution strategy
- Memetic algorithm
- Particle swarm optimization
- Cooperative optimization
- Stochastic tunneling
- see also the section Monte Carlo method
- Simulated annealing
[edit] Theoretical aspects
- Convex analysis
- Duality:
- Farkas's lemma
- Karush-Kuhn-Tucker conditions
- Lagrange multipliers
- No free lunch in search and optimization
- Relaxation technique (mathematics)
[edit] Applications
- Automatic label placement
- Cutting stock problem
- Demand Optimization
- Entropy maximization
- Expenditure minimization problem
- Inventory control problem
- Job-shop problem
- Multidisciplinary design optimization
- Paper bag problem
- Process optimization
- Stigler diet
- Stress majorization
- Trajectory optimization
- Utility maximization problem
[edit] Miscellaneous
- Combinatorial optimization
- Dynamic programming
- Global optimization:
- Infinite-dimensional optimization
- Optimal control
- Pontryagin's minimum principle — infinite-dimensional version of Lagrange multipliers
- Shape optimization
- Optimal control
- Optimal substructure
- Algorithmic concepts:
- Famous test functions for optimization:
- Rosenbrock function — two-dimensional function with a banana-shaped valley
- Shekel function — multimodal and multidimensional
Mathematical Programming Society
[edit] Numerical quadrature
Numerical integration — the numerical evaluation of an integral
- Rectangle method
- Trapezium rule
- Simpson's rule
- Newton-Cotes formulas
- Gaussian quadrature
- Romberg's method
- Tanh-sinh quadrature — variant of Gaussian quadrature which works well with singularities at the end points
- Clenshaw-Curtis quadrature — based on expanding the integrand in terms of Chebyshev polynomials
- Adaptive quadrature — adapting the subintervals in which the integration interval is divided depending on the integrand
- Monte Carlo integration — takes random samples of the integrand
- See also #Monte Carlo method
- T-integration — a non-standard method
- Lebedev grid — grid on a sphere with octahedral symmetry
- Sparse grid
- Numerical differentiation
- Euler-Maclaurin formula
[edit] Numerical ordinary differential equations
Numerical ordinary differential equations — the numerical solution of ordinary differential equations (ODEs)
- Euler integration — the most basic method for solving an ODE
- Explicit and implicit methods — implicit methods need to solve an equation at every step
- Runge–Kutta methods — one of the two main classes of methods for initial-value problems
- Midpoint method — a second-order method with two stages
- Cash-Karp — a fifth-order method with six stages and an embedded fourth-order method
- Dormand-Prince — another fifth-order method with six stages and an embedded fourth-order method
- Runge–Kutta–Fehlberg method — another fifth-order method with six stages and an embedded fourth-order method
- Multistep method — the other main class of methods for initial-value problems
- Methods designed for the solution of ODEs from classical physics:
- Newmark-beta method — based on the extended mean-value theorem
- Verlet integration — a popular second-order method
- Beeman's algorithm — a two-step method extending the Verlet method
- Geometric integrator — a method that preserves some geometric structure of the equation
- Symplectic integrator — a method for the solution of Hamilton's equations that preserves the symplectic structure
- Euler–Cromer algorithm — variant of Euler method which is symplectic when applied to separable Hamiltonians
- Symplectic integrator — a method for the solution of Hamilton's equations that preserves the symplectic structure
- Adaptive stepsize — automatically changing the step size when that seems advantageous
- Stiff equation — roughly, an ODE for which the unstable methods needs a very short step size, but stable methods do not.
- Shooting method — a method for the solution of boundary-value problems
- Methods for solving stochastic differential equations (SDEs):
- Euler-Maruyama method — generalization of the Euler method for SDEs
- Milstein method — a method with strong order one
- Runge–Kutta method (SDE) — generalization of the family of Runge–Kutta methods for SDEs
- Methods for solving integral equations:
- Nyström method — replaces the integral with a quadrature rule
- Bi-directional delay line
[edit] Numerical partial differential equations
Numerical partial differential equations — the numerical solution of partial differential equations (PDEs)
- Methods for the solution of PDEs:
- Finite difference method — based on approximating differential operators with difference operators
- Finite difference — the discrete analogue of a differential operator
- Difference operator — the numerator of a finite difference
- Discrete Laplace operator — finite-difference approximation of the Laplace operator
- Discrete Poisson equation — discrete analogue of the Poisson equation using the discrete Laplace operator
- Five-point stencil — standard finite-difference approximation of the Laplace operator in two dimensions
- Crank-Nicolson method — second-order method for heat and related PDEs
- Lax-Wendroff method — second-order method for hyperbolic PDEs
- Finite-difference time-domain method — a finite-difference method for electrodynamics
- Finite difference — the discrete analogue of a differential operator
- Finite element method, finite element analysis — based on a discretization of the space of solutions
- Finite element method in structural mechanics — a physical approach to finite element methods
- Galerkin method — a finite element method in which the residual is orthogonal to the finite element space
- Rayleigh-Ritz method — a finite element method based on variational principles
- Direct stiffness method — a particular implementation of the finite element method, often used in structural analysis
- Trefftz method
- Modal analysis using FEM — solution of eigenvalue problems to find natural vibrations
- Spectral method — based on the Fourier transformation
- Numerical method of lines — reduces the PDE to a large system of ordinary differential equations
- Boundary element method — based on transforming the PDE to an integral equation on the boundary of the domain
- Analytic element method — similar to the boundary element method, but the integral equation is evaluated analytically
- Finite volume method — based on dividing the domain in many small domains; popular in computational fluid dynamics
- Godunov's scheme — first-order conservative scheme for fluid flow, based on piecewise constant approximation
- MUSCL scheme — second-order variant of Godunov's scheme
- AUSM — advection upstream splitting method
- Discrete element method — a method in which the elements can move freely relative to each other
- Meshfree methods — does not use a mesh, but uses a particle view of the field
- Uniform theory of diffraction — specifically designed for scattering problems in electromagnetics
- Particle-in-cell — used especially in fluid dynamics
- High-resolution scheme
- Split-step method
- Fast marching method
- Finite difference method — based on approximating differential operators with difference operators
- Techniques for improving these methods:
- Multigrid, multigrid method — uses a hierarchy of nested meshes to speed up the methods
- Domain decomposition method — divides the domain in a few subdomains and solves the PDE on these subdomains
- Adaptive mesh refinement — uses the computed solution to refine the mesh only where necessary
- Fast Multipole Method — hierarchical method for evaluating particle-particle interactions
- Broad classes of methods:
- Mimetic methods — methods that respect in some sense the structure of the original problem
- Multiphysics — models consisting of various submodels with different physics
- Analysis:
- Lax equivalence theorem — a consistent method is convergent if and only if it is stable
- Courant–Friedrichs–Lewy condition — stability condition for hyperbolic PDEs
- Numerical diffusion — diffusion introduced by the numerical method, above to that which is naturally present
- Numerical resistivity — the same, with resistivity instead of diffusion
- Weak formulation — a functional-analytic reformulation of the PDE necessary for some methods
- Total variation diminishing — property of schemes that do not introduce spurious oscillations
- Godunov's theorem — linear monotone schemes can only be of first order
- Boundary conditions:
[edit] Monte Carlo method
- Variants of the Monte Carlo method:
- Box-Muller transform
- Low-discrepancy sequence
- Event generator
- Parallel tempering
- Variance reduction techniques:
- Applications:
- Metropolis light transport
- Monte Carlo methods in finance
- Quantum Monte Carlo
- Diffusion Monte Carlo — uses a Green function to solve the Schrödinger equation
- Gaussian quantum Monte Carlo
- Path integral Monte Carlo
- Reptation Monte Carlo
- Variational Monte Carlo
- Auxiliary field Monte Carlo — computes averages of operators in many-body quantum mechanical problems
- Cross-entropy method — for multi-extremal optimization and importance sampling
- Also see the list of statistics topics
[edit] Applications
- Climate model
- Computational chemistry
- Computational electromagnetics
- Computational fluid dynamics
- Large eddy simulation
- Smoothed particle hydrodynamics
- Acoustic analogy — used in numerical aeroacoustics to reduce sound sources to simple emitter types
- Computational physics
- Numerical model of solar system
[edit] Software
For software, see the list of numerical analysis software.