enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Test functions for optimization - Wikipedia

    en.wikipedia.org/wiki/Test_functions_for...

    Test functions for optimization. In applied mathematics, test functions, known as artificial landscapes, are useful to evaluate characteristics of optimization algorithms, such as: Convergence rate. Precision.

  3. Convergence tests - Wikipedia

    en.wikipedia.org/wiki/Convergence_tests

    In mathematics, convergence tests are methods of testing for the convergence, conditional convergence, absolute convergence, interval of convergence or divergence of an infinite series .

  4. Steffensen's method - Wikipedia

    en.wikipedia.org/wiki/Steffensen's_method

    The version of Steffensen's method implemented in the MATLAB code shown below can be found using the Aitken's delta-squared process for accelerating convergence of a sequence. To compare the following formulae to the formulae in the section above, notice that This method assumes starting with a linearly convergent sequence and increases the rate of convergence of that sequence. If the signs of ...

  5. Rosenbrock function - Wikipedia

    en.wikipedia.org/wiki/Rosenbrock_function

    Plot of the Rosenbrock function of two variables. Here , and the minimum value of zero is at . In mathematical optimization, the Rosenbrock function is a non- convex function, introduced by Howard H. Rosenbrock in 1960, which is used as a performance test problem for optimization algorithms. [1] It is also known as Rosenbrock's valley or Rosenbrock's banana function. The global minimum is ...

  6. Line search - Wikipedia

    en.wikipedia.org/wiki/Line_search

    Line search. In optimization, line search is a basic iterative approach to find a local minimum of an objective function . It first finds a descent direction along which the objective function will be reduced, and then computes a step size that determines how far should move along that direction. The descent direction can be computed by various ...

  7. Golden-section search - Wikipedia

    en.wikipedia.org/wiki/Golden-section_search

    The discussion here is posed in terms of searching for a minimum (searching for a maximum is similar) of a unimodal function. Unlike finding a zero, where two function evaluations with opposite sign are sufficient to bracket a root, when searching for a minimum, three values are necessary. The golden-section search is an efficient way to progressively reduce the interval locating the minimum ...

  8. Newton's method - Wikipedia

    en.wikipedia.org/wiki/Newton's_method

    In these situations, it may be appropriate to approximate the derivative by using the slope of a line through two nearby points on the function. Using this approximation would result in something like the secant method whose convergence is slower than that of Newton's method.

  9. Particle swarm optimization - Wikipedia

    en.wikipedia.org/wiki/Particle_swarm_optimization

    In computational science, particle swarm optimization (PSO) [ 1 ] is a computational method that optimizes a problem by iteratively trying to improve a candidate solution with regard to a given measure of quality. It solves a problem by having a population of candidate solutions, here dubbed particles, and moving these particles around in the ...