Source: wikibot/graduated-optimization

= Graduated optimization
{wiki=Graduated_optimization}

Graduated optimization is a computational technique used primarily in the context of optimization and machine learning, particularly for solving complex problems that may be non-convex or have multiple local minima. The general idea behind graduated optimization is to gradually transform a difficult optimization problem into a simpler one, which can be solved more easily.