Source: wikibot/graduated-optimization
= Graduated optimization
{wiki=Graduated_optimization}
Graduated optimization is a computational technique used primarily in the context of optimization and machine learning, particularly for solving complex problems that may be non-convex or have multiple local minima. The general idea behind graduated optimization is to gradually transform a difficult optimization problem into a simpler one, which can be solved more easily.