New Machine Learning Optimization Technique - Part I

New Machine Learning Optimization Technique - Part I

In this series, we discuss a technique to find either the minima or the roots of a chaotic, unsmooth, or discrete function. A root-finding technique that works well for continuous, differentiable functions is successfully adapted and applied to piece-wise constant functions with an infinite number of discontinuities. It even works if the function has no root: it will then find minima instead. In order to work, some constraints must be put on the parameters used in the algorithm, while avoiding over-fitting at the same time. This would also be true true anyway for the smooth, continuous, differentiable case. It does not work in the classical sense where an iterative algorithm converges to a solution. Here the iterative algorithm always diverges, yet there is has a clear stopping rule that tells us when we are very close to a solution, and what to do to find the exact solution.?This is the originality of the method, and why I call it new. Our technique, like many machine learning techniques, can generate false positives or false negatives, and one aspect of our methodology is to minimize this problem.

Applications are discussed, as well as full implementation with results, for a real-life, challenging case, and on simulated data. This first article in this series is a detailed introduction. Interestingly, we also show an example where a continuous, differentiable function, with a very large number of wild oscillations, benefit from being transformed into a non-continuous, non-differentiable function and then use our non-standard technique to find roots (or minima) as the standard technique fails. For the moment, we limit ourselves to the one-dimensional case.?

Read full article here.

Dr.Debasish Banerjee

Turnaround professional in the manufacturing and engineering domains with applications in ML and AI

3 年

The iteration for the root optimization might do well to map out the changes in entropy triggered by incoming fresh data scatter. An algorithm that intuitively traces the logarithmic compression of the randomness and minimizes the same will accelerate the convergence to an approximate root.

要查看或添加评论,请登录

Vincent Granville的更多文章

社区洞察

其他会员也浏览了