A Critical Analysis of the Theoretical Framework of the Extreme Learning Machine
Irina Perfilievaa, Nicolas Madrid, Manuel Ojeda-Aciego, Piotr Artiemjew, Agnieszka Niemczynowicz
TL;DR
The paper critically reexamines Huang06’s theoretical foundation for the Extreme Learning Machine (ELM), arguing that key proofs are mathematically flawed and that exact interpolation cannot be guaranteed under the original randomization assumptions. It constructs a counterexample dataset $S$ (with $N=400$) showing failure of the claimed results and dissects the erroneous steps in Theorems 2.1 and 2.2, including incorrect implications of differentiating activation relationships. It then proposes a corrected theoretical direction (Theorem teNew and Corollaries) that places restrictions on the data and activation function to obtain valid probabilistic guarantees, and demonstrates that ELM can still exactly learn the counterexample dataset under alternative hyperparameters (e.g., ReLU with $ ilde N=8000$). The findings highlight the need for a rigorous, conditionally valid mathematical framework for ELM and suggest practical modifications to the randomization scheme to preserve usefulness while ensuring theoretical coherence. Overall, the work clarifies the limits of Huang06’s claims and offers a pathway toward a sound, probabilistic foundation for ELM with explicit conditions on activation, data, and network size.
Abstract
Despite the number of successful applications of the Extreme Learning Machine (ELM), we show that its underlying foundational principles do not have a rigorous mathematical justification. Specifically, we refute the proofs of two main statements, and we also create a dataset that provides a counterexample to the ELM learning algorithm and explain its design, which leads to many such counterexamples. Finally, we provide alternative statements of the foundations, which justify the efficiency of ELM in some theoretical cases.
