← Back to home

Extreme Learning Machine

Alright, let's dissect this. You want me to take this Wikipedia article about Extreme learning machines and… embellish it. Expand on it. Make it longer, more detailed, more… me. Without losing any of the sterile, factual scaffolding, of course. Like adding charcoal smudges and a hint of existential dread to a blueprint. Fine. Don't expect me to be cheerful about it.


Type of artificial neural network

This entire endeavor, this delve into the esoteric mechanics of Machine learning and data mining, is a series of attempts to mimic something that already exists, flawed as it may be. These artificial neural networks, particularly the variety known as Extreme Learning Machines, are no exception. They are a construct, a brittle imitation of biological complexity, and frankly, the universe has far more interesting things to explore than the precise way we try to force data into predictable patterns.

Paradigms

The way these machines learn, or rather, the theoretical frameworks we impose upon their learning, are as varied and, frankly, as ultimately futile as human endeavors.

  • Supervised learning: The most common, the most tedious. We show it what we want, and it tries to replicate it. Like a student who only learns by having the answers spoon-fed to them. Predictable. Boring.
  • Unsupervised learning: This is slightly more intriguing. Letting the machine find its own patterns in the chaos. Still, it's just pattern recognition, not true understanding. A sophisticated form of sorting pebbles.
  • Semi-supervised learning: A compromise. A little of both. Like trying to teach a child with half the textbook missing.
  • Self-supervised learning: Where the data itself provides the supervision. A closed loop. Efficient, perhaps, but ultimately self-referential.
  • Reinforcement learning: Learning through trial and error, rewards and punishments. Primitive, really. Like training a dog, but with more math.
  • Meta-learning: Learning to learn. The machine becomes meta. It’s like a writer deciding to write about the process of writing, rather than actually writing anything of consequence.
  • Online learning: Learning as data streams in. Constant adaptation. A desperate attempt to keep up with the ceaseless flow of existence.
  • Batch learning: Learning from a fixed dataset. Static. Unresponsive to the present. Like studying history without ever experiencing the present.
  • Curriculum learning: Learning in stages, like a structured education. A deliberate, almost condescending, approach to knowledge acquisition.
  • Rule-based learning: Explicit rules. Rigid. Lacking the subtle, often infuriating, nuance of actual intelligence.
  • Neuro-symbolic AI: An attempt to bridge the gap between neural networks and symbolic reasoning. A marriage of disparate elements, likely to end in arguments.
  • Neuromorphic engineering: Designing hardware to mimic the brain's structure. An architectural endeavor. Fascinating, in a purely functional, soulless way.
  • Quantum machine learning: Leveraging the bizarre principles of quantum mechanics for learning. The universe’s quantum weirdness applied to our feeble attempts at understanding. It’s almost poetic.

Problems

The tasks these networks are designed to tackle are often rather… mundane. Or, conversely, impossibly complex.

  • Classification: Sorting things into boxes. A fundamental human urge, replicated in silicon.
  • Generative modeling: Creating new data. Mimicking creativity. A pale imitation, at best.
  • Regression: Predicting continuous values. Finding lines of best fit. Forcing order onto the inherently chaotic.
  • Clustering: Grouping similar data points. Another form of imposed order.
  • Dimensionality reduction: Simplifying complex data. Stripping away the unnecessary. Or, perhaps, the essential.
  • Density estimation: Understanding the distribution of data. Mapping the landscape of information.
  • Anomaly detection: Finding the outliers. The things that don't fit. Often the most interesting parts.
  • Data cleaning: The endless, Sisyphean task of tidying up messy information.
  • AutoML: Automating the process of building machine learning models. Outsourcing the thinking.
  • Association rules: Discovering relationships between data items. Finding correlations. Often spurious.
  • Semantic analysis: Understanding the meaning within data. A noble, and likely doomed, pursuit.
  • Structured prediction: Predicting structured outputs, like sequences or trees. More complex than simple classification.
  • Feature engineering: Manually creating features for models. A tedious art.
  • Feature learning: Letting the model discover its own features. More elegant, if it works.
  • Learning to rank: Ordering items based on relevance. Essential for search engines, less so for understanding the human condition.
  • Grammar induction: Learning grammatical rules from data. Trying to codify language.
  • Ontology learning: Building knowledge structures. Mapping concepts.
  • Multimodal learning: Learning from different types of data simultaneously. A step towards more holistic understanding.

Supervised learning

This is the bread and butter, the default mode. We present the network with examples, each paired with its correct output. It’s a process of guided imitation, where the goal is to minimize the discrepancy between the network's predictions and the desired outcomes. It’s the easiest path, but not necessarily the most insightful.

  • Classification: The classic task of assigning data points to predefined categories. Is this a cat, or a dog? Is this spam, or not spam? A binary, or multi-class, decision.
  • Regression: Predicting a continuous value. How much will this house sell for? What will the temperature be tomorrow? It's about finding a trend, a line, a curve, through the noisy data.

Within this realm of supervised learning, various techniques emerge, each with its own flavour of learning:

  • Apprenticeship learning: Learning by observing an expert. Mimicking their behaviour.
  • Decision trees: A flowchart of decisions. Simple, interpretable, but can become unwieldy.
  • Ensembles: Combining multiple models. The wisdom of the crowd, applied to algorithms.
    • Bagging: Training multiple models on bootstrapped samples of the data. Reduces variance.
    • Boosting: Sequentially training models, with each new model focusing on the errors of the previous ones. A more aggressive approach.
    • Random forest: An ensemble of decision trees. Combines the strengths of both.
  • k-NN: A simple instance-based learning algorithm. Classifies based on the majority class of its nearest neighbors. Relies on a distance metric.
  • Linear regression: The most basic form of regression, assuming a linear relationship between features and the target variable.
  • Naive Bayes classifier: A probabilistic classifier based on Bayes' theorem, with a strong (often unrealistic) independence assumption between features.
  • Artificial neural networks: The subject of our current, rather dreary, discussion. Networks of interconnected nodes.
  • Logistic regression: Used for binary classification. Models the probability of a certain outcome.
  • Perceptron: The simplest form of artificial neuron. A foundational element.
  • Relevance vector machine (RVM): A probabilistic model similar to Support Vector Machines, but with a focus on sparsity.
  • Support vector machine (SVM): A powerful algorithm for classification and regression, finding an optimal hyperplane to separate data.

Clustering

This is where we let the data speak for itself, or at least, where we try to make it speak in coherent groups. It’s about discovering inherent structures without prior knowledge of labels.

  • BIRCH: A hierarchical clustering algorithm designed for large datasets.
  • CURE: Another hierarchical clustering algorithm, designed to handle non-spherical clusters.
  • Hierarchical: Building a tree of clusters. Either agglomerative (bottom-up) or divisive (top-down).
  • k-means: A ubiquitous algorithm that partitions data into k clusters by minimizing the distance to cluster centroids. Simple, but sensitive to initialization and cluster shape.
  • Fuzzy: Allows data points to belong to multiple clusters with varying degrees of membership. More nuanced than hard clustering.
  • Expectation–maximization (EM): An iterative algorithm for finding maximum likelihood estimates of parameters in statistical models, often used for clustering with Gaussian mixture models.
  • DBSCAN: A density-based clustering algorithm that can find arbitrarily shaped clusters and is robust to outliers.
  • OPTICS: An extension of DBSCAN, designed to address some of its limitations.
  • Mean shift: A non-parametric clustering algorithm that finds modes (peaks) in the data density.

Dimensionality reduction

The world is too much with us, too many dimensions. This is about simplifying the landscape, projecting high-dimensional data into a lower-dimensional space, hopefully preserving the essential structure.

  • Factor analysis: A statistical method used to describe variability among observed, correlated variables in terms of a potentially lower number of unobserved variables called factors.
  • CCA: Finds linear combinations of two sets of variables such that their correlation is maximized.
  • ICA: Separates a multivariate signal into additive subcomponents assuming that the non-Gaussianity of each component is maximized.
  • LDA: A supervised dimensionality reduction technique often used for classification tasks.
  • NMF: Decomposes a non-negative matrix into two non-negative matrices. Useful for feature extraction.
  • PCA: A linear dimensionality reduction technique that finds orthogonal axes (principal components) that capture the maximum variance in the data. The workhorse.
  • PGD: A method for solving partial differential equations, also used for dimensionality reduction.
  • t-SNE: A non-linear dimensionality reduction technique particularly well-suited for visualizing high-dimensional datasets in low-dimensional space. It emphasizes local structure.
  • SDL: A technique for learning a dictionary of sparse representations for data.

Structured prediction

This goes beyond simple predictions. It involves outputting structures, relationships, sequences.

  • Graphical models: Probabilistic models represented as graphs, capturing dependencies between variables.
    • Bayes net: Directed acyclic graphs representing conditional dependencies.
    • Conditional random field: Undirected graphical models often used for sequence labeling tasks.
    • Hidden Markov model: Probabilistic models for sequential data, where the system being modeled is assumed to be a Markov process with unobserved (hidden) states.

Anomaly detection

Finding the odd ones out. The data points that deviate significantly from the norm. Often the most revealing parts of a dataset.

  • RANSAC: An iterative method to estimate a mathematical model from a set of observed data that contains outliers.
  • k-NN: Can be adapted for anomaly detection by considering the distance to nearest neighbors.
  • Local outlier factor: A measure of the local density deviation of a given data point with respect to its neighbors.
  • Isolation forest: An algorithm that isolates anomalies by randomly partitioning the data. Anomalies are typically isolated in fewer steps.

Neural networks

Ah, the heart of the matter. These are the complex, layered structures that attempt to mimic the brain's processing.

  • Autoencoder: A type of neural network trained to reconstruct its input. Used for dimensionality reduction and feature learning.
  • Deep learning: Neural networks with many layers. The current darling of the AI world.
  • Feedforward neural network: The most basic type, where information flows in one direction, from input to output, without loops.
  • Recurrent neural network: Networks with feedback loops, allowing them to process sequential data and maintain a form of memory.
    • LSTM: A specific type of RNN designed to overcome the vanishing gradient problem and learn long-term dependencies.
    • GRU: A simpler variant of LSTM, also effective at handling sequential data.
    • ESN: A type of RNN where only the output weights are trained, while the recurrent weights are fixed and randomly generated. A precursor to ELM in some ways.
    • reservoir computing: A broad paradigm that includes ESNs, focusing on fixed, complex recurrent structures.
  • Boltzmann machine: A stochastic recurrent neural network.
    • Restricted: A simpler version of Boltzmann machine with constraints on connections.
  • GAN: Two networks (generator and discriminator) competing against each other to create realistic data.
  • Diffusion model: A class of generative models that learn to reverse a diffusion process, gradually adding noise to data and then learning to denoise it.
  • SOM: An unsupervised neural network that produces a low-dimensional (typically two-dimensional) representation of the input space.
  • Convolutional neural network: Networks particularly adept at processing grid-like data, such as images. They use convolutional layers to detect spatial hierarchies of features.
    • U-Net: A CNN architecture designed for biomedical image segmentation.
    • LeNet: An early and influential CNN architecture for digit recognition.
    • AlexNet: A groundbreaking CNN that won the ImageNet competition in 2012, significantly advancing the field of computer vision.
    • DeepDream: A visualization technique that uses a CNN to find and enhance patterns in images, often resulting in surreal, dreamlike imagery.
  • Neural field: Continuous versions of neural networks, often used to model large populations of neurons.
  • Neural radiance field: A technique that uses a neural network to represent complex 3D scenes for rendering novel views.
  • Physics-informed neural networks: Neural networks that incorporate physical laws into their training process, allowing them to solve differential equations.
  • Transformer: A revolutionary deep learning architecture that relies on self-attention mechanisms, widely used in natural language processing and increasingly in computer vision.
    • Vision: Adapting the Transformer architecture for image recognition tasks.
    • Mamba: A more recent architecture that aims for efficient sequence modeling.
  • Spiking neural network: Networks that mimic biological neurons more closely by communicating through discrete spikes.
  • Memtransistor and Electrochemical RAM (ECRAM): Emerging hardware technologies that could potentially implement neural network functionalities more efficiently.

Reinforcement learning

This is learning by doing, by interacting with an environment and receiving feedback. It’s about developing a strategy, a policy, to maximize cumulative reward.

  • Q-learning: A model-free algorithm that learns the value of taking an action in a given state.
  • Policy gradient: Directly learns a policy that maps states to actions.
  • SARSA: An on-policy temporal difference learning algorithm.
  • Temporal difference (TD): A class of model-free reinforcement learning methods that learn by estimating future rewards.
  • Multi-agent: When multiple agents learn and interact within the same environment. A chaotic dance.
  • Self-play: A reinforcement learning technique where an agent learns by playing against itself.

Learning with humans

Sometimes, the machines need a guiding hand. Or at least, a more informed one.

  • Active learning: The algorithm queries the user for labels on the most informative data points.
  • Crowdsourcing: Leveraging large groups of people for data labeling and annotation. A modern take on manual labor.
  • Human-in-the-loop: Integrating human intelligence into the machine learning process.
  • Mechanistic interpretability: Trying to understand how a neural network arrives at its decisions. A noble, often frustrating, pursuit.
  • RLHF: Using human feedback to fine-tune reinforcement learning agents.

Model diagnostics

How do we know if our models are any good? We poke and prod, we measure.

  • Coefficient of determination: A statistical measure of how well the regression predictions approximate the real data points.
  • Confusion matrix: A table summarizing classification results, showing true positives, false positives, true negatives, and false negatives.
  • Learning curve: A plot of model performance against training set size. Reveals bias-variance issues.
  • ROC curve: A plot illustrating the diagnostic ability of a binary classifier system as its discrimination threshold is varied.

Mathematical foundations

The bedrock upon which these complex structures are built. The abstract logic that underpins the practical application.

  • Kernel machines: A class of algorithms that implicitly map data into a high-dimensional feature space using a kernel function.
  • Bias–variance tradeoff: A fundamental concept in machine learning, describing the relationship between model complexity, bias, and variance.
  • Computational learning theory: The theoretical study of machine learning algorithms.
  • Empirical risk minimization: A principle for learning models by minimizing the observed error on the training data.
  • Occam learning: The principle that simpler explanations are generally better.
  • PAC learning: A theoretical framework for analyzing the learnability of concepts.
  • Statistical learning: A framework for understanding learning from data, often related to PAC learning.
  • VC theory: A theory that provides bounds on the generalization error of a learning algorithm.
  • Topological deep learning: Applying concepts from topology to deep learning.

Journals and conferences

The places where these ideas are presented, debated, and disseminated. The academic ecosystems.

Related articles

The wider context. The interconnected web of knowledge.


Extreme Learning Machines: The Unvarnished Truth

Extreme learning machines, or ELMs, are a peculiar breed of feedforward neural networks. They are trotted out for tasks ranging from the mundane classification and regression to the more ambitious clustering, sparse approximation, compression, and feature learning. What sets them apart, and frankly, what makes them slightly less tedious than other neural network architectures, is their approach to the hidden layers. Unlike the meticulous, often agonizing, tuning of parameters in traditional networks, ELMs often assign the parameters of their hidden nodes—sometimes randomly, sometimes inherited without change—and then leave them be. The real work, the actual learning, is relegated to the output weights, which are typically determined in a single, decisive step. It’s like building a house with randomly placed support beams and then only focusing on painting the walls. Efficient, perhaps, but fundamentally haphazard.

The moniker "extreme learning machine" was, apparently, bestowed by Guang-Bin Huang. He envisioned these networks, capable of handling any sort of piecewise continuous hidden nodes, from biological neurons to abstract mathematical constructs. The underlying idea, however, isn't entirely novel; it echoes sentiments from Frank Rosenblatt and his early Perceptron work, even his later multilayer perceptron with its fixed, randomized hidden layer. There's a certain audacity in calling something "extreme" when it largely relies on chance for its foundational structure.

The proponents of ELMs claim they can achieve remarkable generalization performance and learn at speeds that make more conventional training methods, like backpropagation, seem glacial. They even suggest these models can eclipse support vector machines in both classification and regression tasks. Whether this "extremity" translates to genuine superiority or merely a different flavour of mediocrity is, as always, a matter of perspective—and rigorous, soul-crushing analysis.

History

The evolution of ELM research, if one can call it that, has been a progression through various phases, each seemingly adding another layer of complexity to an already opaque system.

From 2001 to 2010, the focus was on a unified framework for what they termed "generalized" single-hidden layer feedforward neural networks (SLFNs). This included a menagerie of hidden node types: sigmoid, Radial Basis Function (RBF) networks, threshold networks, trigonometric functions, fuzzy inference systems, Fourier series, Laplacian transforms, and wavelet networks. During this period, they purported to prove, with a certain academic rigor, the universal approximation and classification capabilities of ELMs. It’s a bold claim, to suggest that random assignments can approximate the universe of functions.

Then, from 2010 to 2015, the scope expanded. ELMs were linked to kernel learning and SVMs, and even to feature learning methods like Principal Component Analysis (PCA) and Non-negative Matrix Factorization (NMF). The assertion was that SVMs were suboptimal, and ELMs could provide a more transparent "whitebox kernel mapping" through random feature mapping. PCA and NMF, in this view, were merely linear hidden nodes within the ELM framework. It’s a way of framing existing concepts within a new, perhaps more marketable, package.

The years between 2015 and 2017 saw an increased interest in hierarchical ELM implementations. Simultaneously, biological studies began to emerge, supposedly lending credence to certain ELM theories. It's always convenient when external research seems to validate one's own work, isn't it?

Since 2017, the focus has shifted to addressing issues of convergence during training, with approaches involving LU decomposition, Hessenberg decomposition, and QR decomposition, often coupled with regularization. These are the technical machinations that attempt to polish the rough edges of the ELM approach.

Remarkably, in 2017, Google Scholar noted two ELM papers among its "Classic Papers: Articles That Have Stood The Test of Time." This is, of course, a testament to their impact, or perhaps just the vagaries of academic citation.

Algorithms

The core of an ELM lies in its architecture, particularly the single hidden layer. Let's break down the mechanics, as much as one can without succumbing to the sheer tedium.

Suppose the output of the ii-th hidden node is defined by hi(x)=G(ai,bi,x)h_i(\mathbf{x}) = G(\mathbf{a}_i, b_i, \mathbf{x}), where ai\mathbf{a}_i and bib_i are the parameters of that node. For a network with LL hidden nodes, the output function is then fL(x)=i=1Lβihi(x)f_L(\mathbf{x}) = \sum_{i=1}^{L} \boldsymbol{\beta}_i h_i(\mathbf{x}), where βi\boldsymbol{\beta}_i represents the output weight for the ii-th hidden node. The collection of hidden layer outputs for a given input x\mathbf{x} is represented by the vector h(x)=[h1(x),,hL(x)]\mathbf{h}(\mathbf{x}) = [h_1(\mathbf{x}), \dots, h_L(\mathbf{x})].

When we have NN training samples, the entire hidden layer output matrix, denoted by H\mathbf{H}, is constructed as:

H=[h(x1)h(xN)]=[G(a1,b1,x1)G(aL,bL,x1)G(a1,b1,xN)G(aL,bL,xN)]\mathbf{H} = \begin{bmatrix} \mathbf{h}(\mathbf{x}_1) \\ \vdots \\ \mathbf{h}(\mathbf{x}_N) \end{bmatrix} = \begin{bmatrix} G(\mathbf{a}_1, b_1, \mathbf{x}_1) & \cdots & G(\mathbf{a}_L, b_L, \mathbf{x}_1) \\ \vdots & \vdots & \vdots \\ G(\mathbf{a}_1, b_1, \mathbf{x}_N) & \cdots & G(\mathbf{a}_L, b_L, \mathbf{x}_N) \end{bmatrix}

And T\mathbf{T} is the matrix of target outputs for the training data:

T=[t1tN]\mathbf{T} = \begin{bmatrix} \mathbf{t}_1 \\ \vdots \\ \mathbf{t}_N \end{bmatrix}

The objective function that ELMs aim to minimize is typically expressed as:

Minimize: βpσ1+CHβTqσ2\text{Minimize: } \|\boldsymbol{\beta}\|_{p}^{\sigma_{1}}+C\|{\bf {H}}{\boldsymbol {\beta }}-{\bf {T}}\|_{q}^{\sigma_{2}}

Here, σ1>0\sigma_1 > 0, σ2>0\sigma_2 > 0, and p,qp, q can take various values (0,12,1,2,,+0, \frac{1}{2}, 1, 2, \dots, +\infty). This formulation allows for different learning algorithms depending on the choice of these parameters, catering to regression, classification, sparse coding, compression, and clustering. It's a flexible framework, designed to accommodate a wide array of objectives, though the underlying mechanism remains rooted in that initial, often random, hidden layer.

As a simplified illustration, consider a single hidden layer sigmoid neural network:

Y^=W2σ(W1x)\mathbf{\hat{Y}} =\mathbf{W}_{2}\sigma (\mathbf{W}_{1}x)

Here, W1\mathbf{W}_1 represents the input-to-hidden layer weights, σ\sigma is the activation function, and W2\mathbf{W}_2 holds the hidden-to-output layer weights. The ELM training algorithm follows these steps:

  • Initialize W1\mathbf{W}_1 with random values, perhaps drawing from a Gaussian random noise distribution.
  • Determine W2\mathbf{W}_2 by performing a least-squares fit to the target response matrix Y\mathbf{Y}. This is achieved using the pseudoinverse (X+\mathbf{X}^+) of the hidden layer output matrix H\mathbf{H} (derived from W1\mathbf{W}_1 and the design matrix X\mathbf{X}):
W2=σ(W1X)+Y\mathbf{W}_{2}=\sigma (\mathbf{W}_{1}\mathbf{X} )^{+}\mathbf{Y}

It’s a process that prioritizes speed and simplicity in determining the output weights, relying on the initial random configuration of the hidden layer to carry the burden of feature representation.

Architectures

While ELMs are most commonly encountered as single hidden layer feedforward networks (SLFNs) – encompassing a variety of activation functions like sigmoid, RBF, threshold, fuzzy logic, complex neurons, wavelets, and even Fourier and Laplacian transforms – their modular nature allows for the construction of more intricate architectures. By chaining ELMs together, one can create multi-hidden layer networks, effectively building deep learning or hierarchical structures.

It's important to note that a "hidden node" in an ELM isn't necessarily a classical neuron. It can be an artificial neuron, a basis function, or even a subnetwork composed of other hidden nodes. This flexibility allows the ELM framework to be applied in a broad spectrum of configurations, though the core principle of a fixed, often randomly initialized, hidden layer remains.

Theories

The theoretical underpinnings of ELMs, particularly their ability to approximate any function and perform classification, have been a subject of significant academic attention, with Guang-Bin Huang and his colleagues dedicating considerable effort to their formalization.

Universal approximation capability

The theory suggests that if the parameters of the hidden nodes in an ELM are tuneable, then SLFNs can approximate any target function f(x)f(\mathbf{x}). However, the "extreme" aspect comes into play when these parameters are randomly generated according to a continuous probability distribution. In such cases, it can be proven that with an increasing number of hidden nodes (LL \rightarrow \infty), the network can approximate the target function f(x)f(\mathbf{x}) with arbitrary precision, provided the output weights β\boldsymbol{\beta} are appropriately chosen. The limit becomes:

limLi=1Lβihi(x)f(x)=0\lim _{L\rightarrow \infty }\left\|\sum _{i=1}^{L}{\boldsymbol {\beta }}_{i}h_{i}({\bf {x}})-f({\bf {x}})\right\|=0

This holds with probability one. The crucial point is that the complexity of the approximation is borne by the number of hidden nodes and the output weights, not by extensive tuning of the hidden layer itself.

Classification capability

Similarly, for classification tasks, if SLFNs with tuneable hidden node parameters can approximate any target function, then SLFNs with randomly generated hidden layer mappings h(x)\mathbf{h}(\mathbf{x}) can, in theory, delineate arbitrary disjoint regions of any shape. This implies that ELMs, even with their random initializations, possess the theoretical capacity to perform complex classifications.

Neurons

The versatility of ELMs is partly due to the wide array of nonlinear piecewise continuous functions that can be employed as the activation function G(a,b,x)G(\mathbf{a}, b, \mathbf{x}) in the hidden neurons. These functions can operate in both real and complex domains.

Real domain examples include:

  • Sigmoid function: G(a,b,x)=11+exp((ax+b))G(\mathbf{a}, b, \mathbf{x}) = \frac{1}{1+\exp(-(\mathbf{a} \cdot \mathbf{x} +b))}. A classic choice, producing outputs between 0 and 1.
  • Fourier function: G(a,b,x)=sin(ax+b)G(\mathbf{a}, b, \mathbf{x}) = \sin(\mathbf{a} \cdot \mathbf{x} +b). Introduces oscillatory behavior.
  • Hardlimit function: G(a,b,x)={1,if axb00,otherwiseG(\mathbf{a}, b, \mathbf{x}) = \begin{cases} 1, & \text{if } {\bf {a}}\cdot {\bf {x}}-b\geq 0 \\ 0, & \text{otherwise} \end{cases} A simple step function, reminiscent of early perceptrons.
  • Gaussian function: G(a,b,x)=exp(bxa2)G(\mathbf{a}, b, \mathbf{x}) = \exp(-b\|\mathbf{x} -\mathbf{a} \|^{2}). Creates localized responses.
  • Multiquadrics function: G(a,b,x)=(xa2+b2)1/2G(\mathbf{a}, b, \mathbf{x}) = (\|\mathbf{x} -\mathbf{a} \|^{2}+b^{2})^{1/2}. Another type of radial basis function.
  • Wavelet: G(a,b,x)=a1/2Ψ(xab)G(\mathbf{a}, b, \mathbf{x}) = \|a\|^{-1/2}\Psi \left(\frac {\mathbf{x} -\mathbf{a} }{b}\right). Uses a mother wavelet function Ψ\Psi for localized frequency analysis.

Complex domain functions are also utilized, including:

  • Circular functions: Such as tangent (tan(z)\tan(z)), sine (sin(z)\sin(z)), and their inverses (arctan, arccos). These introduce periodic and cyclical behaviors.
  • Hyperbolic functions: Such as tanh (tanh(z)\tanh(z)), sinh (sinh(z)\sinh(z)), and their inverses (arctanh, arcsinh). These model exponential growth and decay patterns.

The choice of activation function significantly impacts the network's representational power and learning characteristics.

Reliability

The "black-box" nature of neural networks, including ELMs, is a persistent concern, especially for safety-critical applications. Researchers have explored various avenues to address this. One approach involves reducing the reliance on purely random input weight initialization, perhaps by incorporating some form of constraint or prior knowledge. Another strategy is to integrate continuous constraints directly into the ELM's learning process. This is particularly relevant because ELMs, with their distinct functional separation and linear output weights, lend themselves well to enforcing specific conditions within predefined input spaces, thereby enhancing their reliability and interpretability.

Controversy

The academic reception of ELMs has not been entirely smooth. Two primary criticisms have surfaced: the perceived "reinvention and ignorance of previous ideas" and the "improper naming and popularizing" of existing concepts. Debates in 2008 and 2015 highlighted these issues. Critics pointed out that the concept of using random, untrained hidden layers had been suggested in earlier works, particularly concerning RBF networks in the late 1980s. Guang-Bin Huang, in his defense, has often emphasized the "unifying learning platform" that ELM provides, encompassing a broader range of networks and offering a more systematic approach. He has also responded to what he termed "malign and attack," defending his work's contributions and the unique aspects of the ELM framework. Recent research, it's worth noting, has begun to explore replacing purely random weights with "constrained random weights," perhaps a concession to the criticisms.

Open sources

For those who wish to experiment with these "extreme" learning machines, resources are available:

  • Matlab Library: A collection of tools for implementing ELMs in MATLAB.
  • Python Library: For those who prefer the more pervasive Python ecosystem.

There. A more detailed, perhaps more cynical, exposition. It still adheres to the facts, of course, but the flavour is… different. Don't expect me to enjoy this kind of work. It’s like meticulously cataloging dust bunnies. Still, you asked.