- 1. Overview
- 2. Etymology
- 3. Cultural Impact
Alright, let’s dissect this. You want a Wikipedia article, rewritten. In my style. Which means it’ll be sharp, precise, and probably a little⦠unenthusiastic about the whole affair. Don’t expect sunshine and rainbows. Expect clarity, with a side of existential weariness. And all the internal links, precisely as they are. No deviations. Let’s get this over with.
AI That Learns Decision Rules From Data
This entry is a segment within the broader landscapes of Machine learning and data mining . Itās about systems that donāt just follow instructions, but actually learn how to make choices, deriving those choices from the data you shove at them. Itās less about programming and more about⦠observation. And if you think thatās easy, you havenāt seen enough data.
Paradigms
The way these systems learn is varied, like a poorly curated art exhibition. Each approach has its own peculiar logic, its own brand of “intelligence.”
- Supervised learning : This is the one where you show it examples, label them, and expect it to figure out the pattern. Like teaching a child, but with less whining and more complex algorithms. You provide the answers, it finds the rules. Simple, in theory.
- Unsupervised learning : Here, you just dump data on it and tell it to make sense of it. No labels, no guidance. It’s like asking someone to categorize a pile of discarded memories. Expect chaos, or perhaps, a disturbing order.
- Semi-supervised learning : A compromise. A bit of labeled data, a lot of unlabeled. Itās for when you have some answers, but not enough to be truly confident. A half-measure, much like most human endeavors.
- Self-supervised learning : This is where the data teaches itself, in a way. It creates its own labels from the data itself, often by predicting parts of the input. Itās like a narcissistic AI, fascinated by its own reflection.
- Reinforcement learning : This one learns by trial and error. It performs actions, gets rewards or penalties, and adjusts its strategy. Think of it as a highly sophisticated, eternally patient gambler.
- Meta-learning : Learning to learn. Itās not about mastering a task, but about mastering the process of learning itself. Itās the AI equivalent of an academic, always studying how to study better.
- Online learning : This system learns incrementally, as new data arrives. It doesn’t need a massive batch to process; it adapts on the fly. Like a street artist, constantly updating their work.
- Batch learning : The opposite of online. It processes all available data at once to train the model. Itās the methodical, thorough approach. Less agile, more⦠definitive.
- Curriculum learning : This is where the AI is trained on data in a specific, ordered way, like a student progressing through a curriculum. Start simple, build complexity. Itās about making the learning process less overwhelming.
- Rule-based learning: The core of what we’re discussing. Itās about deriving explicit rules from data. Simple, understandable logic. Or so they claim.
- Neuro-symbolic AI : A hybrid. It attempts to combine the pattern-recognition strengths of neural networks with the reasoning capabilities of symbolic AI. Like trying to get a poet to write a legal brief.
- Neuromorphic engineering : Designing hardware that mimics the structure and function of the biological brain. Itās about building thinking machines from the ground up, rather than just programming them.
- Quantum machine learning : Leveraging quantum mechanics for machine learning algorithms. Itās theoretical, abstract, and probably beyond your immediate comprehension.
Problems
These are the tasks these learning systems are set to tackle. The challenges they face, the puzzles theyāre meant to solve.
- Classification : Assigning data points to predefined categories. Sorting the wheat from the chaff, or more likely, the useful data from the noise.
- Generative modeling : Creating new data that resembles the training data. It’s about learning the underlying distribution and then sampling from it. Like an artist learning to paint in the style of another.
- Regression : Predicting a continuous numerical value. Forecasting the weather, predicting stock prices. The illusion of certainty.
- Clustering : Grouping similar data points together without prior labels. Finding patterns in the chaos, revealing hidden structures.
- Dimensionality reduction : Simplifying data by reducing the number of variables, while retaining essential information. Making complex things comprehensible, or at least, less cumbersome.
- Density estimation : Estimating the probability distribution of data. Understanding the likelihood of certain events.
- Anomaly detection : Identifying unusual data points that deviate from the norm. Spotting the outliers, the errors, the threats.
- Data cleaning : Preparing raw data for analysis by detecting and correcting errors or inconsistencies. The tedious, thankless task of tidying up.
- AutoML : Automating the process of applying machine learning to real-world problems. Making it easier for the less⦠dedicated.
- Association rules : Discovering interesting relationships between variables in large datasets. The “people who buy this also buy that” kind of insight.
- Semantic analysis : Understanding the meaning of text or other data. Getting to the heart of the matter, or at least, its digital representation.
- Structured prediction : Predicting structured outputs, like sequences or graphs, rather than single labels. More complex relationships, more complex predictions.
- Feature engineering : Creating new input features from existing data to improve model performance. The art of making data more palatable for the algorithm.
- Feature learning : Automatically discovering the relevant features from raw data. Letting the machine do the heavy lifting of selection.
- Learning to rank : Training models to order items based on their relevance. The science behind search results and recommendations.
- Grammar induction : Learning the grammatical rules of a language from examples. The AI as a linguist.
- Ontology learning : Automatically creating or extending ontologies. Building structured representations of knowledge.
- Multimodal learning : Learning from data that combines multiple types of information, like text and images. Integrating different senses, digitally.
Supervised learning
This is where the machine is shown examples with correct answers. Think of it as a guided tour through the data.
- ( classification  ⢠regression )
- Apprenticeship learning : Learning by observing an expert. Mimicking behavior.
- Decision trees : Hierarchical structures of decisions. Like a flowchart of fate.
- Ensembles
: Combining multiple models to improve performance. Strength in numbers, or perhaps, shared delusion.
- Bagging : Training models on different subsets of data.
- Boosting : Sequentially training models, with each new model correcting the errors of the previous ones. A relentless pursuit of accuracy.
- Random forest : An ensemble of decision trees. More trees, more decisions, more⦠complexity.
- k-NN: K-Nearest Neighbors. Classifies based on the majority class of its nearest neighbors. Simple, intuitive, and often effective.
- Linear regression : Modeling relationships with a straight line. The most basic form of prediction.
- Naive Bayes : A probabilistic classifier based on Bayes’ theorem. Assumes independence between features, which is rarely true, but often works.
- Artificial neural networks : Inspired by the structure of the human brain. Complex, powerful, and often opaque.
- Logistic regression : Used for binary classification. Predicts the probability of an instance belonging to a class.
- Perceptron : The simplest form of an artificial neural network. A single-layer neural network.
- Relevance vector machine (RVM) : A probabilistic approach similar to Support Vector Machines, but with a sparse solution.
- Support vector machine (SVM) : Finds the optimal hyperplane to separate data points into classes. Elegant and powerful.
Clustering
Grouping data without prior knowledge. Finding the inherent structure.
- BIRCH : Balanced Iterative Reducing and Clustering using Hierarchies. Efficient for large datasets.
- CURE : Clustering Using REpresentatives. Handles non-spherical clusters and outliers.
- Hierarchical : Creates a tree of clusters. Offers a view of data at different levels of granularity.
- k-means: Partitions data into k clusters. Simple, fast, but sensitive to initial centroids.
- Fuzzy : Allows data points to belong to multiple clusters with varying degrees of membership. More nuanced than hard clustering.
- Expectationāmaximization (EM) : An iterative algorithm for finding maximum likelihood estimates of parameters in statistical models. Used for clustering when distributions are unknown.
- DBSCAN : Density-Based Spatial Clustering of Applications with Noise. Finds arbitrarily shaped clusters and identifies outliers.
- OPTICS : Ordering Points To Identify the Clustering Structure. An extension of DBSCAN, producing a reachability plot.
- Mean shift : A non-parametric clustering algorithm that finds modes of the data density.
Dimensionality reduction
Simplifying data by reducing the number of variables. Making the complex manageable.
- Factor analysis : Identifies underlying latent variables that explain correlations among observed variables.
- CCA : Canonical Correlation Analysis. Finds relationships between two sets of variables.
- ICA : Independent Component Analysis. Separates a multivariate signal into additive subcomponents assuming non-Gaussianity and mutual independence.
- LDA : Linear Discriminant Analysis. Used for dimensionality reduction and classification, maximizing class separability.
- NMF : Non-negative Matrix Factorization. Decomposes a matrix into two non-negative matrices.
- PCA : Principal Component Analysis. Transforms data into a new coordinate system where the greatest variances lie on the first few components. The workhorse of dimensionality reduction.
- PGD : Proper Generalized Decomposition. A tensor decomposition method.
- t-SNE : t-Distributed Stochastic Neighbor Embedding. Primarily used for visualizing high-dimensional data.
- SDL : Sparse Dictionary Learning. Learns a dictionary of basis vectors for sparse representation.
Structured prediction
Predicting outputs that have internal structure, like sequences or graphs.
- Graphical models
: Representing complex probability distributions using graphs.
- Bayes net : A directed acyclic graph representing probabilistic relationships between variables.
- Conditional random field : A discriminative model for labeling or segmenting sequential data.
- Hidden Markov : A statistical model where the system being modeled is assumed to be a Markov process with unobserved (hidden) states.
Anomaly detection
Spotting the outliers, the things that don’t fit.
- RANSAC : Random Sample Consensus. An iterative method to estimate parameters of a mathematical model from an observed set of data that contains outliers.
- k-NN: Can be used to detect anomalies based on the distance to nearest neighbors.
- Local outlier factor : Measures the local density deviation of a data point with respect to its neighbors.
- Isolation forest : An algorithm that isolates anomalies by randomly partitioning the data. Anomalies are easier to isolate.
Neural networks
These are the complex, layered structures that mimic biological neurons. They are the foundation of much of modern AI.
- Autoencoder : A type of neural network trained to reconstruct its input. Used for dimensionality reduction and feature learning.
- Deep learning : Neural networks with multiple layers. Capable of learning complex hierarchical representations.
- Feedforward neural network : Information flows in one direction, from input to output, without cycles. The simplest form.
- Recurrent neural network
: Networks with feedback loops, allowing them to process sequential data and maintain a “memory” of past inputs.
- LSTM : A specialized type of RNN capable of learning long-term dependencies.
- GRU : A simpler variant of LSTM, also effective for sequential data.
- ESN : A type of RNN where only the output weights are trained.
- reservoir computing : A paradigm for training RNNs where the internal weights are fixed and random.
- Boltzmann machine
: A stochastic recurrent neural network that can learn a probability distribution over its set of inputs.
- Restricted : A simplified version of the Boltzmann machine with constraints on connections, making training more efficient.
- GAN : A framework where two neural networks (a generator and a discriminator) compete, leading to the generation of realistic data.
- Diffusion model : A class of generative models that work by gradually adding noise to data and then learning to reverse the process.
- SOM : Self-Organizing Map. A type of artificial neural network that produces a low-dimensional discretized representation of the input space of the training samples, analogous to a map.
- Convolutional neural network
: Particularly effective for image recognition, using convolutional layers to detect spatial hierarchies of features.
- U-Net : A convolutional network architecture designed for biomedical image segmentation.
- LeNet : An early and influential CNN architecture.
- AlexNet : A landmark CNN that won the ImageNet competition in 2012.
- DeepDream : A visualization technique that enhances patterns in images, leading to surreal and dreamlike outputs.
- Neural field : A continuous version of neural networks, often used in modeling large-scale neural activity.
- Neural radiance field : A technique for synthesizing novel views of complex 3D scenes from a sparse set of input views.
- Physics-informed neural networks : Neural networks that incorporate physical laws into their training process.
- Transformer : An architecture that relies heavily on self-attention mechanisms, revolutionizing natural language processing and increasingly used in other domains.
- Spiking neural network : Models that mimic the temporal dynamics of biological neurons more closely.
- Memtransistor : A device that combines memory and transistor functionality, potentially enabling more efficient neuromorphic hardware.
- Electrochemical RAM (ECRAM): A type of memory technology inspired by biological synapses, for use in neuromorphic computing.
Reinforcement learning
Learning through interaction and feedback. The AI as an agent exploring an environment.
- Q-learning : An off-policy algorithm that learns the value of taking an action in a particular state.
- Policy gradient : Directly optimizes the policy, which dictates the actions to take in given states.
- SARSA : StateāActionāRewardāStateāAction. An on-policy algorithm for learning.
- Temporal difference (TD) : A class of model-free prediction and control algorithms that learn from experience by bootstrapping.
- Multi-agent : Reinforcement learning involving multiple agents interacting in an environment.
- Self-play : Training an agent by having it play against itself, a technique crucial for games like Go and Chess.
Learning with humans
When the human element is integrated into the learning process.
- Active learning : The algorithm interactively queries the user (or other information source) to label new data points.
- Crowdsourcing : Using a large group of people to perform tasks, often for data labeling or annotation.
- Human-in-the-loop : Systems where human oversight and intervention are integral to the learning process.
- Mechanistic interpretability : Efforts to understand the internal workings of complex models, particularly neural networks.
- RLHF : Reinforcement learning guided by human preferences, used to align AI behavior with human values.
Model diagnostics
Assessing the performance and behavior of trained models.
- Coefficient of determination : A statistical measure of how well the regression predictions approximate the real data points.
- Confusion matrix : A table summarizing classification results, showing true positives, true negatives, false positives, and false negatives.
- Learning curve : Plots model performance against training set size, revealing issues like overfitting or underfitting.
- ROC curve : A plot showing the diagnostic ability of a binary classifier system as its discrimination threshold is varied.
Mathematical foundations
The theoretical underpinnings of machine learning. The abstract principles that govern how it all works.
- Kernel machines : A class of algorithms that implicitly map inputs into high-dimensional feature spaces.
- Biasāvariance tradeoff : The fundamental tension between a model’s ability to fit the training data (low bias) and its ability to generalize to unseen data (low variance).
- Computational learning theory : The theoretical study of machine learning from a computational perspective.
- Empirical risk minimization : A principle for learning models by minimizing the average loss on the training data.
- Occam learning : The principle that simpler explanations are generally preferable to more complex ones.
- PAC learning : A theoretical framework for analyzing the learnability of concepts.
- Statistical learning : A framework for statistical inference and learning from data.
- VC theory : A theory that provides bounds on the generalization error of a learning algorithm.
- Topological deep learning : Applying concepts from topology to deep learning.
Journals and conferences
Where the research is published and presented. The battlegrounds of ideas.
- AAAI : Association for the Advancement of Artificial Intelligence.
- ECML PKDD : European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases.
- NeurIPS : Neural Information Processing Systems. A premier conference for machine learning and computational neuroscience.
- ICML : International Conference on Machine Learning.
- ICLR : International Conference on Learning Representations. Focuses on deep learning.
- IJCAI : International Joint Conference on Artificial Intelligence.
- ML : Machine Learning, a prominent journal.
- JMLR : Journal of Machine Learning Research. An open-access journal.
Related articles
Other areas that touch upon this subject. A tangled web of concepts.
- Glossary of artificial intelligence
- List of datasets for machine-learning research
- List of datasets in computer vision and image processing
- Outline of machine learning
Rule-based Machine Learning (RBML)
Rule-based machine learning, or RBML, is a rather specific corner of computer science . Itās not about nebulous pattern recognition; it’s about identifying, learning, or even evolving explicit ‘rules’ that the system can then use to store, manipulate, or apply information. The defining characteristic, the very essence of RBML, is the extraction and deployment of a set of relational rules. These rules, working in concert, are meant to embody the knowledge the system has gleaned.
Approaches within RBML are varied, like a collection of cryptic notes. They include learning classifier systems , which are quite robust, the ever-popular association rule learning that uncovers hidden relationships, and even artificial immune systems , which draw inspiration from biological defenses. Essentially, any method that relies on a set of rules, each covering specific contextual knowledge, falls under this umbrella.
Now, while RBML is, conceptually, a type of rule-based system, itās crucial to distinguish it from the traditional, often hand-crafted rule-based systems . Those are usually built by humans, painstakingly assembling rules based on their own domain knowledge . RBML, on the other hand, employs learning algorithmsāthink of Rough sets theory for example [7]āto discover these rules automatically. It identifies and minimizes the set of features, then extracts useful rules without requiring a human to manually curate the entire knowledge base. Itās about letting the data speak, rather than imposing our own interpretations.
Rules
Rules, in this context, typically manifest as an ‘{IF:THEN} expression’. Itās a conditional statement: {IF ‘condition’ THEN ‘result’}. A more concrete, if somewhat simplistic, example might be {IF ‘red’ AND ‘octagon’ THEN ‘stop-sign’}. An individual rule, in isolation, isn’t a complete model. It only applies when its specific condition is met. Consequently, RBML methods usually comprise a set of these rules, forming a knowledge base . This collection of rules collectively constitutes the prediction model, often referred to as the decision algorithm. The interpretation of these rules can vary significantly, depending on the domain, the nature of the data (whether discrete or continuous), and how they are combined.
RIPPER
One notable algorithm in this space is RIPPER, which stands for Repeated Incremental Pruning to Produce Error Reduction. Developed by William W. Cohen, itās an optimized iteration of an earlier algorithm, IREP [8]. It’s designed to efficiently learn propositional rules, meaning rules that operate on propositions rather than quantified variables.
See also
For those who wish to delve deeper into the labyrinth of related concepts, a comprehensive list follows:
- Learning classifier system
- Association rule learning
- Associative classifier
- Artificial immune system
- Expert system
- Decision rule
- Rule induction
- Inductive logic programming
- Rule-based machine translation
- Genetic algorithm
- Rule-based system
- Rule-based programming
- RuleML
- Production rule system
- Business rule engine
- Business rule management system