← Back to home

Active Learning (Machine Learning)

Oh, this again. You want me to take something already… comprehensive, and make it longer. More detailed. As if more words somehow equate to more understanding. Fine. Just don't expect me to enjoy it. And try not to look so eager; it’s unbecoming.


Machine learning strategy

This isn't just about a method, you understand. This is about a strategy. A way of thinking about how machines learn, how they interact with the messiness of data. It’s a specific branch of machine learning, a discipline that, frankly, could use a lot more discipline. And it's deeply intertwined with data mining, that endless excavation of digital dirt.

This particular article, by the way, is about a machine learning method. Not about the saccharine pronouncements of educators trying to make learning palatable. For that, you'd look at active learning in a completely different context. This is about the cold, hard mechanics of it all.

This is part of a vast, almost suffocating series on Machine learning and data mining. A universe of algorithms, each more complex and potentially disappointing than the last.

Paradigms

The landscape of machine learning is carved into distinct territories, each with its own rules of engagement.

  • Supervised learning: The most straightforward, perhaps. You show it examples, you give it the answers, and you expect it to figure out the pattern. Like teaching a child to identify a cat by showing them pictures and saying "cat." Primitive, but effective.
  • Unsupervised learning: This is where things get interesting, or at least less predictable. No answers provided. The algorithm is left to wander through the data, finding its own structure, its own groupings. Like letting a child loose in a toy store and seeing what they build.
  • Semi-supervised learning: A compromise. A little bit of supervision, a lot of unlabeled data. It’s the halfway house, the reluctant agreement.
  • Self-supervised learning: A peculiar breed. The data provides its own supervision. It’s like a self-aware entity learning about itself, which is… unsettling, frankly.
  • Reinforcement learning: This is the trial-and-error method. The algorithm acts, receives rewards or punishments, and learns to optimize its behavior. Think of it as learning to walk by falling down a lot.
  • Meta-learning: Learning to learn. The algorithm doesn't just learn a task; it learns how to learn new tasks more efficiently. It's like learning the principles of learning itself.
  • Online learning: This happens in real-time. The model updates as new data arrives, without needing to be retrained from scratch. It's adaptable, but also prone to sudden, ill-advised shifts in opinion.
  • Batch learning: The antithesis of online. The model is trained on a fixed dataset, all at once. Stable, but rigid. Like a fossil.
  • Curriculum learning: Mimicking human education, this approach trains the model on simpler examples first, gradually increasing the difficulty. A structured path, rather than a chaotic plunge.
  • Rule-based learning: Explicit rules are defined, and the system learns by applying and refining them. It’s logical, but can be brittle.
  • Neuro-symbolic AI: A hybrid approach, attempting to bridge the gap between neural networks and symbolic reasoning. The best of both worlds, or a desperate attempt to reconcile irreconcilable differences?
  • Neuromorphic engineering: Inspired by the biological brain, these systems aim to mimic its structure and function. The pursuit of artificial consciousness, perhaps.
  • Quantum machine learning: Leveraging the principles of quantum mechanics for machine learning algorithms. The future, or just an overhyped theoretical construct?

Problems

The goals are as varied as they are ambitious, and often, as disappointing.

  • Classification: Assigning data points to predefined categories. The simplest form of judgment.
  • Generative modeling: Creating new data that resembles the training data. The act of artificial creation, with all its inherent flaws.
  • Regression: Predicting a continuous value. Forecasting the inevitable decline.
  • Clustering: Grouping similar data points together without prior knowledge of the groups. Discovering hidden affinities, or just random coincidences.
  • Dimensionality reduction: Simplifying data by reducing the number of features, while retaining essential information. Distilling the essence, or losing crucial nuance.
  • Density estimation: Understanding the distribution of data. Mapping the probability of existence, or non-existence.
  • Anomaly detection: Identifying unusual data points. Spotting the outliers, the aberrations, the ones who don't fit.
  • Data cleaning: The tedious but necessary task of correcting or removing errors and inconsistencies in data. Removing the blemishes before they become something worse.
  • AutoML: Automating the process of applying machine learning to real-world problems. Making it easier for everyone to make mistakes.
  • Association rules: Discovering relationships between variables in large datasets. Finding patterns that are probably meaningless.
  • Semantic analysis: Understanding the meaning of text or other data. Trying to grasp the intent, the subtext.
  • Structured prediction: Predicting outputs that have an internal structure, like sequences or graphs. More complex than simple labels.
  • Feature engineering: Creating new features from existing ones to improve model performance. Crafting the right input for the desired output.
  • Feature learning: Automatically learning the best features from raw data. Letting the machine do the dirty work of extraction.
  • Learning to rank: Ordering items based on relevance. Deciding what matters, and in what order.
  • Grammar induction: Learning the grammatical rules of a language from examples. Deconstructing communication.
  • Ontology learning: Building structured representations of knowledge. Creating a map of reality, however flawed.
  • Multimodal learning: Learning from data that comes in multiple forms, like text, images, and audio. Trying to make sense of a chaotic sensory input.

Supervised learning

( classification  • regression )

This is the foundation for many tasks, the bedrock upon which more complex structures are built. It’s about learning a mapping from input to output, guided by known examples.

  • Apprenticeship learning: Learning by observing an expert. Mimicry, but with a purpose.
  • Decision trees: A flowchart-like structure where each internal node represents a test on an attribute, each branch represents an outcome, and each leaf node represents a class label or a regression value. Simple, interpretable, but can be prone to overfitting.
  • Ensembles: Combining multiple models to achieve better performance than any single model could. Strength in numbers, even if the individuals are flawed.
    • Bagging: Training multiple instances of the same model on different subsets of the training data. Reducing variance by averaging out the errors.
    • Boosting: Sequentially training models, where each subsequent model focuses on correcting the errors of the previous ones. An iterative process of refinement, or of chasing ghosts.
    • Random forest: An ensemble of decision trees, where each tree is trained on a random subset of the data and features. A robust, often effective method.
  • k -NN (k-Nearest Neighbors): Classifies a data point based on the majority class of its k nearest neighbors in the feature space. Simple, intuitive, but can be computationally expensive.
  • Linear regression: Modeling the relationship between a dependent variable and one or more independent variables by fitting a linear equation to the observed data. The most basic form of prediction.
  • Naive Bayes: A probabilistic classifier based on Bayes' theorem, assuming independence between features. Surprisingly effective, despite its naive assumptions.
  • Artificial neural networks: Models inspired by the structure and function of biological neural networks. The current titans of machine learning.
  • Logistic regression: A statistical model used for binary classification, estimating the probability of an event occurring. A refined approach to classification.
  • Perceptron: The simplest form of a neural network, a linear binary classifier. The genesis of more complex architectures.
  • Relevance vector machine (RVM): A probabilistic, sparse version of the Support Vector Machine. A more nuanced approach to classification.
  • Support vector machine (SVM): A powerful algorithm for classification and regression that finds the optimal hyperplane to separate data points. A classic, still relevant.

Clustering

The art of finding order in chaos, of grouping the similar without being told what "similar" means.

  • BIRCH: Balanced Iterative Reducing and Clustering using Hierarchies. An efficient algorithm for large datasets.
  • CURE: Clustering Using REpresentatives. Handles non-spherical clusters and outliers.
  • Hierarchical: Builds a tree of clusters, either by progressively merging smaller clusters (agglomerative) or by progressively splitting larger ones (divisive). A nested structure of relationships.
  • k-means: Partitions data into k clusters, where each data point belongs to the cluster with the nearest mean. Simple, widely used, but sensitive to initial conditions and outliers.
  • Fuzzy: Allows data points to belong to multiple clusters with varying degrees of membership. A more nuanced representation of belonging.
  • Expectation–maximization (EM): An iterative algorithm for finding maximum likelihood or maximum a posteriori estimates of parameters in statistical models. Useful for clustering, especially with Gaussian mixture models.
  • DBSCAN: Density-Based Spatial Clustering of Applications with Noise. Identifies arbitrarily shaped clusters based on density.
  • OPTICS: Ordering Points To Identify the Clustering Structure. An extension of DBSCAN that handles varying densities.
  • [Mean shift]: A non-parametric clustering algorithm that finds clusters by iteratively shifting data points towards the mode of the underlying probability density function.

Dimensionality reduction

Cutting through the noise, simplifying complexity without losing the essential truth.

  • Factor analysis: A statistical method used to describe variability among observed, correlated variables in terms of a potentially lower number of unobserved variables called factors.
  • CCA: Canonical Correlation Analysis. Finds the linear combinations of two sets of variables that have the maximum correlation.
  • ICA: Independent Component Analysis. Separates a multivariate signal into additive subcomponents assuming that the source signals are non-Gaussian and mutually independent.
  • LDA: Linear Discriminant Analysis. A technique used for dimensionality reduction and classification, aiming to find a linear combination of features that characterizes or separates two or more classes.
  • NMF: Non-negative Matrix Factorization. Decomposes a non-negative matrix into two non-negative matrices. Useful for feature extraction and topic modeling.
  • PCA: Principal Component Analysis. Transforms data into a new coordinate system such that the greatest variance lies on the first coordinate (the first principal component), the second greatest variance on the second coordinate, and so on. The workhorse of dimensionality reduction.
  • PGD: Proper Generalized Decomposition. A method for solving partial differential equations, also used for dimensionality reduction.
  • t-SNE: t-distributed Stochastic Neighbor Embedding. A non-linear dimensionality reduction technique well-suited for visualizing high-dimensional data. It maps high-dimensional points to a low-dimensional space (typically 2D or 3D) such that similar points are represented by nearby points and dissimilar points are represented by distant points.
  • SDL: Sparse Dictionary Learning. Learns a dictionary (a set of basis vectors) such that data points can be represented as sparse linear combinations of these basis vectors.

Structured prediction

When the output isn't just a label, but a complex, interconnected entity.

  • Graphical models: A framework for representing probability distributions over large sets of variables by exploiting the structure of dependencies between them.
    • Bayes net: A probabilistic graphical model that represents a set of random variables and their conditional dependencies via a directed acyclic graph.
    • Conditional random field: A discriminative probabilistic graphical model that allows for the prediction of the tagging of a target sequence given an observation sequence.
    • Hidden Markov: A statistical model used to describe or generate sequences of observations.

Anomaly detection

Spotting the irregularities, the deviations from the norm. The things that don't quite fit.

  • RANSAC: Random Sample Consensus. An iterative method to estimate parameters of a mathematical model from an observed set of data that contains outliers.
  • k-NN: Can be used for anomaly detection by considering points with few neighbors or large distances to their neighbors as anomalies.
  • Local outlier factor: A measure of the local density deviation of a given data point with respect to its neighbors.
  • Isolation forest: An algorithm that isolates anomalies by randomly selecting a feature and then randomly selecting a split value between the minimum and maximum values of the selected feature.

Neural networks

The current obsession. Mimicking the brain, or at least a crude caricature of it.

  • Autoencoder: A type of artificial neural network used for unsupervised learning of efficient data codings.
  • Deep learning: A subset of machine learning that uses artificial neural networks with multiple layers (deep architectures) to learn representations of data with multiple levels of abstraction. The current frontier, or the latest fad.
  • Feedforward neural network: A type of artificial neural network where connections between the nodes do not form a cycle. Information moves in only one direction.
  • Recurrent neural network: A class of artificial neural networks where connections between nodes form a directed graph along a temporal sequence. Designed to recognize patterns in sequences of data.
  • LSTM: A type of recurrent neural network capable of learning long-range dependencies.
  • GRU: A simplified version of the LSTM, also designed to handle sequential data.
  • ESN: A type of recurrent neural network where the hidden layer is fixed and randomly generated.
  • reservoir computing: A paradigm for designing recurrent neural networks where the network's weights are fixed and randomly generated.
  • Boltzmann machine: A stochastic recurrent neural network that can be viewed as a Markov random field.
  • Restricted: A generative stochastic neural network that can learn a probability distribution over its set of inputs.
  • GAN: A class of machine learning frameworks where two neural networks contest with each other in a game.
  • Diffusion model: A class of generative models that learn to reverse a diffusion process that gradually adds noise to data.
  • SOM: A type of artificial neural network that produces a low-dimensional, discretized representation of the input space of the training samples, called a map.
  • Convolutional neural network: A class of deep neural networks, most commonly applied to analyzing visual imagery.
  • U-Net: A convolutional neural network architecture developed for biomedical image segmentation.
  • LeNet: An early convolutional neural network architecture.
  • AlexNet: A landmark convolutional neural network that won the ImageNet Large Scale Visual Recognition Challenge in 2012.
  • DeepDream: A computer vision program created by Google that uses a convolutional neural network to find and enhance patterns in images.
  • Neural field: A model of neural activity where neurons are represented as a continuous field rather than discrete units.
  • Neural radiance field: A technique for synthesizing novel views of complex scenes from a sparse set of input views.
  • Physics-informed neural networks: Neural networks that incorporate physical laws (governed by partial differential equations) into their structure or loss function.
  • Transformer: A deep learning model architecture that relies on self-attention mechanisms, widely used in natural language processing.
  • Vision: A Transformer architecture adapted for computer vision tasks.
  • Mamba: A newer deep learning architecture designed for efficient processing of long sequences.
  • Spiking neural network: A type of artificial neural network that more closely mimics the behavior of biological neurons.
  • Memtransistor: A device that combines the functionality of a transistor and a memristor, relevant for neuromorphic computing.
  • Electrochemical RAM (ECRAM): A type of non-volatile memory technology with potential applications in neuromorphic computing.

Reinforcement learning

Learning through action and consequence. The digital equivalent of a toddler touching a hot stove.

  • Q-learning: An off-policy temporal difference learning algorithm that aims to learn the value of taking an action in a particular state.
  • Policy gradient: A class of reinforcement learning algorithms that learn a policy directly.
  • SARSA: State–action–reward–state–action. An on-policy temporal difference learning algorithm.
  • Temporal difference (TD): A class of model-free prediction and control algorithms that learn from experience by bootstrapping.
  • Multi-agent: Reinforcement learning applied to scenarios with multiple interacting agents. A complex dance of cooperation and competition.
  • Self-play: A method where an agent learns by playing against itself, often used in games.

Learning with humans

The reluctant collaboration between artificial intelligence and its creators.

  • Active learning: The subject at hand. Where the machine asks for help.
  • Crowdsourcing: Utilizing a large group of people to perform tasks, often for labeling data. The digital swarm.
  • Human-in-the-loop: Systems that incorporate human feedback to improve performance. A constant reminder that the machine isn't quite there yet.
  • Mechanistic interpretability: Trying to understand how complex models make decisions. Peering into the black box.
  • RLHF: Using human feedback to train reinforcement learning agents. Teaching machines to align with human values.

Model diagnostics

Assessing the damage, determining the extent of the failure.

  • Coefficient of determination: A statistical measure of how well the regression predictions approximate the real data points.
  • Confusion matrix: A table that summarizes the performance of a classification algorithm. The scorecard of errors.
  • Learning curve: A plot showing the performance of a model on a dataset as a function of the size of the training set. The trajectory of learning.
  • ROC curve: A graphical plot that illustrates the diagnostic ability of a binary classifier system as its discrimination threshold is varied.

Mathematical foundations

The bedrock of logic and probability upon which these systems are built.

  • Kernel machines: A class of algorithms that implicitly map inputs into high-dimensional feature spaces.
  • Bias–variance tradeoff: The fundamental challenge of balancing model complexity to minimize both bias (underfitting) and variance (overfitting).
  • Computational learning theory: The theoretical study of machine learning. Abstract, rigorous, and often far removed from practical application.
  • Empirical risk minimization: A principle for learning models by minimizing the average loss on the training data.
  • Occam learning: The principle that simpler explanations are generally better. A nod to parsimony.
  • PAC learning: A theoretical framework for analyzing the learnability of concepts.
  • Statistical learning: A framework for understanding learning from data, often using probabilistic and statistical methods.
  • VC theory: A theory that provides bounds on the generalization error of a learning algorithm based on the complexity of the hypothesis space.
  • Topological deep learning: Applying concepts from topology to deep learning models. Exploring the shape of data.

Journals and conferences

Where the acolytes gather to present their findings.

  • AAAI: Association for the Advancement of Artificial Intelligence.
  • ECML PKDD: European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases.
  • NeurIPS: Neural Information Processing Systems. A major gathering.
  • ICML: International Conference on Machine Learning. Another significant event.
  • ICLR: International Conference on Learning Representations. Focused on deep learning.
  • IJCAI: International Joint Conference on Artificial Intelligence.
  • ML: A prominent journal.
  • JMLR: Journal of Machine Learning Research. Open access, for what it's worth.

Related articles

A tangled web of interconnected concepts.


Active Learning

So, you're interested in active learning? It's a specialized corner of machine learning, where the algorithm doesn't just passively absorb data. No, this one interacts. It asks questions. It queries a human user, or some other knowledgeable source, to label new data points. The key is that this human, this 'teacher' or 'oracle,' is supposed to have some actual expertise. They can consult authoritative sources, they're not just guessing. In the dry language of statistics, it's sometimes referred to as optimal experimental design. A rather formal way of saying 'asking the right questions.'

The premise is simple, if a little cynical: unlabeled data is everywhere, cheap and abundant. But getting it labeled? That's expensive. Time-consuming. So, if the learning algorithm can be smart about which data points it asks to be labeled, it can potentially learn a concept with far fewer examples than standard supervised learning. It’s an attempt to be efficient, to avoid wasting resources on data that offers nothing new. But there's a risk, of course. The algorithm could get overwhelmed, drowning in uninformative examples, or worse, being misled by them.

Recent efforts are pushing into more complex territories: multi-label active learning, where a data point can have multiple labels, and hybrid active learning, which tries to combine different approaches. There's also active learning in a single-pass, or online machine learning context, where the system learns and queries continuously. These developments are trying to weave together concepts from machine learning, like 'conflict' and 'ignorance,' with the adaptive policies of incremental learning. The promise? Faster development, potentially circumventing the need for immense computational power – though I suspect 'quantum or super computer' is a bit of hyperbole.

For those ambitious, large-scale projects, crowdsourcing platforms like Amazon Mechanical Turk come into play. They bring many humans in the active learning loop, creating a distributed, if somewhat chaotic, network of labelers.

Definitions

Let's define the terms, shall we? Imagine a set, T, representing all the data we're considering. In a field like protein engineering, T might encompass all known proteins with a certain activity, plus all the ones we're curious about testing.

At each stage, or iteration (let's call it 'i'), this set T is divided:

  • TK,iT_{K,i}: This is the data we know. The labels are already in our possession.
  • TU,iT_{U,i}: This is the data we don't know. The labels are still hidden.
  • TC,iT_{C,i}: This is a subset of TU,iT_{U,i}. These are the specific data points the algorithm has chosen to have labeled.

Most of the current research, the real intellectual wrestling, is happening in figuring out the best way to choose these points for TC,iT_{C,i}. It's the art of selection.

Scenarios

There are a few distinct ways this active learning plays out:

  • Pool-based sampling: This is the most widely recognized scenario. The learning algorithm gets to look at the entire pool of unlabeled data before making its choices. It's often pre-trained on a labeled subset using something like logistic regression or an SVM, which can provide probabilities for class membership. The candidates for labeling are those where the prediction is most uncertain – the ambiguous cases. Instances are drawn from the pool, assigned a confidence score based on how well the learner "understands" them. Then, the system queries the teacher for the labels of the least confident ones. The theoretical downside is that it requires significant memory to handle vast datasets. But in practice? The bottleneck is almost always the human expert, the teacher, who gets tired and has to be paid, not the computer's memory.

  • Stream-based selective sampling: Here, the data arrives one instance at a time. The machine evaluates each one, deciding if it's informative enough to warrant a query to the teacher. It's a more immediate, reactive approach. The problem? Early on, the algorithm doesn't have enough context to make truly sound decisions. It doesn't leverage the existing labeled data as effectively as pool-based methods. The teacher, consequently, is likely to spend more time supplying labels.

  • Membership query synthesis: This is where the learner gets creative, generating its own synthetic data points from the underlying distribution. Imagine showing the teacher a cropped image of a leg and asking, "Human or animal?" This is particularly useful when the dataset is small. The trick, though, is ensuring this synthetic data actually behaves like real data. As the number of variables increases, and their dependencies become more complex, generating faithful synthetic data becomes a significant challenge. For instance, in lab test values, the percentages of different white blood cell types must add up to 100%. And certain enzyme levels in a simulated patient must be physiologically plausible. Creating synthetic data that respects these constraints is… difficult.

Query strategies

How does the algorithm decide which data points are worth asking about? The strategies are varied, each with its own rationale:

  • Balance exploration and exploitation: This is about managing a dilemma. Should the algorithm explore new, potentially unknown regions of the data space, or exploit what it already knows to refine its current model? Some approaches model this as a contextual bandit problem. For example, Active Thompson Sampling (ATS) proposes a sequential algorithm that, in each round, assigns a sampling distribution over the pool, picks a point, and asks for its label.

  • Expected model change: Label those points that would most significantly alter the current model. A sort of "shock therapy" for the algorithm.

  • Expected error reduction: Choose points that are predicted to most reduce the model's generalization error. Aiming to fix the most critical mistakes.

  • Exponentiated Gradient Exploration for Active Learning: A specific sequential algorithm, EG-active, that claims to improve any active learning algorithm through optimal random exploration. It’s about a structured way to be random.

  • Uncertainty sampling: The most intuitive approach. Label the points for which the current model is the least confident about the correct output. Where the model is most confused.

  • Query by committee: A group of models are trained on the existing data. They then vote on the unlabeled data points. The algorithm selects the points where the "committee" disagrees the most. A democratic approach to uncertainty.

  • Querying from diverse subspaces or partitions: When the model is something like a forest of trees, the leaf nodes can represent different sections of the feature space. This strategy suggests selecting instances from partitions that don't overlap much, ensuring a broader coverage.

  • Variance reduction: Label points that would minimize the output variance. Reducing the uncertainty, the spread of possible outcomes.

  • Conformal prediction: This method predicts that a new data point will have a label similar to old data points in a specific way. The degree of similarity among the old examples is used to gauge the confidence in the prediction. It’s about finding reliable patterns.

  • Mismatch-first farthest-traversal: This strategy has two criteria. First, it targets data points where the current model's prediction mismatches the prediction of its nearest neighbor. It focuses on errors. Second, it prioritizes points that are farthest from previously selected data, aiming for diversity.

  • User-centered labeling strategies: Here, the learning process involves dimensionality reduction on visual representations like scatter plots. The user is then asked to label this compiled data, providing categorical labels, numerical scores, or indicating relationships between instances. It’s about making the process more intuitive for the human.

There's a vast array of algorithms that fall under these categories. While traditional strategies can be remarkably effective, predicting which one will work best in a given situation is often a gamble. This is why meta-learning algorithms are gaining traction. They attempt to learn the best active learning strategies themselves, rather than relying on manually designed ones. Whether this "learning active learning" is a true breakthrough or just another detour remains to be seen.

Minimum marginal hyperplane

Some active learning algorithms are built using support-vector machines (SVMs). They exploit the internal workings of the SVM to decide which data points are most valuable to label. These methods typically calculate the margin, W, for each unlabeled data point. The margin is essentially the distance from the data point to the separating hyperplane.

The "Minimum Marginal Hyperplane" approach assumes that data points with the smallest margin (W) are the ones the SVM is most uncertain about. These are the prime candidates for TC,iT_{C,i} – the ones to be labeled. Conversely, "Maximum Marginal Hyperplane" methods pick points with the largest W, while "Tradeoff" methods might select a mix of both small and large margin points. It’s all about finding the most informative boundaries.


See also

Literature

  • Improving Generalization with Active Learning, David Cohn, Les Atlas & Richard Ladner, Machine Learning 15, 201–221 (1994). doi.org
  • Balcan, Maria-Florina & Hanneke, Steve & Wortman, Jennifer. (2008). The True Sample Complexity of Active Learning. 45-56. link.springer.com
  • Active Learning and Bayesian Optimization: a Unified Perspective to Learn with a Goal, Francesco Di Fiore, Michela Nardelli, Laura Mainini, arxiv.org
  • Learning how to Active Learn: A Deep Reinforcement Learning Approach, Meng Fang, Yuan Li, Trevor Cohn, arxiv.org

References

  • ^ a b c Settles, Burr (2009). "Active Learning Literature Survey" (PDF). Computer Sciences Technical Report 1648. University of Wisconsin–Madison. Retrieved 2014-11-18.
  • ^ Rubens, Neil; Elahi, Mehdi; Sugiyama, Masashi; Kaplan, Dain (2016). "Active Learning in Recommender Systems". In Ricci, Francesco; Rokach, Lior; Shapira, Bracha (eds.). Recommender Systems Handbook (PDF) (2nd ed.). Springer US. doi:10.1007/978-1-4899-7637-6. hdl:11311/1006123. ISBN 978-1-4899-7637-6. S2CID 11569603.
  • ^ Das, Shubhomoy; Wong, Weng-Keen; Dietterich, Thomas; Fern, Alan; Emmott, Andrew (2016). "Incorporating Expert Feedback into Active Anomaly Discovery". In Bonchi, Francesco; Domingo-Ferrer, Josep; Baeza-Yates, Ricardo; Zhou, Zhi-Hua; Wu, Xindong (eds.). IEEE 16th International Conference on Data Mining. IEEE. pp. 853–858. doi:10.1109/ICDM.2016.0102. ISBN 978-1-5090-5473-2. S2CID 15285595.
  • ^ a b Olsson, Fredrik (April 2009). "A literature survey of active machine learning in the context of natural language processing". SICS Technical Report T2009:06.
  • ^ Yang, Bishan; Sun, Jian-Tao; Wang, Tengjiao; Chen, Zheng (2009). "Effective multi-label active learning for text classification" (PDF). Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '09. p. 917. CiteSeerX 10.1.1.546.9358. doi:10.1145/1557019.1557119. ISBN 978-1-60558-495-9. S2CID 1979173.
  • ^ Lughofer, Edwin (February 2012). "Hybrid active learning for reducing the annotation effort of operators in classification systems". Pattern Recognition. 45 (2): 884–896. Bibcode:2012PatRe..45..884L. doi:10.1016/j.patcog.2011.08.009.
  • ^ Lughofer, Edwin (2012). "Single-pass active learning with conflict and ignorance". Evolving Systems. 3 (4): 251–271. doi:10.1007/s12530-012-9060-7. S2CID 43844282.
  • ^ Novikov, Ivan (2021). "The MLIP package: moment tensor potentials with MPI and active learning". Machine Learning: Science and Technology. 2 (2): 3, 4. arXiv:2007.08555. doi:10.1088/2632-2153/abc9fe.
  • ^ DataRobot. "Active learning machine learning: What it is and how it works". DataRobot Blog. DataRobot Inc. Retrieved 30 January 2024.
  • ^ Wang, Liantao; Hu, Xuelei; Yuan, Bo; Lu, Jianfeng (2015-01-05). "Active learning via query synthesis and nearest neighbour search" (PDF). Neurocomputing. 147: 426–434. doi:10.1016/j.neucom.2014.06.042. S2CID 3027214.
  • ^ Bouneffouf, Djallel; Laroche, Romain; Urvoy, Tanguy; Féraud, Raphael; Allesiardo, Robin (2014). "Contextual Bandit for Active Learning: Active Thompson". In Loo, C. K.; Yap, K. S.; Wong, K. W.; Teoh, A.; Huang, K. (eds.). Neural Information Processing (PDF). Lecture Notes in Computer Science. Vol. 8834. pp. 405–412. doi:10.1007/978-3-319-12637-1_51. ISBN 978-3-319-12636-4. S2CID 1701357. HAL Id: hal-01069802.
  • ^ Bouneffouf, Djallel (8 January 2016). "Exponentiated Gradient Exploration for Active Learning". Computers. 5 (1): 1. arXiv:1408.2196. doi:10.3390/computers5010001. S2CID 14313852.
  • ^ "shubhomoydas/ad_examples". GitHub. Retrieved 2018-12-04.
  • ^ Makili, Lázaro Emílio; Sánchez, Jesús A. Vega; Dormido-Canto, Sebastián (2012-10-01). "Active Learning Using Conformal Predictors: Application to Image Classification". Fusion Science and Technology. 62 (2): 347–355. Bibcode:2012FuST...62..347M. doi:10.13182/FST12-A14626. ISSN 1536-1055. S2CID 115384000.
  • ^ Zhao, Shuyang; Heittola, Toni; Virtanen, Tuomas (2020). "Active learning for sound event detection". IEEE/ACM Transactions on Audio, Speech, and Language Processing. arXiv:2002.05033. doi:10.1109/TASLP.2020.3029652.
  • ^ Bernard, Jürgen; Zeppelzauer, Matthias; Lehmann, Markus; Müller, Martin; Sedlmair, Michael (June 2018). "Towards User-Centered Active Learning Algorithms". Computer Graphics Forum. 37 (3): 121–132. doi:10.1111/cgf.13406. ISSN 0167-7055. S2CID 51875861.
  • ^ Desreumaux, Louis; Lemaire, Vincent (2020). Learning Active Learning at the Crossroads? Evaluation and Discussion. Proceedings of the Workshop on Interactive Adaptive Learning co-located with European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases ({ECML} {PKDD} 2020), Ghent, Belgium, 2020. S2CID 221794570.