QUICK FACTS
Created Jan 0001
Status Verified Sarcastic
Type Existential Dread
history of speech recognition, gottfried wilhelm leibniz, rené descartes, paper tape, peter troyanskii, esperanto, alan turing

History Of Natural Language Processing

“This article needs to be updated. Please help update this article to reflect recent events or newly available information. (April...”

Contents
  • 1. Overview
  • 2. Etymology
  • 3. Cultural Impact

This article needs to be updated. Please help update this article to reflect recent events or newly available information. (April 2023)

The ceaseless, often repetitive, saga of machines attempting to grasp the nuances of human communication is chronicled within the history of natural language processing . This particular narrative, like a tangled thread, inevitably overlaps with the equally convoluted tales found in the history of machine translation , the history of speech recognition , and, of course, the grand, overarching history of artificial intelligence . One might say it’s all part of humanity’s persistent, and frankly, rather exhausting, quest to build something that can talk back without judgment.

Early history

The notion of machines deciphering and transforming human language isn’t a modern conceit; it’s a persistent, almost ancient, fantasy. The earliest theoretical musings on what we might now call machine translation can be traced back to the rather optimistic seventeenth century. Visionary philosophers, minds like Gottfried Wilhelm Leibniz and René Descartes , toyed with ambitious proposals for universal codes. These were grand, theoretical frameworks designed to establish direct, unambiguous relationships between words across different languages. One might imagine them sketching out elegant, albeit utterly impractical, blueprints for linguistic harmony. However, despite their intellectual prowess, these lofty proposals remained firmly in the realm of the abstract. Not a single one ever coalesced into anything resembling a functional machine; they were intellectual exercises, not engineering blueprints.

Centuries later, as the world lumbered towards the mid-1930s, the first actual patents for these fantastical “translating machines” began to emerge. It seems the dream, though dormant, was never truly extinguished. One such proposal, put forth by a certain Georges Artsrouni, was commendably straightforward, if somewhat underwhelming in its ambition: an automatic bilingual dictionary. This rather rudimentary device was designed to operate using nothing more complex than paper tape , a testament to the technological constraints of its era. It was less a translator and more a glorified, automated lookup tool.

However, a far more intricate and ambitious vision materialized from the mind of a Russian pioneer, Peter Troyanskii . Troyanskii’s proposal wasn’t content with mere word-for-word substitution. His detailed design encompassed not only the foundational bilingual dictionary but also incorporated a sophisticated method for navigating the treacherous waters of grammatical roles and structural differences between languages. His approach, perhaps surprisingly, was rooted in the principles of Esperanto , the constructed international auxiliary language, which offered a degree of grammatical regularity that natural languages notoriously lack. This early reliance on a simplified, logical linguistic structure foreshadowed many of the rule-based systems that would dominate future decades of NLP research. [1] [2]

Logical period

The mid-20th century ushered in an era where the logical underpinnings of computation and language began to intertwine more explicitly. In 1950, Alan Turing , a name now synonymous with the very concept of machine intelligence, published his seminal article, “Computing Machinery and Intelligence ”. Within its pages, he introduced what would become the enduring benchmark for artificial intelligence: the Turing test . This criterion, deceptively simple yet profoundly challenging, posited that true intelligence could be attributed to a machine if a human judge, engaged in a real-time written conversation, could not reliably distinguish between the machine’s responses and those of an actual human being. The test, in essence, shifted the focus from how a machine thinks to how convincingly it can pretend to think. A fascinating parlour trick, if you ask me, but hardly a measure of genuine sentience.

Seven years later, in 1957, the field of linguistics itself was irrevocably altered by Noam Chomsky ’s groundbreaking work, “Syntactic Structures ”. Chomsky introduced the concept of ‘universal grammar ’, proposing an innate, rule-based system of syntactic structures common to all human languages. This theory profoundly influenced subsequent NLP research, steering it towards rule-based parsing and generation systems, with the belief that language could be deconstructed and reconstructed through a finite set of elegant rules. [3]

The mid-1950s also saw a brief, dazzling flash of optimism in the realm of machine translation. The Georgetown experiment in 1954, a collaborative effort, showcased the fully automatic translation of over sixty Russian sentences into English. The researchers, perhaps intoxicated by their preliminary success, boldly declared that the problem of machine translation would be “solved” within three to five years. [4] Such pronouncements, as history repeatedly demonstrates, are rarely met with the swift gratification promised. Real progress, as it invariably does, proved far more glacial. The subsequent ALPAC report in 1966 delivered a sobering dose of reality, concluding that a decade of intensive research had failed spectacularly to deliver on those initial, grandiose expectations. The report’s damning assessment led to a dramatic and precipitous reduction in funding for machine translation research, effectively sending the field into a protracted hibernation. Little substantial work in machine translation was undertaken for the next two decades, until the late 1980s, when a new paradigm, statistical machine translation systems, began to tentatively emerge from the ashes of earlier failures.

Despite these setbacks, the 1960s did yield some notably successful, albeit highly constrained, NLP systems. Among these was SHRDLU , a natural language system that operated within severely restricted “blocks worlds ” and possessed an equally restricted vocabulary. SHRDLU could understand and execute commands related to manipulating virtual blocks, demonstrating an impressive, if narrow, grasp of context and instruction. It worked extremely well within its tiny, carefully constructed universe, much like a perfectly trained pet in a very small cage.

The theoretical landscape continued to evolve. In 1969, Roger Schank introduced his conceptual dependency theory for natural language understanding. [5] This model, which drew partial inspiration from the work of Sydney Lamb , proposed that language could be represented using a small set of primitive conceptual actions and states, allowing for deeper semantic understanding beyond mere syntactic parsing. Schank’s students at Yale University , including prominent figures like Robert Wilensky, Wendy Lehnert, and Janet Kolodner , extensively adopted and developed this model, applying it to various aspects of language comprehension and generation.

The 1970s saw further architectural innovations. In 1970, William A. Woods unveiled the augmented transition network (ATN) as a powerful mechanism for representing natural language input. [6] Moving beyond rigid phrase structure rules , ATNs utilized an equivalent set of finite-state automata that could call upon themselves recursively, offering greater flexibility and power in parsing complex sentence structures. Both ATNs and their more generalized successors, “generalized ATNs,” remained prominent tools for natural language processing for a considerable period. Throughout this decade, a significant trend emerged among programmers: the development of ‘conceptual ontologies’. These were intricate structures designed to organize real-world information into a format that computers could genuinely “understand” and process. Notable examples from this period include MARGIE (Schank, 1975), SAM (Cullingford, 1978), PAM (Wilensky, 1978), TaleSpin (Meehan, 1976), QUALM (Lehnert, 1977), Politics (Carbonell, 1979), and Plot Units (Lehnert 1981). It was also during this time that numerous early chatterbots made their debut, including the somewhat infamous PARRY , the bizarrely creative Racter , and the perpetually conversational Jabberwacky . These programs, for all their rudimentary nature, hinted at humanity’s enduring desire for digital companionship, even if it was with something fundamentally misunderstanding them.

Statistical period

Up until the late 1980s, the vast majority of NLP systems were meticulously constructed upon elaborate sets of hand-written rules. These systems, often brittle and difficult to scale, reflected the Chomskyan influence that prioritized explicit grammatical structures. However, the late 1980s heralded a significant paradigm shift, a quiet revolution in NLP marked by the integration of machine learning algorithms for language processing. This pivot was driven by a confluence of factors: the relentless march of Moore’s law , which delivered ever-increasing computational power, and a gradual, almost grudging, decline in the absolute dominance of strictly Chomskyan theories of linguistics (such as transformational grammar ). These highly theoretical frameworks, with their emphasis on “corner cases ” and thought experiments, often discouraged the kind of extensive, real-world data analysis inherent in corpus linguistics – the very foundation of the machine-learning approach. [7]

Initially, some of the pioneering machine learning algorithms, like decision trees , produced systems composed of hard if-then rules, not dissimilar in their logical structure to the hand-written rules they were replacing. However, the trajectory of research increasingly gravitated towards statistical models . These models adopted a more nuanced approach, making soft, probabilistic decisions by assigning real-valued weights to the various features extracted from the input data. The cache language models that underpin many modern speech recognition systems serve as excellent examples of such statistical frameworks. The inherent advantage of these statistical models lies in their robustness; they tend to perform far more gracefully when confronted with unfamiliar or erroneous input, a common occurrence with real-world linguistic data. Furthermore, their probabilistic nature allows for more reliable integration into larger, multi-component systems, where multiple subtasks must cooperate seamlessly. It turns out, language isn’t a neat set of rules; it’s a statistical anomaly, a chaotic symphony that sometimes requires a more flexible, forgiving interpretation.

Datasets

The ascendance of statistical methodologies was not merely a triumph of algorithmic innovation; it was fundamentally enabled by two crucial developments: the aforementioned explosion in computing power and, perhaps more importantly, the burgeoning availability of vast datasets. During this period, large multilingual corpora began to materialize, almost as a happy accident of bureaucracy. Notably, some of the most extensive early corpora were generated by legislative bodies such as the Parliament of Canada and the European Union . These invaluable resources were a direct consequence of legal mandates requiring the translation of all governmental proceedings into all official languages of their respective systems. What started as administrative necessity became a goldmine for linguistic research, providing the raw material for machine learning algorithms to train on.

Many of the most significant early breakthroughs achieved through this statistical paradigm occurred within the demanding field of machine translation . By 1993, the groundbreaking IBM alignment models were being deployed for statistical machine translation . [8] These systems represented a profound departure from their predecessors, which were typically symbolic systems meticulously hand-coded by computational linguists. The statistical nature of the IBM models allowed them to automatically learn intricate patterns and relationships directly from colossal textual corpora . While these early statistical systems struggled in scenarios where only limited corpora were available, their success underscored the immense potential of data-driven approaches. Consequently, the development of more data-efficient methods continues to be an active and critical area of research and development.

The hunger for data only grew. In 2001, a truly monumental text corpus, comprising a staggering one billion words, was compiled through systematic scraping of the Internet. This dataset, described at the time as “very very large,” was instrumental in advancing the challenging task of word disambiguation – determining the correct meaning of a word in context, a task that often stumps even humans. [9]

To fully leverage these enormous, often unlabelled datasets, researchers dedicated themselves to developing algorithms for unsupervised and self-supervised learning . These methods aimed to uncover inherent structures and patterns within data without explicit human annotation, a far more challenging endeavor than supervised learning . While unsupervised learning typically yields less accurate results for a given quantity of input data, the sheer, unimaginable volume of non-annotated data available (including, most notably, the entirety of the World Wide Web ) often compensates for this deficit, making it an indispensable tool for scaling NLP capabilities. It was the desperate scramble to make sense of the vast, unannotated digital ocean, a noble effort, if often less precise.

Neural period

Timeline of natural language processing models

The dawn of the 1990s marked the quiet emergence of a new computational paradigm: neural language models . In 1990, the Elman network , a pioneering application of a recurrent neural network (RNN), introduced a revolutionary concept: encoding each word in a training set as a distinct numerical vector, often referred to as a word embedding . The entire vocabulary was thus transformed into a vector database , allowing the network to capture semantic relationships and perform tasks like sequence-prediction that were beyond the capabilities of a simpler multilayer perceptron . However, these early static embeddings possessed a significant limitation: they failed to differentiate between the multiple meanings of homonyms , treating “bank” (financial institution) and “bank” (river edge) as identical. [10] A subtle flaw, perhaps, but one that highlights the enduring complexity of language.

The field saw another pivotal moment in 2000 when Yoshua Bengio developed the first neural probabilistic language model. [11] This work laid crucial groundwork, demonstrating how neural networks could learn to predict sequences of words, capturing statistical regularities in language with greater fidelity. The subsequent two decades witnessed an accelerating pace of innovation. A confluence of novel algorithms, the ever-increasing availability of larger and more diverse datasets, and the relentless rise in processing power (particularly with the advent of specialized hardware like GPUs) made it feasible to train neural networks of unprecedented scale and complexity, leading to the development of larger and larger language models.

A significant architectural leap occurred with the introduction of the attention mechanism by Bahdanau et al. in 2014. [12] This innovation allowed neural networks to dynamically weigh the importance of different parts of the input sequence when generating an output, effectively enabling them to “focus” on relevant information. This foundational work directly paved the way for the now-legendary “Attention is All You Need” paper [13] in 2017, which unveiled the Transformer architecture . Transformers, with their self-attention mechanisms, revolutionized sequence processing, drastically improving performance and parallelization capabilities.

The concept of the large language model (LLM) truly solidified and captured public imagination in the late 2010s. An LLM is, at its core, a language model trained through self-supervised learning on truly colossal amounts of text data. The earliest public LLMs, while impressive, typically featured hundreds of millions of parameters [14]. However, this number rapidly escalated, soaring into the billions and even trillions of parameters within a few short years [15]. This exponential growth in model size, coupled with sophisticated training regimes, has endowed LLMs with capabilities that were once the exclusive domain of science fiction, enabling them to generate coherent, contextually relevant, and remarkably human-like text across a vast array of tasks.

In recent years, the advancements in deep learning, particularly the development and scaling of large language models, have fundamentally transformed the capabilities of natural language processing. These powerful models have led to widespread and increasingly sophisticated applications across diverse sectors. In healthcare, they assist with medical transcription, diagnostic support, and patient interaction. In customer service, they power intelligent virtual assistants and chatbots, streamlining support and improving user experience. And in content generation, they are revolutionizing the creation of articles, marketing copy, and even creative writing, demonstrating a remarkable ability to understand and generate human language with unprecedented fluency and nuance. [16] A triumph of mimicry, if nothing else.

Software

SoftwareYearCreatorDescriptionReference
Georgetown experiment1954Georgetown University and IBMThis collaborative effort involved the fully automatic translation of over sixty Russian sentences into English, sparking a brief, intense period of over-optimism regarding the imminent “solution” to machine translation.
STUDENT1964Daniel BobrowA pioneering program capable of solving high school algebra word problems, demonstrating early success in understanding and processing mathematical language within natural language input. [17]
ELIZA1964Joseph WeizenbaumA deceptively simple program that simulated a Rogerian psychotherapist by cleverly rephrasing user input with a few grammar rules. It famously fooled many into believing they were conversing with a human, proving that sometimes, just appearing to listen is enough. [18]
SHRDLU1970Terry WinogradAn influential natural language system that operated within highly restricted “blocks worlds ” and used a finite vocabulary. It worked extremely well within its narrow domain, manipulating virtual objects based on natural language commands.
PARRY1972Kenneth ColbyA notable chatterbot designed to simulate a paranoid individual, often engaging users in defensive and suspicious conversations.
KL-ONE1974Sondheimer et al.A foundational knowledge representation system developed in the tradition of semantic networks and frames. It is considered an early example of a frame language , structuring knowledge hierarchically.
MARGIE1975Roger SchankOne of the early conceptual dependency-based natural language understanding systems, aiming to represent the meaning of sentences using a set of primitive conceptual actions.
TaleSpin (software)1976MeehanA program that generated simple stories based on pre-defined plot units and character goals, an early foray into automated narrative generation.
QUALMLehnertAn early question-answering system that utilized conceptual dependency theory to understand questions and retrieve relevant information from a knowledge base.
LIFER/LADDER1978HendrixA sophisticated natural language interface designed to query a database containing information about US Navy ships, demonstrating practical application of NLP for database interaction.
SAM (software)1978CullingfordAnother conceptual dependency-based system focused on story understanding, capable of answering questions about narratives it had “read.”
PAM (software)1978Robert WilenskyA “Plan Applier Mechanism” that understood stories by inferring the goals and plans of characters, contributing to deeper narrative comprehension.
Politics (software)1979CarbonellAn NLP system designed to understand and answer questions about political events and arguments, using conceptual dependency to represent political knowledge.
Plot Units (software)1981LehnertDeveloped further from QUALM, this system used “plot units” to represent common narrative structures, aiding in story understanding and summarization.
Jabberwacky1982Rollo CarpenterA long-running chatterbot with the ambitious goal of simulating natural human chat in an interesting, entertaining, and humorous manner, constantly learning from its interactions.
MUMBLE (software)1982McDonaldAn early natural language generation system that focused on producing coherent and grammatically correct English text from conceptual representations.
Racter1983William Chamberlain and Thomas EtterA unique chatterbot that gained notoriety for generating surprisingly coherent, albeit often nonsensical, English language prose at random, blurring the lines between programmed response and creative output.
MOPTRANS [19]1984LytinenA machine translation system that utilized Memory Organization Packets (MOPs), a knowledge representation scheme, to translate based on conceptual understanding rather than purely linguistic rules.
KODIAK (software)1986WilenskyA knowledge representation system that built upon conceptual dependency, aiming for a more robust and flexible way to represent and reason about knowledge.
Absity (software)1987HirstA system that explored lexical ambiguity and word sense disambiguation, attempting to resolve multiple meanings of words in context.
Dr. Sbaitso1991Creative LabsA simple, early speech synthesis and natural language processing program often bundled with sound cards, providing a rudimentary interactive “psychologist” experience.
IBM Watson2006IBMA highly sophisticated question-answering system that achieved widespread fame by winning the Jeopardy! contest in February 2011, decisively defeating the best human players through its ability to process natural language questions and access vast knowledge bases.
Siri2011AppleApple’s pioneering virtual assistant, which brought voice-activated natural language interaction to mainstream consumer devices, allowing users to perform tasks and ask questions using spoken commands.
Cortana2014MicrosoftMicrosoft’s virtual assistant, integrated into Windows and other platforms, offering personalized assistance, search capabilities, and task automation through natural language.
Amazon Alexa2014AmazonAmazon’s ubiquitous virtual assistant, primarily known for its role in smart speakers and home automation, providing voice control for various tasks and information retrieval.
Google Assistant2016GoogleGoogle’s intelligent virtual assistant, deeply integrated across Android devices and Google services, offering conversational interaction, personalized information, and smart home control.
ChatGPT2022OpenAIA highly influential generative chatbot that captivated the public with its ability to produce remarkably coherent, contextually relevant, and human-like text across a vast array of prompts and conversational styles. The latest iteration of the chatbot, now capable of generating prose that sounds suspiciously human. A triumph of mimicry, if nothing else.