Natural language processing (NLP)
- Definition
- The branch of AI focused on enabling machines to understand, interpret, and generate human language. NLP underpins chatbots, translation, search, and document analysis across every industry.
- Why it matters
- NLP is the foundation that LLMs are built on, and understanding its evolution helps contextualize the current moment. Before transformers, NLP relied on rule-based systems, statistical models, and recurrent neural networks, each generation improving but still brittle. The transformer architecture and pre-training paradigm changed everything: instead of building task-specific NLP models, you train one massive language model and adapt it to any task. NLP as a distinct discipline is being absorbed into general-purpose LLM capability, but the problems it defined, machine translation, sentiment analysis, named entity recognition, information extraction, remain commercially important.
- In practice
- Traditional NLP tasks that once required specialized models (sentiment analysis, named entity recognition, text classification) are now handled by general-purpose LLMs with simple prompting. Google Translate's shift from phrase-based to neural machine translation, and then to LLM-based translation, illustrates the NLP evolution. Enterprise NLP applications include: contract analysis (extracting clauses and obligations from legal documents), customer feedback analysis (classifying and routing support tickets), and regulatory monitoring (scanning documents for compliance language). The NLP market, now largely synonymous with the LLM applications market, is projected to exceed $50B by 2027.
We cover models & architecture every week.
Get the 5 AI stories that matter — free, every Friday.
Related terms
LLM (Large Language Model)
A neural network trained on massive text corpora to predict and generate language. LLMs like GPT-4, Claude, and Gemini are the foundation of the current AI wave, powering chatbots, coding tools, and enterprise automation.
Transformer
The neural network architecture behind virtually all modern language and multi-modal models. Introduced in Google's 2017 'Attention Is All You Need' paper, transformers use self-attention to process sequences in parallel.
Embedding
A numerical vector representation of text, images, or other data that captures semantic meaning. Embeddings power search, recommendations, and RAG systems by letting you find conceptually similar content.
Tokenizer
The algorithm that splits text into tokens before a model can process it. Different models use different tokenizers, which affects how efficiently they handle various languages, code, and specialized content.
Know the terms. Know the moves.
Get the 5 AI stories that matter every Friday — free.
Free forever. No spam.