|Part of a series on
Cognitive linguistics is an interdisciplinary branch of linguistics, combining knowledge and research from cognitive science, cognitive psychology, neuropsychology and linguistics. Models and theoretical accounts of cognitive linguistics are considered as psychologically real, and research in cognitive linguistics aims to help understand cognition in general and is seen as a road into the human mind.
There has been scientific and terminological controversy around the label "cognitive linguistics"; there is no consensus on what specifically is meant with the term.
The roots of cognitive linguistics are in Noam Chomsky's 1959 critical review of B. F. Skinner's Verbal Behavior. Chomsky's rejection of behavioural psychology and his subsequent anti-behaviourist activity helped bring about a shift of focus from empiricism to mentalism in psychology under the new concepts of cognitive psychology and cognitive science.
Chomsky considered linguistics as a subfield of cognitive science in the 1970s but called his model transformational or generative grammar. Having been engaged with Chomsky in the linguistic wars, George Lakoff united in the early 1980s with Ronald Langacker and other advocates of neo-Darwinian linguistics in a so-called "Lakoff–Langacker agreement". It is suggested that they picked the name "cognitive linguistics" for their new framework to undermine the reputation of generative grammar as a cognitive science.
Consequently, there are three competing approaches that today consider themselves as true representatives of cognitive linguistics. One is the Lakoffian–Langackerian brand with capitalised initials (Cognitive Linguistics). The second is generative grammar, while the third approach is proposed by scholars whose work falls outside the scope of the other two. They argue that cognitive linguistics should not be taken as the name of a specific selective framework, but as a whole field of scientific research that is assessed by its evidential rather than theoretical value.
Generative grammar functions as a source of hypotheses about language computation in the mind and brain. It is argued to be the study of 'the cognitive neuroscience of language'. Generative grammar studies behavioural instincts and the biological nature of cognitive-linguistic algorithms, providing a computational–representational theory of mind.
This in practice means that sentence analysis by linguists is taken as a way to uncover cognitive structures. It is argued that a random genetic mutation in humans has caused syntactic structures to appear in the mind. Therefore, the fact that people have language does not rely on its communicative purposes.
For a famous example, it was argued by linguist Noam Chomsky that sentences of the type "Is the man who is hungry ordering dinner" are so rare that it is unlikely that children will have heard them. Since they can nonetheless produce them, it was further argued that the structure is not learned but acquired from an innate cognitive language component. Generative grammarians then took as their task to find out all about innate structures through introspection in order to form a picture of the hypothesised language faculty.
Generative grammar promotes a modular view of the mind, considering language as an autonomous mind module. Thus, language is separated from mathematical logic to the extent that inference cannot explain language acquisition. The generative conception of human cognition is also influential in cognitive psychology and computer science.
One of the approaches to cognitive linguistics is called Cognitive Linguistics, with capital initials, but it is also often spelled cognitive linguistics with all lowercase letters. This movement saw its beginning in early 1980s when George Lakoff's metaphor theory was united with Ronald Langacker's Cognitive Grammar, with subsequent models of Construction Grammar following from various authors. The union entails two different approaches to linguistic and cultural evolution: that of the conceptual metaphor, and the construction.
Cognitive Linguistics defines itself in opposition to generative grammar, arguing that language functions in the brain according to general cognitive principles. Lakoff's and Langacker's ideas are applied across sciences. In addition to linguistics and translation theory, Cognitive Linguistics is influential in literary studies, education, sociology, musicology, computer science and theology.
According to American linguist George Lakoff, metaphors are not just figures of speech, but modes of thought. Lakoff hypothesises that principles of abstract reasoning may have evolved from visual thinking and mechanisms for representing spatial relations that are present in lower animals. Conceptualisation is regarded as being based on the embodiment of knowledge, building on physical experience of vision and motion. For example, the 'metaphor' of emotion builds on downward motion while the metaphor of reason builds on upward motion, as in saying “The discussion fell to the emotional level, but I raised it back up to the rational plane." It is argued that language does not form an independent cognitive function but fully relies on other cognitive skills which include perception, attention, motor skills, and visual and spatial processing. Same is said of various other cognitive phenomena such as the sense of time:
In Cognitive Linguistics, thinking is argued to be mainly automatic and unconscious. Cognitive linguists study the embodiment of knowledge by seeking expressions which relate to modal schemas. For example, in the expression "It is quarter to eleven", the preposition to represents a modal schema which is manifested in language as a visual or sensorimotoric 'metaphor'.
Constructions, as the basic units of grammar, are conventionalised form–meaning pairings which are comparable to memes as units of linguistic evolution. These are considered multi-layered. For example, idioms are higher-level constructions which contain words as middle-level constructions, and these may contain morphemes as lower-level constructions. It is argued that humans do not only share the same body type, allowing a common ground for embodied representations; but constructions provide common ground for uniform expressions within a speech community. Like biological organisms, constructions have life cycles which are studied by linguists.
According to the cognitive and constructionist view, there is no grammar in the traditional sense of the word. What is commonly perceived as grammar is an inventory of constructions; a complex adaptive system; or a population of constructions. Constructions are studied in all fields of language research from language acquisition to corpus linguistics.
There is also a third approach to cognitive linguistics, which neither directly supports the modular (Generative Grammar) nor the anti-modular (Cognitive Linguistics) view of the mind. Proponents of the third view argue that, according to brain research, language processing is specialized although not autonomous from other types of information processing. Language is thought of as one of the human cognitive abilities, along with perception, attention, memory, motor skills, and visual and spatial processing, rather than being subordinate to them. Emphasis is laid on a cognitive semantics that studies the contextual–conceptual nature of meaning.
Cognitive linguistics offers a scientific first principle direction for quantifying states-of-mind through natural language processing.  As mentioned earlier Cognitive Linguistics, approaches grammar with a nontraditional view. Traditionally grammar has been defined as a set of structural rules governing the composition of clauses, phrases and words in a natural language. From the perspective of Cognitive Linguistics, grammar is seen as the rules of arrangement of language which best serve communication of the experience of the human organism through its cognitive skills which include perception, attention, motor skills, and visual and spatial processing. Such rules are derived from observing the conventionalized pairings of meaning to understand sub-context in the evolution of language patterns.  The cognitive approach to identifying sub-context by observing what comes before and after each linguistic construct provides a grounding of meaning in terms of sensorimotoric embodied experience.  When taken together, these two perspectives form the basis of defining approaches in computational linguistics with strategies to work through the symbol grounding problem which posits that, for a computer, a word is merely a symbol, which is a symbol for another symbol and so on in an unending chain without grounding in human experience.  The broad set of tools and methods of computational linguistics are available as natural language processing or NLP. Cognitive linguistics adds a new set of capabilities to NLP. These cognitive NLP methods enable software to analyze sub-context in terms of internal embodied experience. 
The goal of natural language processing (NLP) is to enable a computer to "understand" the contents of text and documents, including the contextual nuances of the language within them. The perspective of traditional Traditional Chomskyan Linguistics offers NLP three approaches or methods to identify and quantify the literal contents, the who, what, where and when in text – in linguistic terms, the semantic meaning or semantics of the text. The perspective of cognitive linguistics offers NLP a direction to identify and quantify the contextual nuances, the why and how in text – in linguistics terms, the implied pragmatic meaning or pragmatics of text.
The three NLP approaches to understanding literal semantics in text based on traditional linguistics are symbolic NLP, statistical NLP, and neural NLP. The first method, symbolic NLP (1950s – early 1990s) is based on first principles and rules of traditional linguistics. The second method, statistical NLP (1990s–2010s), builds upon the first method with a layer of human curated & machine-assisted corpora for multiple contexts. The third approach neural NLP (2010 onwards), builds upon the earlier methods by leveraging advances in deep neural network-style methods to automate tabulation of corpora & parse models for multiple contexts in shorter periods of time.  All three methods are used to power NLP techniques like stemming and lemmatisation in order to obtain statistically relevant listing of the who, what, where & when in text through named-entity recognition and Topic model programs. The same methods have been applied with NLP techniques like a bag-of-words model to obtain statistical measures of emotional context through sentiment analysis programs. The accuracy of a sentiment analysis system is, in principle, how well it agrees with human judgments. Because evaluation of sentiment analysis is becoming more and more specialty based, each implementation needs a separate training model and specialized human verification raising Inter-rater reliability issues. However, the accuracy is considered generally acceptable for use in evaluating emotional context at a statistical or group level.  
A developmental trajectory of NLP to understand contextual pragmatics in text involving emulating intelligent behavior and apparent comprehension of natural language is cognitive NLP. This method is a rules based approach which involves assigning meaning to a word, phrase, sentence or piece of text based on the information presented before and after the piece of text being analyzed.
The specific meaning of cognitive linguistics, the proper address of the name, and the scientific status of the enterprise have been called into question. Criticism includes an overreliance on introspective data, a lack of experimental testing of hypotheses and little integration of findings from other fields of cognitive science. Some researchers go as far as to consider calling the field 'cognitive' at all a misnomer.
"It would seem to me that [cognitive linguistics] is the sort of linguistics that uses findings from cognitive psychology and neurobiology and the like to explore how the human brain produces and interprets language. In other words, cognitive linguistics is a cognitive science, whereas Cognitive Linguistics is not. Most of generative linguistics, to my mind, is not truly cognitive either."— Bert Peeters
There has been criticism regarding the brain-related claims of both Chomsky's generative grammar, and Lakoff's Cognitive Linguistics. These are said to advocate too extreme views on the axis of modular versus general processing. The empirical evidence points to language being partially specialized and interacting with other systems. However, to counter behaviorism, Chomsky postulated that language acquisition occurs inside an autonomous module, which he calls the language faculty, thus suggesting a very high degree of specialization of language in the brain. To offer an alternative to his view, Lakoff, in turn, postulated the opposite by claiming that language acquisition is not specialized at all because language does not constitute a cognitive capacity of its own but occurs in the sensory domains such as vision and kinesthesis. According to the critical view, these ideas were not motivated by brain research but by a struggle for power in linguistics. Members of such frameworks are also said to have used other researchers' findings to present them as their own work. While this criticism is accepted for most part, it is claimed that some of the research has nonetheless produced useful insights.