Natural language processing (NLP) refers to enabling computers to understand human language and communicate with people in a natural way. It involves natural language understanding to analyze language and natural language generation to produce understandable language. NLP analyzes aspects of language like syntax, semantics, and pragmatics. The main components of NLP are lexical analysis, syntactic analysis, semantic analysis, and discourse/pragmatic analysis. Context-free grammars are commonly used for syntactic analysis in NLP.
Natural language processing (NLP) refers to enabling computers to understand human language and communicate with people in a natural way. It involves natural language understanding to analyze language and natural language generation to produce understandable language. NLP analyzes aspects of language like syntax, semantics, and pragmatics. The main components of NLP are lexical analysis, syntactic analysis, semantic analysis, and discourse/pragmatic analysis. Context-free grammars are commonly used for syntactic analysis in NLP.
Natural language processing (NLP) refers to enabling computers to understand human language and communicate with people in a natural way. It involves natural language understanding to analyze language and natural language generation to produce understandable language. NLP analyzes aspects of language like syntax, semantics, and pragmatics. The main components of NLP are lexical analysis, syntactic analysis, semantic analysis, and discourse/pragmatic analysis. Context-free grammars are commonly used for syntactic analysis in NLP.
Natural language processing (NLP) refers to enabling computers to understand human language and communicate with people in a natural way. It involves natural language understanding to analyze language and natural language generation to produce understandable language. NLP analyzes aspects of language like syntax, semantics, and pragmatics. The main components of NLP are lexical analysis, syntactic analysis, semantic analysis, and discourse/pragmatic analysis. Context-free grammars are commonly used for syntactic analysis in NLP.
Download as PPTX, PDF, TXT or read online from Scribd
Download as pptx, pdf, or txt
You are on page 1of 21
At a glance
Powered by AI
The key takeaways are that natural language processing (NLP) refers to enabling computers to understand, interpret and generate human language. The two main components of NLP are natural language understanding and natural language generation.
The two main components of NLP are natural language understanding (NLU) and natural language generation (NLG). NLU involves mapping input text to representations and analyzing aspects of language, while NLG is the process of generating natural language from internal representations.
There are different levels of ambiguity in natural language including lexical ambiguity at the word level, syntactic ambiguity in parsing sentences, and referential ambiguity in pronouns. One input can have multiple meanings and multiple inputs can have the same meaning.
Natural Language Processing (NLP) refers to
AI method of communicating with an
intelligent systems using a natural language such as English. Processing of Natural Language is required when we want an intelligent system like robot to perform as per our instructions, when we want to hear decision from a dialogue based clinical expert system, etc. There are two components of NLP as given − ◦ Natural Language Understanding (NLU) ◦ Natural Language Generation (NLG) Understanding involves the following tasks − Mapping the given input in natural language into useful representations. Analyzing different aspects of the language. It is the process of producing meaningful phrases and sentences in the form of natural language from some internal representation. It involves − Text planning − It includes retrieving the relevant content from knowledge base. Sentence planning − It includes choosing required words, forming meaningful phrases, setting tone of the sentence. Text Realization − It is mapping sentence plan into sentence structure. NL is very ambiguous. There can be different levels of ambiguity − Lexical ambiguity − It is at very primitive level such as word- level. For example, treating the word “board” as noun or verb? “ Bank” can be used for River Bank also and for depositing Money also. Syntax Level ambiguity − A sentence can be parsed in different ways. For example, “I saw a boy with Telescope.” has 2 different meaning. Referential ambiguity − Referring to something using pronouns. For example, Rima went to Gauri. She said, “I am tired.” − Exactly who is tired? One input can mean different meanings. Many inputs can mean the same thing Phonology − It is study of organizing sound systematically. Morphology − It is a study of construction of words from primitive meaningful units. Morpheme − It is primitive unit of meaning in a language. Syntax − It refers to arranging words to make a sentence. It also involves determining the structural role of words in the sentence and in phrases. Semantics − It is concerned with the meaning of words and how to combine words into meaningful phrases and sentences. Pragmatics − It deals with using and understanding sentences in different situations and how the interpretation of the sentence is affected. Discourse − It deals with how the immediately preceding sentence can affect the interpretation of the next sentence. World Knowledge − It includes the general knowledge about the world There are general five steps − It involves identifying and analyzing the structure of words. Lexicon of a language means the collection of words and phrases in a language. Lexical analysis is dividing the whole chunk of txt into paragraphs, sentences, and words. It involves analysis of words in the sentence for grammar and arranging words in a manner that shows the relationship among the words. The sentence such as “The school goes to boy” is rejected by English syntactic analyzer. It draws the exact meaning or the dictionary meaning from the text. The text is checked for meaningfulness. It is done by mapping syntactic structures and objects in the task domain. The semantic analyzer disregards sentence such as “hot ice-cream”. The meaning of any sentence depends upon the meaning of the sentence just before it. In addition, it also brings about the meaning of immediately succeeding sentence. During this, what was said is re-interpreted on what it actually meant. It involves deriving those aspects of language which require real world knowledge. There are a number of algorithms researchers have developed for syntactic analysis, but we consider only the following simple methods − ◦ Context-Free Grammar ◦ Top-Down Parser It is the grammar that consists rules with a single symbol on the left-hand side of the rewrite rules. Let us create grammar to parse a sentence − “The bird pecks the grains” Articles (DET) − a | an | the Nouns − bird | birds | grain | grains Noun Phrase (NP) − Article + Noun | Article + Adjective + Noun = DET N | DET ADJ N Verbs − pecks | pecking | pecked Verb Phrase (VP) − NP V | V NP Adjectives (ADJ) − beautiful | small | chirping The parse tree breaks down the sentence into structured parts so that the computer can easily understand and process it. In order for the parsing algorithm to construct this parse tree, a set of rewrite rules, which describe what tree structures are legal, need to be constructed. These rules say that a certain symbol may be expanded in the tree by a sequence of other symbols. According to first order logic rule, if there are two strings Noun Phrase (NP) and Verb Phrase (VP), then the string combined by NP followed by VP is a sentence. The rewrite rules for the sentence are as follows − S → NP VP NP → DET N | DET ADJ N VP → V NP Lexocon − DET → a | the ADJ → beautiful | perching N → bird | birds | grain | grains V → peck | pecks | pecking Merit − The simplest style of grammar, therefore widely used one. Demerits They are not highly precise. For example, “The grains peck the bird”, is a syntactically correct according to parser, but even if it makes no sense, parser takes it as a correct sentence. To bring out high precision, multiple sets of grammar need to be prepared. It may require a completely different sets of rules for parsing singular and plural variations, passive sentences, etc., which can lead to creation of huge set of rules that are unmanageable.