Open navigation menu
Close suggestions
Search
Search
en
Change Language
Upload
Sign in
Sign in
Download free for days
0 ratings
0% found this document useful (0 votes)
256 views
Natural Language Understanding
Uploaded by
Akshat
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here
.
Available Formats
Download as PDF or read online on Scribd
Download now
Download
Save Natural Language Understanding For Later
Download
Save
Save Natural Language Understanding For Later
0%
0% found this document useful, undefined
0%
, undefined
Embed
Share
Print
Report
0 ratings
0% found this document useful (0 votes)
256 views
Natural Language Understanding
Uploaded by
Akshat
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here
.
Available Formats
Download as PDF or read online on Scribd
Download now
Download
Save Natural Language Understanding For Later
Carousel Previous
Carousel Next
Save
Save Natural Language Understanding For Later
0%
0% found this document useful, undefined
0%
, undefined
Embed
Share
Print
Report
Download now
Download
You are on page 1
/ 675
Search
Fullscreen
Fq Atma ie IN PNOraE LANGUAGE UNDERSTANDING SECOND EDITIONNatural Language Understanding James Allen i it sterAcqutisitions Editor: 1, Caner Shanklin Production Editor: Ray Kanarr Executive Editor: Dan Joraanstad Copy Editor: Barbara Conway Editorial Assistant: Melissa Standen Proofreader: Joe Ruddick Cover Designer: Yvo Riezebos Design Design Consultant: Michael Rogondino Technical Assistant: Peg Meeker Macintosh is a trademark of Apple Computer, Inc. Word is a trademark of Microsoft, Inc. Canvas is a wademark of Deneba Software. Camera-ready copy for this book was prepared on a Macintosh with Microsoft Word and Canvas. The programs and the applications presented in this book have been included for their instructional value. They have been tested with care but are not guaranteed for any particular purpose. The publisher does not offer any warranties or representations, nor does it accept any liabilities with respect to the programs or applications. Copyright © 1995 by The Benjamin/Cummings Publishing Company, Inc. All rights reserved, No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photo- copying, recording, or otherwise, without the prior written permission of the publisher. Printed in the United States of America. Published simultaneously in Canada, Library of Congress Cataloging-in-Publication Data Allen, James. Natural language understanding / James Allen. Pp. am. Includes bibliographical references and index. ISBN 0-8053-0334-0 1. Programming languages (Electronic computers) --Semantics. 2, Language and logic. 3, artificial intelligence, I, Title. QATE.7.a4a 1994 006.3°6--dc20 94-18218 CIF 4123456789 10—MA—98 97 96 95 94 ‘The Benjamin/Cummings Publishing Company, Inc. 390 Bridge Parkway Redwood City, CA 94065Brief Contents Preface — xii Chapter | Introduction to Natural Language Understanding 1 PARTI Syntactic Processing Chapter 2 Linguistic Background: An Outline of English Syntax 23 Chapter 3. Grammars and Parsing 41 Chapter 4 Features and Augmented Grammars 83 Chapter 5 Grammars for Natural Language 123 Chapter 6 Toward Efficient Parsing 159 Chapter 7 Ambiguity Resolution: Statistical Methods 189 PART II Semantic Interpretation Chapter 8 Semantics and Logical Form 227 Chapter 9 Linking Syntax and Semantics 263 Chapter 10 Ambiguity Resolution 295 Chapter (1 Other Strategies for Semantic Interpretation 328 Chapter 12 Scoping and the Interpretation of Noun Phrases 351 PART IIL Context and World Knowledge Chapter 13 Knowledge Representation and Reasoning 391 Chapter 14 Local Discourse Context and Reference 429 Chapter 15 Using World Knowledge 465 Chapter 16 Discourse Sucture 503 Chapter 17. Defining a Conversational Agem — 541 Appendix A An Introduction to Logic and Model-Theoretic Semantics 579 Appendix B Symbolic Computation 595 Appendix C Speech Recognition and Spoken Language 611 Bibliography 629 Index 645Contents PREFACE xi Chapter 1 Introduction to Natural Language Understanding 1.1 The Study of Language | 12 Applications of Natural Language Understanding 3 13 Evaluating Language Understanding Systems 6 14 The Different Levels of Language Analysis 9 1.5 Representations and Understanding 11 1,6 The Organization of Natural Language Understanding Systems 15 PARTI SYNTACTIC PROCESSING Chapter 2 Linguistic Background: An Outline of English Syntax 21 Words 23 2.2 The Elements of Simple Noun Phrases 25 2.3. Verb Phrases and Simple Sentences 28 24 Noun Phrases Revisited 33 25 Adjective Phrases 35 2.6 Adverbial Phrases 35 Chapter 3 Grammars and Parsing 3.1 Grammars and Sentence Structure 41 3.2 What Makes aGoodGrammar 44 3.3 A Top-Down Parser 47 34 ABotom-UpChan Parser 53 3.5 Transition Network Grammars 61 © 36 Top-Down Chart Parsing 65 © 37 Finite State Models and Morphological Processing 70 © 3.8 Grammars and Logic Programming 72 (o—optional topic)Chapter 4 Features and Augmented Grammars 4.1 Feature Systems and Augmented Grammars 42 Some Basic Feature Systems for English 86 43 Morphological Analysis and the Lexicon 90 44 A Simple Grammar Using Features 94 45 Parsing with Features 98 e 46 Augmented Transition Networks 101 ° 47 Definite Clause Grammars 106 ° 48 Generalized Feature Systems and Unification Grammars 109 Chapter 5 Grammars for Natural Language 5.1 Auxiliary Verbs and Verb Phrases 123 52 Movement Phenomena in Language 127 5.3. Handling Questions in Context-Free Grammars o 54 Relative Clauses 141 5.5 The Hold Mechanismin ATNs = 144 56 GapThreading 148 Chapter 6 Toward Efficient Parsing 6.1 Human Preferences in Parsing 159 62 Encoding Uncertainty: Shift-Reduce Parsers 63 A Deterministic Parser 170 64 Techniques for Efficient Encoding of Ambiguity 65 Partial Parsing 180 Chapter 7 Ambiguity Resolution: Statistical Methods 7.1 Basic Probability Theory 189 7.2 Estimating Probabilities 192 73 Part-of-Speech Tagging 195 7.4 Obtaining Lexical Probabilities 204 75. Probabilistic Context-Free Grammars 209 7.6 Best-First Parsing 213 7. A Simple Context-Dependent Best-First Parser i CONTENTS 132 163 176 216CONTENTS vii RTI SEMANTIC INTERPRETATION Chapter 8 Semantics and Logical Form 8.1 8.2 83 84 85 86 87 ° 88 Semantics and Logical Form 227 Word Senses and Ambiguity 231 The Basic Logical Form Language 233 Encoding Ambiguity in the Logical Form — 238 Verbs and States in Logical Form 241 Thematic Roles 244 Speech Acts and Embedded Sentences 250 Defining Semantic Structure: Model Theory 251 Chapter 9 Linking Syntax and Semantics OL 92 93 94 95 96 ° OF Semantic Interpretation and Compositionality 263 A Simple Grammar and Lexicon with Semantic Interpretation 267 Prepositional Phrases and Verb Phrases 271 Lexicalized Semantic Interpretation and Semantic Roles 275 Handling Simple Questions 280 Semantic Interpretation Using Feature Unification 283 Generating Sentences from Logical Form 286 Chapter 10 Ambiguity Resolution 10.1 10.2 10.3 10.4 10.5 10.6 Selectional Restrictions 295 Semantic Filtering Using Selectional Restrictions 302 Semantic Networks 305 Statistical Word Sense Disambiguation 310 Statistical Semantic Preferences 314 Combining Approaches to Disambiguation 318 Chapter 11 Other Strategies for Semantic Interpretation Wl W2 13 14 Grammatical Relations 328 Semantic Grammars 332 Template Matching 334 Semantically Driven Parsing Techniques 341viii CONTENTS Chapter 12 Scoping and the Interpretation of Noun Phrases 12.1 12.2 12.3 124 12.5 12.6 12.7 PART IIL Scoping Phenomena 351 Definite Descriptions and Scoping 359 A Method for Scoping While Parsing 360 Co-Reference and Binding Constraints 366 Adjective Phrases 372 Relational Nouns and Nominalizations 375 Other Problems in Semantics 378 CONTEXT AND WORLD KNOWLEDGE Chapter 13 Knowledge Representation and Reasoning 13.1 13.2 13.3 13.4 e 13.5 ° 136 ° 137 13.8 Chapter 14 Knowledge Representation 392 A Representation Based on FOPC 397 Frames: Representing Stereotypical Information 400 Handling Natural Language Quantification 404 Time and Aspectual Classes of Verbs 406 Automating Deduction in Logic-Based Representations 410 Procedural Semantics and Question Answering 414 Hybrid Knowledge Representations 419 Local Discourse Context and Reference 141 14.2 14.3 14.4 ° 145 146 14.7 Defining Local Discourse Context and Discourse Entities 429 A Simple Model of Anaphora Based on History Lists 433 Pronouns and Centering 435 Definite Descriptions 440 Definite Reference and Sets 445 Ellipsis 449 Surface Anaphora 455conTENTS ix Chapter 15 Using World Knowledge 15.1 Using World Knowledge: Establishing Coherence 465 15.2 Matching Against Expectations 466 15.3. Reference and Matching Expectations 471 15.4 Using Knowledge About Action and Causality 473 15.5 Scripts: Understanding Stereotypical Situations 477 15.6 Using Hierarchical Plans 480. ° 15.7 Action-Effect-Based Reasoning 483 ° 15.8 Using Knowledge About Rational Behavior 490 Chapter 16 Discourse Structure 16.1 The Need for Discourse Structure 503 16.2 Segmentation and Cue Phrases 504 16.3 Discourse Structure and Reference 510 16.4 Relating Discourse Structure and Inference 512 16.5 Discourse Structure, Tense, and Aspect 517 16.6 Managing the Attentional Stack 524 16.7. AnExample 530 Chapter 17 Defining a Conversational Agent 17.1 17.2 17.3 174 17.5 17.6 17.7 17.8 179 What's Necessary to Build a Conversational Agent? 541 Language as a Multi-Agent Activity 543 Representing Cognitive States: Beliefs 545 Representing Cognitive States: Desires, Intentions, and Plans 551 Speech Acts and Communicative Acts 554 Planning Communicative Acts 557 Communicative Acts and the Recognition of Intention $61 The Source of Intentions in Dialogue 564. Recognizing Mocutionary Acts 567 17.10 Discourse-Level Planning 570xX CONTENTS APPENDIX A An Introduction to Logic and Model-Theoretic Semantics A.1 Logic and Natural Language 579 A.2 Model-Theoretic Semantics 584 A3_ A Semantics for FOPC: Set-Theoretic Models 588 APPENDIX B Symbolic Computation B.l Symbolic Data Structures 595 B.2 Matching 598 B.3 Search Algorithms 600 BA Logic Programming 603 B.5 The Unification Algorithm 604 APPENDIX C Speech Recognition and Spoken Language C.L_ Issues in Speech Recognition 611 C.2. The Sound Structure of Language 613 C3 Signal Processing 616 C4 Speech Recognition 619 CS Speech Recognition and Natural Language Understanding 623 C.6 Prosody and Intonation 625 BIBLIOGRAPHY 629 INDEX 645Preface The primary goal of this book is to provide a comprehensive, in-depth description of the theories and techniques used in the field of natural language understanding. To do this in a single book requires eliminating the idiosyncratic complexities of particular approaches and identifying the underlying concepts of the field as a whole. It also requires developing a small number of notations that can be used to describe a wide range of techniques. It is not possible to capture the range of specific notations that are used in the literature, and attempting to do so would mask the important ideas rather than illuminate them. This book also attempts to make as few assumptions about the background of the reader as possible. All issues and techniques are described in plain English whenever possible. As a result, any reader having some familiarity with the basic notions of programming will be able to understand the principal ideas and techniques used. There is enough detail in this book, however, to allow a sophisticated programmer to produce working systems for natural language understanding. The book is intended both as textbook and as a general reference source for those imerested in natural language processing. As a text, it can be used both in under- graduate and graduate courses in computer science or computational linguistics. As a reference source, it can be used by researchers in areas related to natural language, by developers building natural language systems, and by individuals who are simply interested in how computers can process language Work in natural language understanding requires background in a wide range of areas, niost importantly computer science, linguistics, logic, psycholinguistics, and the philosophy of language. As very few people have all this background, this book into- duces whatever background material is required, as needed. For those who have no familiarity with symbolic programming or logical formalisms, two appendices are provided that describe enough of the basic ideas to enable nonprogrammers to tead the book. Background in linguistics and other areas is introduced as needed throughout the book. The book is organized by problem area rather than by technique. This allows the overlap in different approaches to be presented once, and facilitates the comparison of the different techniques. For example, rather than having a chapter on context-free grammars, one on transition network grammars, and another on logic programming techniques, these three techniques are discussed in many chapters, each focusing on a specific set of problems. There is a chapter on basic parsing techniques. then another on using features, and another on handling movement phenomena in language. Each chapter lays out its problem, and then discusses how the problem is addressed in each approach. This way, the similarities and differences between the approaches can be clearly identified.xii PREFACE Why a Second Edition? Thave taught a course based on the first edition regularly since it first appeared. Given developments in the field, 1 began to find it lacking in some areas. In other areas the book had the right content, but not the right emphasis. Another motivation came from what I saw as a potential split in the fietd with the inveduction of statistically-based techniques and corpus-based research. Claims have been made as to how the new techniques make the old obsolete. But on examining the issues carefully, it was clear to me that the old and new approaches are complementary, and neither is a substitute for the other. This second edition attempts to demonstrate this by including new material on statistical methods, with an emphasis on how they interact with traditional schemes. Another motivation was to improve the pedagogical aspects of the book. While the original edition was intended for use by students and researchers at different levels of background and with different needs, its organization sometimes made this difficult. The second edition addresses this problem in several ways, as described below in the section on using this book. As a result, it should be accessible to students with little background—say, undergraduates in a computer science program with little linguistics background, or undergraduates in a linguistics program with only a basic course in programming—yet also comprehensive enough for more advanced graduate-level courses, and for use as a general reference text for researchers in the field. There are many other changes throughout the book, reflecting how the field has changed since the first edition, or at least reflecting how my views have changed. Much of this is a change in emphasis. The sections on syntax, for instance, now use feature- based context-free grammars as the primary formalism for developing the ideas, rather than augmented transition networks. This better reflects current work in the field, and also allows for a better integration of insights from linguistics. Transition network focmatisms are still covered in detail, however, as they underlie much work in the field. The sections on semantics have also changed substantially, focusing more on under~ lying principles and issues rather than specific computational techniques. ‘The logical form language was updated to cover more complex sentences, and the primary method of semantic interpretation is a compositional rule-by-rule analysis. This allows for both a beter discussion of the underlying problems in semantic interpretation and a better integration of insights from linguistics. New material has also been added on domain- specific semantic interpretation techniques that have been an active focus in practical systems in the last few years. The material on contextual processing and discourse has generally been rewritten and updated to reflect new work and my better understanding of the principles underlying the techniques in the literature. The other significant change is the development of software to accompany the book. Most of the principal examples and algorithms have been implemented and tested. This code is available so that students can run the examples to get a better understanding of the algorithms. It also provides a strong base from which students can extend algorithms and develop new grammars for assignments. The code also helps to clarify and simplify the discussion of the algorithms in the book, both because they can be more precise, being based on running programs, and because implementation details can be omitted from the text and still be available to the student in the code.PREFACE iii How to Use This Book To cover the entire book in detail would require a two-semester sequence at the under- graduate level. As a graduate seminar, it is possible to move more quickly and cover most of the book in a single semester. The most common use of this book, however, is as an undergraduate-level course for a single semester. The book is designed to make this possible, and to allow the instructor to choose the areas on which to focus. There are several mechanisms to help customize the book to your needs. First, each chapter is divided into core material and optional material. Optional sections are marked with an open dot (0). In general, if you are using a chapter, all of the core materiat should be covered, The optional material can be selected as best fits the design of your course. Additional optional material is included in boxes. Material in boxes is never required, ‘but can be used to push a little deeper in some areas. The chapters themselves have the general dependencies shown in the figure on the next page. This means that the student should cover the core material in the earlier chapters to best understand the later chapters. In certain cases, material in a section may require knowledge of a technique discussed earlier but not indicated in the dependencies shown here. For example, one of the techniques discussed in Chapter 11 uses some partial parsing techniques described in Chapter 6. In such a case, the earlier section will need to be covered even though the chapter as a whole is skipped. Readers familiar with the first edition will notice that the two finat chapters, dealing with question answering and generation, are not present in the second edition. In keeping with the organization by topic rather than technique, the material from these chapters has been integrated into other chapters. Sections dealing with generation appear as each topic area is discussed. For example, a head-driven realization algorithm is described in Chapter 9 on linking syntax and semantics, and the procedural semantics technique used in database query systems appears in Chapter 13 on knowledge representation techniques. Software The software accompanying this book is available via anonymous ftp from be.aw.com in the subdirectory “be/allen.” All software is written in standard COMMON LISP, and should run with any COMMON LISP system. To retrieve the software, fip to be.aw.com, by typing fip be.aw.com and log in as “anonymous.” Change to the directory for this book by typing cd be/allen Before retrieving the files, it is a good idea to look at the “readme” fite to see if changes have been made since this book went to press. To retrieve this file, type get README,xiv PREFACE T. Introduction, 2, Linguistic Background, 3, Grammars and Parsing, 4, Features and Augmented Grammars §. Grammars for Natural Language 6. Efficient Parsing 7. Ambiguity Resolution: Statistical Methods 8, Semantics and Logical Form, 9, Linking Symtax and Semantics ‘“ 10. Ambiguity Resolution 1. Other Strategies for Semantic Imerpret: 12. Scoping and the Interpretation of Noun Phrases 13. Knowledge Representation —> [16. Discourse Structure 15. Using Knowledge] ——» |17. A Conversational Agent Quit ftp to tog off and read the file. (Although the file can be read online, it is courteous not to tie up the login for reading.) Log back on when you are ready to download. You can also get a listing of filenames available using either the conventional UNIX “Is” command or the DOS “dit” command. Assuming that you are in the proper subdirectory, these commands require no arguments: dir or Is Using ftp and then de-archiving files can get complicated. Instructions vary as to whether you are downloading Macintosh, DOS, or UNIX files. More instructions are included in the README file. If you ate new to using ftp, it is best to consult your favorite UNIX guide or wizard,PREFACE XV References This book contains extensive references to literature in this field, so that the particular detaits of the techniques described here can easily be found. In addition, references are provided to other key papers that deal with issues beyond the scope of the book to help identify sources for research projects. References are chosen based on their general availability. Thus I have tried to confine citations (o journal articles, books, and major conferences, avoiding technical reports and unpublished notes that are hard to find ten years after the fact. In generat, most work in the field is published first in technical Teports; more accessible publications often do not appear until many years after the ideas were introduced, Thus the year of publication in the references is not a reliable indicator of the time when the work was done, or when it started to influence other work in the field. While I've tried to provide a wide range of sources and to give credit to the ideas described in this book, I know I have omitted key papers that | will regret immensely when I remember them. To these authors, I extend my apologies. Acknowledgments This book would not have been completed except for a sabbatical from the University of Rochester, and for the continual support of the Computer Science Department over the last three years. I was provided with the extra resources needed to complete the project without question, and | am very grateful. I also must thank Peggy Meeker for her efforts in editing and preparing endless drafts as well as the final pages you sce here. Peggy has had amazing patience and good humor through countless revisions, and has enforced a consistency of notation and grammaticat style that appears to be foreign to my natural way of writing! Thave been fortunate to have a wide range of reviewers on various drafts, both those working for the publisher and many other friends who have taken much time to tead and comment on drafts. The tist is long, and each one has devoted such effort that listing their names hardly seems sufficient. But here goes. Many thanks to Glean Blank, Eugene Charniak, Michelle Degraff, George Ferguson, Mary Harper, Peter Heeman, Chung-Hee Hwang, Susan McRoy, Brad Miller, Johanna Moore, Lynn Peterson, Victor Raskin, Len Schubert, Mark Steedman, Mike Tanenhaus, David ‘Traum, and the students in CSC 447 who used the book in the Spring of 1994, The book is significantly different and improved from the early drafts because of all the honest and critical feedback I received. Finally, I must thank my wife, Judith Hook, for putting up with all this for a second time after experiencing the trauma of doing the first edition. Pushing the book to completion became an obsession that occupied me day and night for nearly a year. Thanks to Judi, and to Jeffrey, Daniel, and Michael, for putting up with me during this time and keeping an appropriate sense of humor about it. James F. Allen Rochester, New YorkCHAPTER Introduction to Natural Language Understanding 11 The Study of Language 1.2 Applications of Natural Language Understanding 13 Evaluating Language Understanding Systems 14 The Different Levels of Language Analysis 15 Representations and Understanding 1.6 The Organization of Natural Language Understanding SystemsIntroduction to Natural Language Understanding 1 11 This chapter describes the field of natural language understanding and introduces some basic distinctions. Section 1.1 discusses how natural language understand- ing research fits into the study of language in general. Section 1.2 discusses some applications of natural language understanding systems and considers what it means for a systema to understand language. Section 1.3 describes how you might evaluate whether a system understands language. Section 1.4 introduces a few basic distinctions that are made when studying language, and Section 1.5 discusses how computational systems often realize these distinctions. Finally, Section 1.6 discusses how natural language systems are generally organized, and introduces the particular organization assumed throughout this book. The Study of Language | aspects of human behavior and is a crucial component of our lives. In written form it serves as a long-term record of knowl- edge from one generation to the next. In spoken form it serves as our primary means of coordinating our day-to-day behavior with others. This book describes research about how language comprehension and production work. The goal of this research is to create computational models of language in enough detail that you could write computer programs to perform various tasks involving natural language. The ultimate goal is to be able to specify models that approach human performance in the linguistic tasks of reading, writing, hearing, and speaking. This book, however, is not concerned with problems related to the specific medium used, whether handwriting, keyboard input, or speech. Rather, it is con- cerned with the processes of comprehending and using language once the words are recognized. Computational models are useful both for scientific purposes— for exploring the nature of linguistic communication—and for practical pur- poses—for enabling e icati ‘es tens rel for instance, studies the structure of language itself, consider- ing questions such as why certain combinations of words form sentences but others do not, and why a sentence can have some meanings but not others. The psycholingnist, on the other hand, studies the processes of human language pro- duction and comprehension, considering questions such as how people identify the appropriate structure of a sentence and when they decide on the appropriate meaning for words. The philosopher considers how words can mean anything at all and how they identify objects in the world, Philosophers also consider what it means to have beliefs, goals, and intentions, and how these cognitive capabilities relate to language. The goal of the computational linguist is to develop a compu- tational theory of language, using the notions of algorithms and data structures from computer science. Of course, to build a computational model, you must take advantage of what is known from all the other disciplines. Figure 1.1 summarizes these different approaches to studying language.CHAPTER 1 Discipline Typical Problems Tools Linguists How do words form phrases Intuitions about well-formedness and sentences? What constrains and meaning; mathematical models. the possible meanings for a of structure (for example, formal sentence? language theory, model theoretic semantics) Psycholinguists How do people identify the Experimental techniques based on suucture of sentences? How are measuring human performance; ‘word meanings identified? When statistical analysis of observations does understanding take place? Philosophers What is meaning, and how do —_- Natural language argumentation words and sentences acquire it? using intuition about counter- How do words identify objects examples; mathematical models in the world? (for example, logic and model theory) Computational How is the structure of sentences Algorithms, data siructures; Linguists identified? How can knowledge formal models of representation and reasoning be modeled? and reasoning; Al techniques How can language be used io (search and representation methods) accomplish specific tasks? Figure Ll The major disciplines studying language ioned, there : wo motivations for developing computational models. The scientific motivation is to obtain a better understand- “ing of how language works. It recognizes that any one of the other traditional disciplines does not have the tools to completely address the problem of how language comprehension and production work. Even if you combine all the approaches, a comprehensive theory would be too complex to be studied using traditional methods. But we may be able to realize such complex theories as computer programs and then test them by observing how well they perform. By seeing where they fail, we can incrementally improve them. Computational models may provide very specific predictions about human behavior that can then be explored by the psycholinguist. By continuing in this process, we may eventually acquire a deep understanding of how human language processing occurs. To realize such a dream will take the combined efforts of linguists, psychotinguists. philosophers. and computer scientists. This common goal has motivated a new area of interdiscipli _ The practical, or technological, motivation is that natural language proces~ “sing capabilities would revolutionize the way computers are used. Since most of human knowledge is recorded in linguistic form, computers that could understand natural language could access all this information. In addition, natural language interfaces to computers would allow complex systems to be accessible to12 Introduction to Natural Language Understanding 3 BOX 1.1 Boxes and Optional Sections This book uses several techniques to allow you to identify what material is central and what is optional. In addition, optional material is sometimes classified as advanced, indicating that you may need additional background not covered in this book to fully appreciate the text. Boxes, like this one, always contain optional material, either providing more detail on a particular epproach discussed in the main text or discussing additional issues that are related to the text. Sections and subsections may be marked as optional by means of an open dot (c) before che beading. Optional sections provide more breadth and depth to chapters, but are not necessary for understanding material in later chapters. Depending on your interests and focus, you can choose among the optional sections to fill out the core material presented in the regular sections. In addition, there are dependencies between the ‘chapters, so that entire chapters can be skipped if the material does not address your interests, The chapter dependencies are not marked explicitly in the text, but a chart of dependencies is given in the preface. everyone, Such systems woutd be considerably more flexible and intelligent than is possible with current computer technology. For technological purposes it does not matter if the model used reflects the way humans process language. It only matters that it works. This book takes a middle ground between the scientific and technological goats, On the one hand, this reflects a belief that natural language is so complex that an ad hoc approach without a well-specified underlying theory will not be successful. Thus the technological goal cannot be realized without using sophis- licated underlying theories on the level of those being developed by linguists, psycholinguists, and philosophers. On the other hand, the present state of knowl - edge about natural language processing is so preliminary that attempting to build a cognitively correct model is not feasible. Rather, we are still attempting to constrict any model that appears to work. The goal of this book is to describe work that aims to produce linguistically motivated computational models of language understanding and production that can be shown to perform well in specific example domains. While the book focuses on computational aspects of language processing, considerable space is spent introducing the relevant background knowledge from the other disciplines that motivates and justifies the computational approaches taken. It assumes only a basic knowledge of programming, although the student with some background in linguistics, artificial intelligence (Al), and logic will appreciate additional subtleties in the development. Applications of Natural Language Understanding A good way to define natural language research is to consider the different applications that researchers work on. As you consider these examples, it willCHAPTER 1 also be a good opportunity to consider what it would mean to say that a computer system understands natural language. The applications can be divided into two “major classes: text-based applications and dialogue-based applications. Text-based applications involve the processing of written text, suc! books, newspapers, reports, manuals, e-mail messages, and so on. The -reading-based tasks. Text-based natural language research is ongoing in appli cations such as |* finding appropriate documents on certain topics from a data- base of texts (for example, finding relevant books in a library) (+ extracting information from messages or articles on certain topics (for example, building a database of all stock transac- tions described in the news on a given day) * — tanslating documents from one language to another (for ex- ample, producing automobile repair manuals in many different languages) © summarizing texts for certain purposes (for example, producing 4 3-page summary of a L000-page government report) Not all systems that perform such tasks must be using natural language understanding techniques in the way we mean in this book. For example, consider the task of finding newspaper articles on a certain topic in a large database. Many techniques have been developed that classify documents by the presence of certain keywords in the text. You can then retrieve articles on a certain topie by looking for articles that contain the keywords associated with that topic. Articles on law, for instance, might contain the words lawyer, court, sue, affidavit, and so oa, while artictes on stock transactions might contain wotds such as stocks, takeover, leveraged buyout, options, and so on. Such a system could retrieve articles on any topic that has been predefined by a set of keywords. Clearly, we would not say that this system is understanding the text, rather, it is using a simple matching technique. While such techaiques may produce useful applications, they are inherently limited. It is very unlikely, for example, that they could be extended to handle complex retrieval tasks that are easily expressed in natural language, such as the query Find me all articles on leveraged buyouts involving more than 100 million dollars that were attempted but failed during 1986 and 1990. To handle such queries, the system would have to be able to extract enough information from each article in the database to determine whether the article meets the criteria defined by the query; that is, it would have to build a representation of the information in the articles and then use the representation to do the retrievals. This identifies a crucial characteristic of an understanding system: it must compute some representation of the information that can be used for later inference. Consider another example. Some machine translation systems have been built that are based on pattern matching; that is, a sequence of words in one language is associated with a sequence of words in another language. TheIntroduction to Natural Language Understanding 5 uanslation is accomplished by finding the best set of patterns that match the input and producing the associated output in the other language. This technique can produce reasonable results in some cases but sometimes produces completely wrong translations because of its inability co use an understanding of content to disambiguate word senses and sentence meanings appropriately. In contrast, other machine translation systems operate by producing a representation of the meaning of each sentence in one language, and then producing a sentence in the other language that realizes the same meaning. This latter approach, because it involves the computation of a representation of meaning, is using natural Tanguage understanding techniques. a very attractive domain for text-based research is story understanding. In this task the system processes a story and then must answer questions about it. This is similar to the type of reading comprehension tests used in schools and provides a very rich method for evaluating the depth of understanding the system isable to achieve. - Dialogue-based applications involve human-machine com communication. Most nguage, but it also includes interaction using ds, Typical potential applications include | question-answering systems, where natural language is used 10 query a database (for example, a query system to a personnel database) * automated customer service over the telephone (for example. to perform banking transactions or order items from a catalogue) tutoring systems, where the machine interacts with a student (for example, an automated mathematics tutoring system) © spoken language control of a machine (for example, voice con- trol of a VCR or computer) © general cooperative problem-solving systems (for example, a system that helps a person plan and schedule freight shipments) Some of the problems faced by dialogue systems are quite different than in text-based systems. First, the language used is very different, and the system needs to participate actively in order to maintain a natural, smooth-flowing dialogue. Dialogue requires the use of acknowledgments to verify that things are understood, and an ability to both recognize and generate clarification sub- dialogues when something is not clearly understood. Even with these differences, however, the basic processing techniques are fundamentally the same. Its important to distinguish the problems of speech recognition from the problems of language understanding. A spees ition system need not involve any language understanding, For instance, voice-controlled computers and VCRs are entering the market now. These do not involve natural language understanding in any general way. Rather, the words recognized are used as commands, much like the commands you send to.a VCR using a remote control. Speech recognition is concemed only with identifying the words spoken from a6 — cHaprer1 given speech signal, not with understanding how words aré used to communicate. To be an understanding system, the speech recognizer would need to feed its input to a natural language understanding system, producing what is often called a spoken language understanding system. With few exceptions, all the techniques discussed in this book are equally relevant for text-based and dialogue-based language understanding, and apply equally well whether the input is text, keyboard, or speech. The key cha derstanding system is that it represents the meaning of sentences in presentation language that can be used later for further processing 1.3. Evaluating Language Understanding Systems As you can see, what counts as understanding might vary from application to application. If this is so, how can you tell if a system works? One obvious way to evaluate a system is to rum the program and see how well it performs the task it was designed (o do, If the program is meant to answer questions about a database of facts, you might ask it questions to see how good itis at producing the correct answers. If the system is designed to participa in simple conversations on a early evaluation reals can be alsa! Sometimes the techniques that pro~ duce the best results in the short term will not lead to the best results in the long term, For instance, if the overall performance of all known systems in a given application is uniformly low, few conclusions can be drawn. The fact that one system was correct 50 percent of the time while another was correct only 40 percent of the time says nothing about the long-term viability of either approach. Only when the success rates become high, making a practical application feasi- ble, can much significance be given to overall system performance measures. An alternative method of evaluation is to identify various subcomponents te ¢ pproptia is ass si¢ th stem. The prob- various ge system should be. consensus is an area of considerable activity at the present. Despite the difficulties, evaluations are of crucial importance to the field, as very little can be concluded from examining a few examples that a particular system can handle. If you cannot interact freely with the program yourself but must depend on transcripts of others’ interactions, you can easily be mislead into overestimating a program’s abilities. It is very hard for people not to attribute much more intelligence to the program than is actually present. ple—the : ELIZA program that was developed in must be one of the most popular AlIntroduction to Natural Language Understanding, 7 Users Men are all alike, ‘They're always bugging us about something or other. Well, my boyfriend made me come here, He says I'm depressed much of the time, It’s true, J am unhappy. T need some help; that much seems certain. ELIZA: In what way? Can you think of a specific example? Your boyfriend made you come here, Lam sorry to hear you are depressed. Do you think coming here will help you not to be unhappy? ‘What would it mean to you if you ot some help? Perhaps { could learn to get along with my mother. My mothes takes care of me My father. You are like my father in some ways. Tell me more about your fara Who else in your family takes care of you? Your father. What sesemblance do you see? Figure L2 A dialogue with ELIZA and versions of it e 1 "personal computers. ELIZA was never claimed to embody a theory of language comprehension and production, but it serves as an excellent example as it transcript, or even playing with the system yourself for a few minutes, ELIZA's performance certainly seems impressive. : Here is a simple description of how ELIZA works. There is a database of | __ particular words that are called keywords. For each keyword, the system stores an integer, a pattern to match against the input, and a specification of the output. ‘The algorithm is as follows: Given a sentence S, find a keyword in $ whose pattern matches S. If there is more than one keyword, pick the one with the | highest integer value. Use the output specification that is associated with this keyword to generate the next sentence. If there are no keywords, generate an |. innocuous continuation statement, such as Tell me more or Go on. Figure 1.3 shows a fragment of a database of keywords. In this database a patiern consists of words and variables. The prefix ? before a letter indicates a variable, which can match any sequence of words. For example, the pattern ‘2X are you 7 would match the sentence Why are you looking at me?, where the variable ’X matches Why and 2Y matches looking at me. The output specification may also use the same variables. In this case, ELIZA insert ‘s the words that match the Variables in the input into the output after making some minor changes in the8 — cHapreri Outputs ™ In what way? What resemblance do you see? 2X are you 2¥ Would you prefer it if I weren't ?Y? °K are ¥ What if they were not ?Y? * Can you think of a specific example? Whea? Really, always? Why do you ask? Does that interest you? Figure 1.3 Sample data from ELIZA, ‘pronouns (for example, replacing me with you), Thus, for the pattern above, if the output specification is Would you preferit if I weren't 2¥? the rule woutd generate a response Would you prefer it if I weren't looking at you? When the database lists multiple output speci s stern, ELIZA selects a different one each time a keyword rule is used, thereby prevent- ing unnatural repetition in the conversation. Using these rules, you can see how ELIZA produced the first two exchanges in the conversation in Figure 1.2, ELIZA generated the first response from the first output of the keyword alike and the second response from the first output of the keyword always. This description covers all of the essential points of the program. You will probably agree that the program does not understand the conversation itis partci- in, Rather, it is a collection of tricks. Why then does ELIZA appear to so well? There are several reasons. Perhaps the most important reason is that, when people hear or read a sequence of words that they understand as a sentence, they attribute meaning to the sentence and assume that the person (or machine) that produced the sentence actually intended that meaning. People are extremely good at distinguishing word meanings and interpreting sentences to fit the context. Thus ELIZA aj be intelligent because you use your own ‘intelligence to ee what it says. Other crucial characteristics of the conversational setting also aid in sus- taining the illusion of intelligence. For instance, the system does not need any world knowledge because it never has to make a claim, support an argument, or answer a question. Rather, it simply asks a series of questions. Except in a patient-therapist situation, this would be unacceptable, ELIZA evades all direct questions by responding with another question, such as Why do you ask? There is no way to force the program to say something concrete about any topic,Introduction to Natural Language Understanding 9 Even in such a restricted situation, however, it is relatively easy to demonstrate that the program does not understand. It sometimes produces com- pletely off-the-wall responses. For instance, if you say Necessity is the mother of might respond with Tell me more about your family, based on its pattern for the word mother. In addition, since ELIZA has no knowledge about the structure of language, it accepts gibberish just as readily as valid sentences. If you enter Green the adzabak are the a ran four, ELIZA will respond with something like Whar if they were not the a ran four? Also, as a conversation progresses, it becomes obvious that the program does not retain any of the ntent in the conversation. It begins to ask questions that are inappropriate in light of earlier exchanges, and its responses in general begin to show a lack of focus. Of course, if you are not able to play with the program and must depend only on transcripts of conversations by others, you would have no way of detecting these flaws, unless they are explicitly mentioned. Suppose you need to build a natural language program for a certain application in only six months. If you start to construct a general mode] of language understanding, it will not be completed in that time frame and so will perform miserably on the tests. An ELIZA-like system, however, could easily produce behavior like that previously discussed with less than a few months of programming and will appear to far outperform the other system in testing, The differences will be especially marked if the test data only includes typical domain interactions that are not designed to test the limits of the system. Thus, if we take short-term performance as our only criteria of progress, everyone will build and fine-tune ELIZA-style systems, and the field will not progress past the limitations of the simple approach. To avoid this problem, either we have to accept certain theoretical assump- tions about the architecture of natural language systems and develop specific evaluation measures for different components, or we have to discount overall evaluation results until some reasonably high level of performance is obtained. ‘Only then will cross-system comparisons begin to reflect the potential for long- term success in the field. 1.4. The Different Levels of Language Analysis @ A natural language system must use considerable knowledge about the structure ___ of the language itself, including what the words are, how words combine to form & | sentences, what the words mean, how word meanings contribute to sentence meanings, and so on. However, we cannot completely account for linguistic _ behavior without also taking into account another aspect of what makes humans intelligemt—their general world knowledge and their reasoning abilities. For ‘example, to answer questions or to participate in a conversation, a person not only must know a lot about the structure of the language being used, but also must know about the world in general and the conversational setting in particular.10 CHAPTER 1 The following are some of the different forms of knowledge relevant for natural language understanding: Phonetic and phonological knowledge—concems how words are related to the sounds that realize them. Such knowledge is crucial for speech-based systems and is discussed in more i detail in Appendix C Morphological knowledge—concems how words are constructed from more basic meaning units called morphemes. A mor- pheme is the primitive unit of meaning ia a language (for example, the meaning of the word friendly is derivable from the meaning of the noun friend and the suffix -éy, which transforms a noun into an adjective). Syntactic knowledge—concerns how words can be put together to form correct sentences and determines what structural role each word plays in the sentence and what phrases are subparts of what other phrases. Semantic knowledge—concerns what words mean and how these meanings combine in sentences to form sentence meanings. This is the study of comtext-independent meaning—the mean- ing a sentence has regardless of the context in which it is used. Pragmatic knowledge—concems how sentences are used in 4 different situations and how use affects the interpretation of the | sentence, Discourse knowledge—concems how the immediately preceding sentences affect the interpretation of the next sentence. This | information is especially important for interpreting pronouns and for interpreting the temporal aspects of the information conveyed. ~ World knowledge— includes the general knowledge about the struc- ture of the world that language users must have in order to, for example, maintain a conversation. It includes what each lan- guage user must know about the other user's beliefs and goals. These definitions are imprecise and are more characteristics of knowledge than actual distinct classes of knowledge. Any particular fact might include aspects from several different levels, and an algorithm might need to draw from several different levels simultaneously. For teaching purposes, however, this book is organized into three parts, each describing a set of techniques that natu- rally cluster together. Part I focuses on syntactic and morphological processing, Part I focuses on semantic processing, and Part III focuses on contextual effects cluding pragmatics, discourse, and world knowledge.Introduction to Natural Language Understanding = 11. BOX 1.2 Syntax, Semantics, and Pragniatics ‘The following examples may help you understand the distinction berween syntax, semantics, and pragmatics. Consider each example as a candidate for the initial sentence of this book, which you know discusses natural language processing: 1, Language is one of the fundamental aspects of human behavior and is a crucial component of our lives, 2 Green frogs have large noses. 3 Green ideas have large noses 4, Large have green ideas nose, Sentence | appears to be a reasonable start (1 hope!). It agrees with all that is known about syntax, semantics, and pragmatics. Each of the other sentences violates one or more of these levels. Semence 2 is well-formed syntactically and semantically, but not pragmatically. It fates poorly as the first sentence of the book because the reader would find no reason for using it. But however bad semtence 2 would be as a slart, sentence 3 is much worse, Not only is it obviously prag- matically ill-formed, it is also semantically il-formed. To see this, consider that ‘you and T could argue about whether sentence 2 is tue of not, but we cannot do so with sentence 3, T cannot affirm or deny semence 3 in coherent conversation, However, the sentence does have some structure, for we cun discuss what is wrong ‘with it: Tdeas cannot be green and, even if they could, they certainly cannot have large noses. Sentence 4 is even worse. In fact, it is unintelligible, even though it contains the same words as sentence 3. It does not even have enough simncture to allow you to say what is wrong with it. Thus it is syntactically ill-formed. Inci- dentally, there are cases in which 4 sentence may be ptagmatically well-formed but not syntactically well-formed. For example, if I ask you where you are going and you reply “I go store,” the response would be understandable even though it is syntactically ill-formed. Thus it is at least pragmatically well-formed and may even be semantically well-formed. 1.5. Representations and Understanding AAs previously stated, a crucial component of understanding involves computing @ nces and texts. Without defining the notion of representation, however, this assertion has little content, For instance, why not simply use the sentence itself as a representation of its meaning? One reason is that most words have multiple meanings, which we will call senses. The word cook, for example, has a sense as a verb and a sense as a noun; dish has multiple senses as a noun as well as a sense as a verb; and still has senses as a noun, verb, adjective, and adverb. This ambiguity would inhibit the system from making the appropriate inferences needed 10 model understanding. The disambiguation problem appears much easier than it actually is because peopte do not generally notice ambiguity. While a person does not seem 10 consider each of the possible12 CHAPTER] senses of a word when understanding a sentence, a program must explicitly consider them one by one. To represent meaning, we must have a more precise language. The tools to do this come from mathematics and logic and involve the use of formally speci- fied representation languages. Formal Janguages are specified from very simple building blocks. The most fundamental is the notion of an atomic symbol, which is distinguishable from any other atomic symbol simply based on how it is written. Useful representation languages have the following two properties: + The representation must be precise and unambiguous. You should be able to express every distinct reading of a sentence as, a distinct formula in the representation. (The representation should capture the intuitive structure of the natural language sentences that it represents. For example, sentences that appear to be structurally similar should have similar structural representations, and the meanings of two sentences that are paraphrases of each other should be closely related to each other. Several different representations will be used that correspond to some of the levels of analysis discussed in the last section. In particular, we will develop formal languages for expressing syntactic structure, for context-independent word and sentence meanings, and for expressing general world knowledge. Syntax: Representing Sentence Structure The syntactic structure of a sentence indicates the way that words in the sentence are related to each other. This structure indicates how the words are grouped together into phrases, what words modify what other words, and what words are of central importance in the sentence. In addition, this structure may identify the types of relationships that exist between phrases and can store other information "about the particular sentence structure that may be needed for later processing. For example, consider the following sentences: 1. John sold the book to Mary. 2 The book was sold to Mary by John. ‘These sentences share certain structural properties. In each, the noun phrases are John, Mary, and the book, and the act described is some selling action. In other respects, these sentences are significantly different. For instance, even though both sentences are always cither true or false in the exact same situations, you could only give sentence I as an answer (o the question What did John do for Mary? Sentence 2 is a much better continuation of a sentence beginning with the phrase After ir fell in the river, as sentences 3 and 4 show. Following the standard convention in linguistics, this book will use an asterisk (*) before any example of an ill-formed or questionable sentence.Introduction to Natural Language Understanding 13 Figure 1.4 Two structural representations of Rice fies like sand. 3. *After it fell in the river, John sold Mary the book. 4, After it fell in the river, the book was sold to Mary by John. Many other structural properties can be revealed by considering sentences that are not well-formed. Sentence 5 is ill-formed because the subject and the verb do not agree in number (the subject is singular and the verb is plural), while 6 is ill-formed because the verb pur requires some modifier that describes where John put the object. 5. *John are in the comer. 6 John put the book. Making judgments on grammaticality is not a goal in natural language understanding. In fact, a robust system should be able to understand ill-formed sentences whenever possible. This might suggest that agreement checks can be ignored, but this is not so. Agreement checks are essential for eliminating poten- tial ambiguities. Consider sentences 7 and 8, which are identical except for the number feature of the main verb, yet represent two quite distinct interpretations, 7. Flying planes are dangerous. 8. Flying planes is dangerous. If you did not check subject-verb agreement, these two sentences would be indis- tinguishable and ambiguous. You could find similar examples for every syntactic feature that this book introduces and uses. Most syntactic representations of language are based on the notion of context-free grammars, which represent sentence structure in terms of what phrases are subparts of other phrases. This information is often presented in a tee form, such as the one shown in Figure 1.4, which shows two different structures
You might also like
Natural Language Processing: Dr. Abdulfetah A.A
PDF
No ratings yet
Natural Language Processing: Dr. Abdulfetah A.A
25 pages
Unit -1 NLP-R20
PDF
No ratings yet
Unit -1 NLP-R20
10 pages
Automata Theory
PDF
No ratings yet
Automata Theory
48 pages
Natural Language Processing
PDF
No ratings yet
Natural Language Processing
20 pages
Text Processing, Tokenization & Characteristics
PDF
100% (1)
Text Processing, Tokenization & Characteristics
89 pages
Parsing
PDF
No ratings yet
Parsing
38 pages
Question Bank
PDF
No ratings yet
Question Bank
13 pages
NLP Unit 5
PDF
No ratings yet
NLP Unit 5
10 pages
Compiler Design
PDF
No ratings yet
Compiler Design
188 pages
Advances in Computers, Vol.72, High Performance Computing (AP, 2008) (ISBN 0123744113) (369s) - CsAl
PDF
No ratings yet
Advances in Computers, Vol.72, High Performance Computing (AP, 2008) (ISBN 0123744113) (369s) - CsAl
369 pages
03 Cet 404 Comprehensive Course Viva Slot e S8
PDF
No ratings yet
03 Cet 404 Comprehensive Course Viva Slot e S8
1 page
NLP: Background and Overview: Introduction To Natural Language Processing (CSE5321)
PDF
No ratings yet
NLP: Background and Overview: Introduction To Natural Language Processing (CSE5321)
30 pages
Tcs Paper Soln
PDF
No ratings yet
Tcs Paper Soln
37 pages
Semantic Analysis: Natural Language Processing (CSE 5321)
PDF
No ratings yet
Semantic Analysis: Natural Language Processing (CSE 5321)
35 pages
Context-Free Languages & Grammars (Cfls & CFGS) : Reading: Chapter 5
PDF
No ratings yet
Context-Free Languages & Grammars (Cfls & CFGS) : Reading: Chapter 5
40 pages
Natural Language Processing Inside Pages 2
PDF
No ratings yet
Natural Language Processing Inside Pages 2
159 pages
NLP Akash
PDF
No ratings yet
NLP Akash
4 pages
Information Retrieval, Question Answering Systems, and Chatgpt: Technology, Capability, and Intelligence
PDF
No ratings yet
Information Retrieval, Question Answering Systems, and Chatgpt: Technology, Capability, and Intelligence
15 pages
Natural Language Processing-Wiki
PDF
No ratings yet
Natural Language Processing-Wiki
237 pages
Compiler Design LectureNotes
PDF
No ratings yet
Compiler Design LectureNotes
45 pages
The Theory of Automata in Natural Language Processing
PDF
No ratings yet
The Theory of Automata in Natural Language Processing
6 pages
Be Computer Engineering Semester 7 2023 May Dloc III Natural Language Processing Rev 2019 C Scheme
PDF
0% (1)
Be Computer Engineering Semester 7 2023 May Dloc III Natural Language Processing Rev 2019 C Scheme
2 pages
Word Semantics, Sentence Semantics and Utterance Semantics
PDF
No ratings yet
Word Semantics, Sentence Semantics and Utterance Semantics
11 pages
Conversion DFA To Regular Expression
PDF
No ratings yet
Conversion DFA To Regular Expression
6 pages
Natural Language Processing
PDF
No ratings yet
Natural Language Processing
49 pages
Advances in Computers, Vol.65 (Elsevier, 2005) (ISBN 9780120121656) (447s) - CsAl
PDF
No ratings yet
Advances in Computers, Vol.65 (Elsevier, 2005) (ISBN 9780120121656) (447s) - CsAl
447 pages
Using A Hierarchy of Domain Specific Languages in Complex Software Systems Design
PDF
No ratings yet
Using A Hierarchy of Domain Specific Languages in Complex Software Systems Design
8 pages
What Is Natural Language Processing?
PDF
No ratings yet
What Is Natural Language Processing?
5 pages
Natural Language Processing
PDF
No ratings yet
Natural Language Processing
5 pages
NLP Unit I
PDF
No ratings yet
NLP Unit I
30 pages
The Go Programming Language Specification - The Go Programming Language
PDF
No ratings yet
The Go Programming Language Specification - The Go Programming Language
95 pages
Theory of Computation - Context Free Languages
PDF
No ratings yet
Theory of Computation - Context Free Languages
25 pages
Qurana: Corpus of The Quran Annotated With Pronominal Anaphora
PDF
No ratings yet
Qurana: Corpus of The Quran Annotated With Pronominal Anaphora
8 pages
NLP Unit 1 and 2
PDF
No ratings yet
NLP Unit 1 and 2
106 pages
Closure Properties of Context Free Languages (Proof)
PDF
No ratings yet
Closure Properties of Context Free Languages (Proof)
2 pages
NATURAL LANGUAGE PROCESSING IN Education
PDF
No ratings yet
NATURAL LANGUAGE PROCESSING IN Education
17 pages
Concrete Semantics With Isabelle/HOL
PDF
No ratings yet
Concrete Semantics With Isabelle/HOL
308 pages
7 Knowledge Representation
PDF
No ratings yet
7 Knowledge Representation
42 pages
Natural Language Processing
PDF
100% (1)
Natural Language Processing
3 pages
Semantic Information Retrieval
PDF
No ratings yet
Semantic Information Retrieval
168 pages
NLP Textbook Star Edu
PDF
No ratings yet
NLP Textbook Star Edu
103 pages
24 Steps of Compiler Design
PDF
No ratings yet
24 Steps of Compiler Design
11 pages
MG 02 1 FiniteAutomata Anim
PDF
No ratings yet
MG 02 1 FiniteAutomata Anim
12 pages
Faster Generalized LR Parsing PDF
PDF
No ratings yet
Faster Generalized LR Parsing PDF
15 pages
Intelligent Compilers
PDF
No ratings yet
Intelligent Compilers
9 pages
CS 464 Question Bank
PDF
No ratings yet
CS 464 Question Bank
5 pages
JACL: A Common Lisp For Developing Single-Page Web Applications - Alan Dipert
PDF
No ratings yet
JACL: A Common Lisp For Developing Single-Page Web Applications - Alan Dipert
6 pages
NLP Iat QB
PDF
No ratings yet
NLP Iat QB
10 pages
Multilingual Information Retrieval
PDF
No ratings yet
Multilingual Information Retrieval
18 pages
NLP Qb-Ese
PDF
No ratings yet
NLP Qb-Ese
2 pages
CD Unit - 1
PDF
No ratings yet
CD Unit - 1
38 pages
Natural Language Processing: Dr. Tulasi Prasad Sariki SCOPE, VIT Chennai
PDF
No ratings yet
Natural Language Processing: Dr. Tulasi Prasad Sariki SCOPE, VIT Chennai
29 pages
AT&FL Lab 11
PDF
No ratings yet
AT&FL Lab 11
6 pages
Formal Languages and Automata Theory QB PDF
PDF
No ratings yet
Formal Languages and Automata Theory QB PDF
5 pages
Compiler Design - Chapter 4 - Syntax Directed Translation
PDF
No ratings yet
Compiler Design - Chapter 4 - Syntax Directed Translation
49 pages
21AD3202 - Natural LanguageProcessing-Record
PDF
No ratings yet
21AD3202 - Natural LanguageProcessing-Record
64 pages
CFG 2
PDF
No ratings yet
CFG 2
6 pages
NLU2e James Allen
PDF
No ratings yet
NLU2e James Allen
459 pages
05 and 06-CCS-lectures.pptx
PDF
No ratings yet
05 and 06-CCS-lectures.pptx
106 pages
Chapter 6
PDF
100% (1)
Chapter 6
28 pages
Introduction To Data Repositories - Unlocked
PDF
No ratings yet
Introduction To Data Repositories - Unlocked
182 pages
Data Analytics Basics - Unlocked
PDF
No ratings yet
Data Analytics Basics - Unlocked
59 pages
MDA Book
PDF
No ratings yet
MDA Book
68 pages
Compiler Final Modified
PDF
No ratings yet
Compiler Final Modified
33 pages
Theory of Computation: Delhi Technological University, Delhi
PDF
No ratings yet
Theory of Computation: Delhi Technological University, Delhi
40 pages
ESE Notes 1.0
PDF
No ratings yet
ESE Notes 1.0
15 pages