File Information
File: 05-lr/acl_arc_1_sum/cleansed_text/xml_by_section/intro/87/j87-3004_intro.xml
Size: 10,878 bytes
Last Modified: 2025-10-06 14:04:37
<?xml version="1.0" standalone="yes"?> <Paper uid="J87-3004"> <Title>COMMONSENSE METAPHYSICS AND LEXICAL SEMANTICS</Title> <Section position="3" start_page="0" end_page="0" type="intro"> <SectionTitle> 1. INTRODUCTION </SectionTitle> <Paragraph position="0"> In the TACITUS project for using commonsense knowledge in the understanding of texts about mechanical devices and their failures, we have been developing various commonsense theories that are needed to mediate between the way we talk about the behavior of such devices and causal models of their operation. Of central importance in this effort is the axiomatization of what might be called &quot;commonsense metaphysics&quot;.</Paragraph> <Paragraph position="1"> This includes a number of areas that figure in virtually every domain of discourse, such as scalar notions, granularity, time, space, material, physical objects, causality, functionality, force, and shape. Our approach to lexical semantics is to construct core theories of each of these areas, and then to define, or at least characterize, a large number of lexical items in terms provided by the core theories. In the TACITUS system, processes for solving pragmatics problems posed by a text will use the knowledge base consisting of these theories, in conjunction with the logical forms of the sentences in the text, to produce an interpretation. In this paper we do not stress these interpretation processes; this is another, important aspect of the TACITUS project, and it will be described in subsequent papers (Hobbs and Martin, 1987).</Paragraph> <Paragraph position="2"> This work represents a convergence of research in lexical semantics in linguistics and efforts in artificial intelligence to encode commonsense knowledge. Over the years, lexical semanticists have developed formalisms of increasing adequacy for encoding word meaning, progressing from simple sets of features (Katz and Fodor, 1963) to notations for predicate-argument structure (Lakoff, 1972; Miller and Johnson-Laird, 1976), but the early attempts still limited access to world knowledge and assumed only very restricted sorts of processing. Workers in computational linguistics introduced inference (Rieger, 1974; Schank, 1975) and other complex cognitive processes (Herskovits, 1982) into our understanding of the role of word meaning. Recently linguists have given greater attention to the cognitive processes that would operate on their representations (e.g., Talmy, 1983; Croft, 1986). Independently, in artificial intelligence an effort arose to encode large amounts of commonsense knowledge (Hayes, 1979; Hobbs and Moore, 1985; Hobbs et al. 1985). The research reported here represents a convergence of these various developments. By constructing core theories of certain fundamental phenomena and defining lexical items within these theories, using the full power of predicate calculus, we are able to cope with complexities of word meaning that have hitherto escaped lexical semanticists.</Paragraph> <Paragraph position="3"> Moreover, we can do this within a framework that gives full scope to the planning and reasoning processes that manipulate representations of word meaning.</Paragraph> <Paragraph position="4"> Copyright 1987 by the Association for Computational Linguistics. Permission to copy without fee all or part of this material is granted provided that the copies are not made for direct commercial advantage and the CL reference and this copyright notice are included on the first page. To copy otherwise, or to republish, requires a fee and/or specific permission. 0362-613X/87/030241-250503.00 Computational Linguistics Volume 13, Numbers 3-4, July-December 1987 241 Jerry R. Hobbs et al. Commonsense Metaphysics and Lexical Semantics In constructing the core theories we are attempting to adhere to several methodological principles: 1. One should aim for characterization of concepts, rather than definition. One cannot generally expect to find necessary and sufficient conditions for a concept.</Paragraph> <Paragraph position="5"> The most we can hope for is to find a number of necessary conditions and a number of sufficient conditions. This amounts to saying that a great many predicates are primitives, but they are primitives that are highly interrelated with the rest of the knowledge base.</Paragraph> <Paragraph position="6"> 2. One should determine the minimal structure necessary for a concept to make sense. In efforts to axiomatize an area, there are two positions one may take, exemplified by set theory and by group theory. In axiomatizing set theory, one attempts to capture exactly some concept that one has strong intuitions about. If the axiomatization turns out to have unexpected models, this exposes an inadequacy. In group theory, by contrast, one characterizes an abstract class of structures.</Paragraph> <Paragraph position="7"> If it turns out that there are unexpected models, this is a serendipitous discovery of a new phenomenon that we can reason about using an old theory. The pervasive character of metaphor in natural language discourse shows that our commonsense theories of the world ought to be much more like group theory than set theory. By seeking minimal structures in axiomatizing concepts, we optimize the possibilities of using the theories in metaphorical and analogical contexts. This principle is illustrated below in the section on regions.</Paragraph> <Paragraph position="8"> One consequence of this principle is that our approach will seem more syntactic than semantic. We have concentrated more on specifying axioms than on constrncting models. Our view is that the chief role of models in our effort is for proving the consistency and independence of sets of axioms, and for showing their adequacy. As an example of the lastpoint, many of the spatial and temporal theories we construct are intended at least to have Euclidean space or the real numbers as one model, and a subclass of graph-theoretical structures as other models.</Paragraph> <Paragraph position="9"> 3. A balance must be struck between attempting to cover all cases and aiming only for the prototypical cases. In general, we have tried to cover as many cases as possible with an elegant axiomatization, in line with the two previous principles, but where the formalization begins to look baroque, we assume that higher processes will block some inferences in the marginal cases.</Paragraph> <Paragraph position="10"> We assume that inferences will be drawn in a controlled fashion. Thus, every outr6, highly context-dependent counterexample need not be accounted for, and to a certain extent, definitions can be geared specifically to a prototype.</Paragraph> <Paragraph position="11"> 4. Where competing ontologies suggest themselves in a domain, one should try to construct a theory that accommodates both. Rather than commit oneself to adopting one set of primitives rather than another, one should show how either set can be characterized in terms of the other. Generally, each of the ontologies is useful for different purposes, and it is convenient to be able to appeal to both. Our treatment of time illustrates this.</Paragraph> <Paragraph position="12"> 5. The theories one constructs should be richer in axioms than in theorems. In mathematics, one expects to state half a dozen axioms and prove dozens of theorems from them. In encoding commonsense knowledge, it seems to be just the opposite. The theorems we seek to prove on the basis of these axioms are theorems about specific situations that are to be interpreted, in particular, theorems about a text that the system is attempting to understand.</Paragraph> <Paragraph position="13"> 6. One should avoid falling into &quot;black holes&quot;. There are a few &quot;mysterious&quot; concepts that crop up repeatedly in the formalization of commonsense metaphysics. Among these are &quot;relevant&quot; (that is, relevant to the task at hand) and &quot;normative&quot; (that is, conforming to some norm or pattern). To insist upon giving a satisfactory analysis of these before using them in analyzing other concepts is to cross the event horizon that separates lexical semantics from philosophy. On the other hand, our experience suggests that to avoid their use entirely is crippling; the lexical semantics of a wide variety of other terms depends upon them. Instead, we have decided to leave them minimally analyzed for the moment and use them without scruple in the analysis of other commonsense concepts. This approach will allow us to accumulate many examples of the use of these mysterious concepts, and in the end, contribute to their successful analysis. The use of these concepts appears below in the discussions of the words &quot;immediately&quot;, &quot;sample&quot;, and &quot;operate&quot;.</Paragraph> <Paragraph position="14"> We chose as an initial target the problem of encoding the commonsense knowledge that underlies the concept of &quot;wear&quot;, as in a part of a device wearing out. Our aim was to define &quot;wear&quot; in terms of predicates characterized elsewhere in the knowledge base and to be able to infer some consequences of wear. For something to wear, we decided, is for it to lose imperceptible bits of material from its surface due to abrasive action over time. One goal, which we have not yet achieved, is to be able to prove as a theorem that, since the shape of a part of a mechanical device is often functional and since loss of material can result in a change of shape, wear of a part of a device can cause the failure of the device as a whole. In addition, as we have proceeded, we have characterized a number of words found in a set of target texts, as it has become possible.</Paragraph> <Paragraph position="15"> We are encoding the knowledge as axioms in what is for the most part a first-order logic, described by Hobbs (1985a), although quantification over predicates is sometimes convenient. In the formalism there is a nominalization operator ..... for reifying events and conditions, as expressed in the following axiom schema:</Paragraph> <Paragraph position="17"> That is, p is true of x if and only if there is a condition e ofp's being true ofx and e exists in the real world.</Paragraph> <Section position="1" start_page="0" end_page="0" type="sub_section"> <SectionTitle> 242 Computational Linguistics Volume 13, Numbers 3-4, July-December 1987 </SectionTitle> <Paragraph position="0"> Jerry R. Hobbs et al. Commonsense Metaphysics and Lexicai Semantics In our implementation so far, we have been proving simple theorems from our axioms using the CG5 theorem-prover developed by Mark Stickel (1982), and we are now beginning to use the knowledge base in text processing.</Paragraph> </Section> </Section> class="xml-element"></Paper>