Towards a type-theoretical account of lexical semantics

Journal of Logic, Language and Information 19 (2):229-245 (2010)
  Copy   BIBTEX

Abstract

After a quick overview of the field of study known as “Lexical Semantics”, where we advocate the need of accessing additional information besides syntax and Montague-style semantics at the lexical level in order to complete the full analysis of an utterance, we summarize the current formulations of a well-known theory of that field. We then propose and justify our own model of the Generative Lexicon Theory, based upon a variation of classical compositional semantics, and outline its formalization. Additionally, we discuss the theoretical place of informational, knowledge-related data supposed to exist within the lexicon as well as within discourse and other linguistic constructs. The formalization of the structure of natural language utterances around a surface form (phenogrammatics), a deep structure (tectogrammatics) and the meaning thereof as a logical form (semantics) has developed from the original theories of Curry and Montague to form coherent, type-driven models. Most of these new theories rely upon variations of the compositional analysis of the sentence: from pheno to tectogrammatics, and then to semantics. Our contribution to this work aims at giving such a model a means to overcome the problems posed by polysemous lexical units during the semantical analysis of the tectogrammatical form. Building upon an assumed “deep structure”, we formalize parts of Pustejovsky’s Generative Lexicon Theory, linguistically motivated in Pustejovsky (The generative lexicon, MIT Press, Cambridge, MA, 1995), in a pre-processing of the semantics of the sentence. The mechanisms of Lexical Semantics we propose are an additional layer of classical Montague compositional semantics, and, as such, integrate smoothly within such an analysis; we proceed by converting the lexical data to modifiers of the logical form. This treatment of Lexical Semantics furthermore induces us to think that some sort of non-evident background knowledge of the common use of words is necessary to perform a correct semantic analysis of an utterance. This “commonsense metaphysics” would therefore not be strictly confined to pragmatics, as is often assumed.

Other Versions

No versions found

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 100,448

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Analytics

Added to PP
2009-12-28

Downloads
144 (#154,860)

6 months
1 (#1,884,392)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Christian Retoré
Université de Montpellier

Citations of this work

Natural Language Inference in Coq.Stergios Chatzikyriakidis & Zhaohui Luo - 2014 - Journal of Logic, Language and Information 23 (4):441-480.
Natural Language Semantics and Computability.Richard Moot & Christian Retoré - 2019 - Journal of Logic, Language and Information 28 (2):287-307.

Add more citations