Next Article in Journal
Image Inpainting for 3D Reconstruction Based on the Known Region Boundaries
Next Article in Special Issue
Preimage Problem Inspired by the F-Transform
Previous Article in Journal
Lie Symmetry Analysis and Conservation Laws of the Axially Loaded Euler Beam
Previous Article in Special Issue
A Fuzzy Grammar for Evaluating Universality and Complexity in Natural Language
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Describing Linguistic Vagueness of Evaluative Expressions Using Fuzzy Natural Logic and Linguistic Constraints

by
Adrià Torrens-Urrutia
1,*,†,
Vilém Novák
2,† and
María Dolores Jiménez-López
1,†
1
Research Group on Mathematical Linguistics (GRLMC), Universitat Rovira i Virgili, 43002 Tarragona, Spain
2
Institute for Research and Applications of Fuzzy Modeling (IRAFM), University of Ostrava, 701 03 Ostrava, Czech Republic
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2022, 10(15), 2760; https://doi.org/10.3390/math10152760
Submission received: 30 June 2022 / Revised: 20 July 2022 / Accepted: 29 July 2022 / Published: 3 August 2022
(This article belongs to the Special Issue Fuzzy Natural Logic in IFSA-EUSFLAT 2021)

Abstract

:
In recent years, the study of evaluative linguistic expressions has crossed the field of theoretical linguistics and has aroused interest in very different research areas such as artificial intelligence, psychology or cognitive linguistics. The interest in this type of expressions may be due to its relevance in applications such as opinion mining or sentiment analysis. This paper brings together Fuzzy Natural Logic and Fuzzy Property Grammars to approach evaluative expressions. Our contribution includes the marriage of mathematical and linguistic methods for providing a formalism to deal with the linguistic vagueness of evaluative expressions by describing the syntax and semantics of these structures. We contribute to the study of evaluative linguistic expressions by proposing a formal characterization of them where the concepts of semantic prime, borderline evaluative expressions and markedness are defined and where the relation between the semantic constraints of evaluations and their sentiment can be established. A proof-of-concept of how to create a lexicon of evaluative expressions for future computational applications is presented. The results demonstrate that linguistic evaluative expressions are gradient, have sentiment, and that the evaluations work as a relation of hypernym and hyponym, the hypernym being a semantic prime. Our findings provide the basis for building an ontology of evaluative expressions for future computational applications.

1. Introduction

This article focuses on evaluative linguistic expressions, i.e., expressions such as small, medium, large, very deep, rather shallow, beautiful, etc. Our main objective is to provide a formal characterization of this type of expressions by defining them in syntactic terms (determining their canonical and borderline structures) and by describing them semantically in order to account for their vagueness and characterizing their sentiment polarity. In this work, we propose a model with a single scale that allows grading and computing the semantic orientation of an evaluative expression by means of a fuzzy relationship between the semantics of these expressions and the concepts of ‘positive–negative’. With this model, we can calculate the semantic orientation of words.
Fuzzy Natural Logic (FNL) and its theory of evaluative linguistic expressions is a well-known mathematical theory [1,2,3,4,5,6,7]. This theory is based on Lakoff’s hypothesis of universal meaning [8]. Lakoff argues that natural logic is a collection of terms and rules that come with natural language and that allow us to reason and argue in it. FNL tries to account for the basic principle by which people understand each other. This principle is universal and makes communication possible between people of different origins and cultures. An example of this universal representation of meaning is found when any speaker of any place and language and in any situation is able to understand an expression such as “that person is far away”. Regardless of whether the meaning of "far" is “800 km” or “5 km”, we understand that “far” means “a high value on a scale of distance”, with regards to a specific context. A component of FNL is also the theory of intermediate quantifiers [9,10,11] that are expressions such as most, many, a lot of, a few, several, etc.
FNL has been successfully applied in time series analysis [12,13,14], or data mining [15]. We claim that it can be useful for a richer extraction of semantic meaning and for more accurate sentiment analysis. The main advantage of our formalism is that it provides a unified formal analysis for the semantics of all kinds of evaluative expressions.
However, Fuzzy Natural Logic is a hardly known theory in linguistics and is barely used for linguistic semantic analysis. Most of the recent works in linguistics that involve the study of the semantics of the evaluative language is found in the area of sentiment analysis [16].
Some of the works in linguistics integrating FNL as a fundamental part of their framework can be found in the modelling of a Fuzzy Property Grammar [17,18,19,20]. FNL is essential to define the basis of a fuzzy grammar with linguistic features in such work. Furthermore, FNL describes the grammar on a higher (abstract) level. In contrast, linguistic constraints describe the grammar on a local level (at the level of the sentence). The notion of linguistic constraint stands for a relation that puts together two or more linguistic elements, allowing Fuzzy Property Grammar to model vague, highly controversial phenomena for linguists at both levels. Some of the phenomena that a Fuzzy Property Grammar models using the theory of evaluative linguistic expressions are linguistic gradience in natural language, grammaticality, linguistic complexity, and linguistic universality.
This paper provides modifications and suggestions to expand the theory of evaluative expressions towards a linguistic theory as well as future Natural Language Processing (NLP) applications. We have implemented in FNL the notions of theoretical linguistics, linguistic constraints, and characterization of borderline semantic linguistic phenomena in natural language.
Our proposal can contribute to the lexicon-based models in sentiment analysis. The issue of extracting sentiment automatically has been approached in two different ways [21,22]: by using Machine learning techniques or by considering Lexicon-based approaches. The advantage of the machine learning approach is that with a labeled data set training, a classifier can be easily built with existing classifier tools [23]. The main drawback of these techniques is that, being trained with very specific data, it is difficult to apply the model to new contexts. In contrast, lexical-based methods are robust across domains without changing dictionaries. Furthermore, applying dictionaries to a new language is a less demanding task than labeling data in a new language for a classifier [23].
Lexicon-based techniques promote a good synergy between the computational and linguistic approach since they use the linguistic information contained in the text [23]. Moreover, these approaches have been highly regarded in the recent studies.
Some techniques based on the lexicon start from the following two assumptions [23]: (1) words have a “prior polarity”, having a semantic orientation independent of the context; and (2) it is possible to express the semantic orientation with a numerical value. Another important issue to consider is the fact that the semantic orientation of words is affected by the linguistic context; for this reason, it is essential to take into account contextual valence shifters.
Lexicon-based techniques provide for each word classifications of its semantic orientation. To assign each word the value of ‘positive’ or ‘negative’, these classifications are based on Distribution Semantics (DSm) methods. These classifications are either discrete or gradual.
As shown in Table 1, both [24]’s SentiWordNet and [25] Macquire’s Dictionary show only two classifications related to semantic sentiment: positive and negative. On the contrary, the Subjectivity Dictionary proposed by [26,27] considers in addition to the values ‘positive’ and ‘negative’ two additional elements for each of them: strong and weak. Despite a certain graduality shown by the dictionary of subjectivity, this approach still fails to capture the vague and gradient spectrum between ‘positive’ and ‘negative’. In contrast, Ref. [23] offers a richer and more formal model in its proposed Semantic Orientation Calculator (SO-CAL) that features a graded rating on a 10-point scale, from −5 to +5. Unlike previous models, Ref. [23] makes an effort to capture both the vague and gradient nature of semantically oriented word meaning and provides a numerical value for each evaluative item that allows for more detailed extraction of sentiment and of meaning.
Although the models presented in Table 1 generally offer good solutions for calculating the semantic orientation of words, in this paper, we want to propose a different model that uses a single scale for grading and computing the semantic orientation of words and its sentiment together with the description of their structure in a sentence with a fuzzy grammar, i.e., describing prototypical and borderline structures in which such words appear.
The remainder of this paper is organized as follows. Section 2 introduces the basics of Fuzzy Natural Logic. Section 3 presents Fuzzy Property Grammars as a constraint-based grammar that deals with linguistic vagueness. In Section 4, our proposal of a formal characterization of evaluative expressions is presented by defining the concept and by providing a syntactic and semantic explanation of evaluative expression. Section 5 describes a proof of concept to identify trichotomic expressions in Spanish and English from SO-CAL corpus. Section 6 presents some conclusions and future research lines.

2. Fuzzy Natural Logic: Formal Prerequisites

This section summarizes the basic concepts of Fuzzy Natural Logic. It is underpinned on six key concepts:
  • Fuzzy set. A fuzzy set A in a universe U, is a function A : U L where L = [ 0 , 1 ] is a support of the standard Łukasiewicz MV-algebra [ 0 , 1 ] , , , , , 0 , 1 . The , are lattice operations (here equal to min and max), , are Łukasiewicz conjunction and implication, respectively defined by
    a b = max 0 , a + b 1 , a b = min 1 , 1 a + b
    where a , b [ 0 , 1 ] . The set of all fuzzy sets on U is denoted by F ( U ) . If A is a fuzzy set then its kernel is a set K e r ( A ) = { u A ( u ) = 1 } .
  • Lakoff’s hypothesis of universal meaning. According to Lakoff [8], natural logic is a collection of terms and rules that comes with natural language and allows people to reason and argue in it. This idea captures the basic principle of why people understand each other. This principle is universal in all languages and makes communication among people possible.
  • Evaluative linguistic expressions are those expressions of natural language used by people to characterize features of objects or their parts [1,5,12] such as length, beauty, size, complexity, economical value, kindness, among others. The most elaborated evaluative expressions in FNL simple evaluative ones. They have the general form
    linguistic hedge TE - adjective
    TE - adjective can be grouped to form a fundamental evaluative trichotomy consisting of two antonyms and a middle term, for example g o o d , n o r m a l , b a d , b e a u t i f u l , n o r m a l , u g l y , e a s y , a v e r a g e , c o m p l e x , etc. The triple of adjectives s m a l l , m e d i u m , b i g is taken as canonical. On the other hand, linguistic hedge makes more or less specific the meaning of the TE - adjective . Usually it is represented by an intensifying adverb such as “very”, “roughly”, “approximately”, “significantly”, etc. This is a special expression representing a linguistic phenomenon called hedging. In some papers of FNL, linguistic hedge is found as linguistic modifier .
  • Possible world. It is a specific context in which a linguistic expression is used. In the case of evaluative expressions, it is characterized by a triple w = v L , v S , v R . Without loss of generality, it can be defined by three real numbers v L , v S , v R R where v L < v S < v R . These numbers represent an interval of reals [ v L , v S ] [ v S , v R ] where v S is marked to emphasize position of “typically medium”.
  • Intension. The intension of a linguistic expression is a property denoted by a specific word or expression. It is independent on a concrete possible world (context) and does not change when the context is changed.
  • Extension. The extension of a linguistic expression is the referent. It depends on the particular context and changes when the possible world changes. With the exception of few specific cases, the elements falling into an extension are delineated vaguely. In our theory, they are accompanied by degrees. For example, the expression “very long distance” has always a high degree in any context of distance (possible world). Thus, “very long distance in Europe” is “1000 km” when “driving a car”, while it is “5 km” when “walking on foot”. In such case, extensions are fuzzy sets in the intervals [ 0 , 1000 ] and [ 0 , 5 ] .
Definition 1.
Let A be an expression of natural language. Then, its intension Int( A ) is a function
I n t ( A ) : W F ( V )
where W is a set of possible worlds, V is a set of some objects and F is a set of all fuzzy sets defined on V. The extension of A in a given possible world (context) w W is a fuzzy set E x t w ( A ) = I n t ( A ) ( w ) F ( V ) .
In Figure 1 (from [6]), extensions of the expressions forming the fundamental evaluative trichotomy are depicted. Part (b) of the figure contains informal justification, how the shapes of extensions of evaluative expressions are constructed. Namely, in the given context v L , v S , v R , “small” values are those around the left bound v L , while, e.g., “very small” or “extremely small” are part of the small ones. Similarly for “big”, that are values around the right bound v R . “Medium” values are values around the central point v S (note that it can lay on an arbitrary place between v L and v R ).

3. Fuzzy Property Grammar

Fuzzy Property Grammars combine the formalism of Fuzzy Natural Logic and linguistic constraints typically used in linguistics. The higher-order fuzzy logic as a formalism describes the grammar at a higher level (abstractly), enabling a mathematical formalization of the degrees of grammaticality. In comparison, linguistic constraints of a Property Grammar [28] allow us to describe vague phenomena on a local-sentence level. However, unlike a standard Property Grammar, FPGr characterizes the objects (constraints) as prototypical and borderline. Therefore, both assets are necessary to build a Fuzzy Property Grammar. In this section, the basic concepts of Fuzzy Property Grammars are presented.
A linguistic constraint is a relation that puts together two or more linguistic elements such as linguistic categories, or parts-of-speech. Formally, a linguistic constraint is an n-tuple A 1 , , A n where A i are linguistic categories. We usually have n = 2 . We consider the following linguistic categories: D E T (determiner), A D J (adjective), N O U N (noun), P R O P N (proper noun), V E R B (verb), A D V (adverb), C O N J (conjunction), S C O N J (subordinate conjunction) and A D P (preposition).
To identify linguistic categories, it is necessary to either use a part-of-speech tagger or to identify them according to the constraints of a specific language. For example, Spanish and English would have their constraints to identify a N O U N and an A D J , which will not necessarily be shared with Chinese or Japanese. It could happen that natural languages would require an extension or a shortening of these part-of-speech. Because FPGr is closely related to the framework of Universal Dependencies (UD) [29,30], we recommend using the UD parser to identify each part of speech. UD recognises a maximum of 17 elements in a part-of-speech parsing. However, when working with FPGr in the syntactic domain, the model proposes to shorten from 17 part-of-speech to 9 (at maximum) to simplify the model and the analysis of the inputs from UD. The fact of how FPGr would identify part-of-speech with constraints UNIVERSALLY is not explored yet, since it is such a new model and mainly applied for the syntactic domain, mainly in Spanish and English. A further explanation of this can be found in Torrens–Urrutia in [19] (Chapter 7). One of the options to describe linguistic categories with constraints universally is to implement formal definitions of General Natural Syntax (GNS) from Manca and Jiménez-López [31] into FPGr.
The constraints provide a non-derivational definition of several linguistics. Linguistic constraints indicate the properties that an object (or objects) must satisfy. Thus, linguistic constraints define both the linguistic knowledge of a speaker, and the membership of a linguistic input towards a specific grammar in terms of constraint satisfaction. There are four types of constraints in Fuzzy Property Grammars:
  • General or universal constraints that are valid for a universal grammar (any language).
  • Specific constraints that are applicable to a specific grammar.
  • Prototypical constraints that definitely belong to a specific grammar, i.e., their degree of membership is 1.
  • Borderline constraints that belong to a specific language with some degree only (we usually measure it by a number from [0,1]).
The constraints that FPGr work with are the following (the A and B are understood as linguistic categories):
-
Linearity of precedence order between two elements: A precedes B, in symbols A B . Therefore, a violation is triggered when B precedes A. Example: “The ( D E T ) professor ( N O U N )”, C α ( D E T N O U N ). C α stands for satisfied constraint.
-
Co-occurrence between two elements: A requires B, in symbols A B . A violation is triggered if A occurs, but B does not. Example: “The ( D E T ) girl ( N O U N ) plays football C α ( N O U N D E T ), but “girl ( N O U N ) plays football C β ( N O U N D E T ). C β stands for violated constraint.
-
Exclusion between two elements: A and B never appear in co-occurrence in the specified construction, in symbols A B . That is, only A or only B occurs. Example: “He ( P R O N ) does yoga”, C α ( P R O N N O U N ), but “He ( P R O N ) boy ( N O U N ) does yoga C β ( P R O N N O U N ).
-
Uniqueness means that neither a category nor a group of categories (constituents) can appear more than once in a given construction. For example, in a construction X, U n i q = { a , b , c , d } . A violation is triggered if one of these constituents is repeated in a construction. Example: “The ( D E T ) the ( D E T ) kid that who used to be my friend C β (in nominal construction: U n i q = { D e t , R e l } ).
-
Dependency. An element A has a dependency i on an element B, in symbols A i B . Typical dependencies (but not exclusively) for i are s u b j (subject), m o d (modifier), o b j (object), s p e c (specifier), v e r b (verb), c o n j (conjunction). A violation is triggered if the specified dependency does not occur. Example: “Andorra is a small ( A D J ) country ( N O U N )”, C α ( A D J m o d N O U N ), but “Andorra is a small ( A D J ) goodly ( A D V )”, C β ( A D J m o d N O U N ).
-
Obligation. This property determines which elements are heads. It is expressed by the symbol □. This property tends to be avoided since FPGr pursues to define the relationship of part-of-speech without hierarchy or derivation, only at a local level. However, it is useful to define specific constructions to express the mandatory appearance of a category or linguistic feature.
Additionally, two key concepts for the formalization of Fuzzy Property Grammars are:
  • Linguistic Feature. Features specifies when properties are going to be applied to a category. The typical feature to be represented is a linguistic function, such as the function of subject: A [ s u b j ] . Features are always written and subindexed in a linguistic category, i.e., N O U N [ s u b j ] .
    Example 1.
    A property for an English grammar such as N V might be inaccurate since the noun can both precede and be preceded by a verb. We can specify functions and other values for a category to provide proper linguistic information thanks to the features. Therefore, property grammars can specify that a noun as a subject precedes a verb: N [ s u b j ] V . Features reinforce properties as a tool that can describe linguistic information independently of a context and more precisely represent grammatical knowledge by taking into account linguistic variation. However, features can express any type of specification from any domain. For this paper, we are interested in semantic features, which will be expressed like the following A [ s e m : x ] . Such characterization is useful for words that even being defined with the same category trigger different meanings, i.e., I have a snake, (snake N O U N [ s e m : a n i m a l ] ), John is a snake, (snake N O U N [ s e m : t r e a c h e r o u s ] ).
  • xCategory. An x C a t e g o r y is a feature which specifies that a certain category is displaying a syntactic fit from another category, for example, an adjective with a syntactic fit of a noun. All the x C a t e g o r i e s are marked with a x before a prototypical category, i.e., x A D J . For example, in Spanish, el boxeador ( N O U N ) león ( N O U N ) (the lion boxer), león ( N O U N ) is performing as an adjective; therefore, el boxeador ( N O U N ) león ( N O U N [ x A D J ] ).
Definition 2.
A Fuzzy Property Grammar ( F P G r ) is a couple
F P G r = U , F G r
where U is a universe
U = P h ρ × M r μ × X χ × S δ × L θ × P r ζ × P s κ .
The subscripts ρ , , κ denote types and the sets in (4) are sets of the following constraints:
  • P h ρ = { p h ρ p h ρ is a phonological constraint} is the set of constraints that can be determined in phonology.
  • M r μ = { m r μ m r μ is a morphological constraint } is the set of constraints that can be determined in morphology.
  • X χ = { x χ x χ is a syntactic constraint } is the set of constraints that characterize syntax.
  • S δ = { s δ s δ is a semantic constraint } is the set of constraints that characterize semantic phenomena.
  • L θ = { l θ l θ is a lexical constraint } is the set of constraints that occur on lexical level.
  • P r ζ = { p r ζ p r ζ is a pragmatic constraint } is the set of constraints that characterize pragmatics.
  • P s κ = { p s κ p s κ is a prosodic constraint } is the set of constraints that can be determined in prosody.
The second component is a function
F G r : U [ 0 , 1 ]
which can be obtained as a composition of functions F ρ : P h ρ [ 0 , 1 ] , …, F κ : P s κ [ 0 , 1 ] . Each of the latter functions characterizes the degree in which the corresponding element x belongs to each of the above linguistic domains (with regards to a specific grammar).
Technically speaking, F P G r in (5) is a fuzzy set with the membership function computed as follows:
F G r ( x ρ , x μ , , x κ ) = min { F ρ ( x ρ ) , F μ , , F κ ( x κ ) }
where x ρ , x μ , , x κ U .
Let us now consider a set of constraints from an external linguistic input D = { d d is a dialect constraint } . Each d D can be observed as an n-tuple d = d ρ , d μ , , d κ . Then, the membership degree F G r ( d ) [ 0 , 1 ] is a degree of grammaticality of the given utterance that can be said in arbitrary dialect (of the given grammar).
FPGr operates taking into account the notion of linguistic construction, originally from [32,33]. A linguistic construction is understood as a pair of structure and meaning.
In FPGr, linguistic constructions in written language stands for a simplified version of a F P G r because only three linguistic domains are relevant for it: the morphological domain ( m r ), the syntactical domain (x), the semantic domain (s), < m r , x , s > , the others are neglected: F P G r = < U ¯ , F G r ¯ > .
Definition 3.
A construction is:
U ¯ = < m r , x , s >
Examples of constraints in the linguistic domains in Equation (7):
  • The morphological domain < m r > , which defines the part-of-speech and the constraints between lexemes and morphemes. For example, in English, the lexeme of an adjective ≺ (precedes) the morpheme -ly, and the morpheme -ly⇒ (requires) an adjective as a lexeme.
  • The syntactical domain < x > , which defines the structure relations between categories in a linguistic construction or phrase. For example, in English, an adjective as a modifier A D J [ m o d ] of a noun N O U N is dependent ( ) of such a noun ( A D J [ m o d ] N O U N ).
  • The semantic domain < s > , which defines the network-of-meanings of a language and its relation with the syntactical domain. This can be defined with semantic frames [34,35]. It is also responsible for explaining semantic phenomena as metaphorical meaning, metonymy, and semantic implausibility. For example, in English, d r i n k a b l e object (i.e., w a t e r [ s e m : d r i n k a b l e ] ) ⇒ (requires) r e c i p i e n t (i.e., g l a s s [ s e m : r e c i p i e n t ] ). A metonymy can be triggered with the follow I F T H E N rule: If asking for something d r i n k a b l e without r e c i p i e n t (i.e., “may I have some water, please?”), then r e c i p e n t is included as a feature in the frame of d r i n k a b l e object, i.e., w a t e r [ s e m : d r i n k a b l e , r e c i p i e n t ] .
All the linguistic descriptions from now on are conducted following the framework of Fuzzy Property Grammars (FPGr). For a deeper insight of the theory, see [18,19,20].

4. Evaluative Linguistic Expressions with Linguistic Features

4.1. Linguistic Prerrequisites

Fuzzy Natural Logic must include the following linguistic concepts in the theory of evaluative expressions. Those concepts are essential to characterize prototypical and borderline evaluative words.
  • Referent. A referent is a person or thing to which a linguistic expression refers. For example, in the sentence Paul talks to me, the referent of the word Paul is the particular person called Paul who is being spoken of, while the referent of the word me is the person uttering the sentence. In the case of the evaluative linguistic expressions, we will understand the referent as the entity which is being evaluated.
    The referent of the evaluation can be: (1) in the linguistic input, (2) close to the evaluative word, or (3) it can be absent in the linguistic input but present in the extra-linguistic context.
    Usually, in English and Spanish language, the referent will typically be the subject of the evaluative sentence ( X [ s u b j ] ) if the referent and the evaluation are in a dependency relationship with a V E R B . Otherwise, the syntactic head of the evaluative word will establish a modifier dependency with it. This syntactic head will usually be a N O U N from the evaluation. For example, in Paul runs fast, the referent of the evaluation runs fast is Paul, not the V E R B runs. Therefore, the fact that the evaluative word has a dependency of modifying a V E R B , A D J m o d V E R B , does not imply that it will be the referent, even though V E R B is a higher element in the syntactic hierarchy. Sometimes, a referent can be co-referential. That is, the same referent appears twice in a sentence.
    For example, Paul had his huge backpack with him. In such case, Paul and him refers to the same person. But, following our definition, the referent of the evaluation is not Paul, but the backpack since it is the one receiving the evaluation.
    When the referent is in the linguistic input, the evaluative word and the referent will have a relationship of dependence. Such as in Mark watches a bad show on TV. In this case, the referent is a show, and the evaluative word is bad. On the other hand, typically, when the evaluation and the referent are connected with a dependency through a verb, the referent will usually be the subject. For example, Cristina thinks that Claudia is smart enough to pass her degree. The evaluation is is smart enough, and the referent is Claudia. Both constructions have a relationship of dependence, and they are connected because of the verb. However, further exploring to set a frequent pattern of dependency between referent and evaluation is needed.
  • Linguistic pairing stands for the task or phenomena in which linguistic features associate two linguistic elements. These can be syntactic or semantic features or both. The most interesting linguistic pairing for adding linguistic features in the evaluative linguistic expressions is the one which pairs the referent with the evaluative word.
  • Markedness arises to represent the importance of the linguistic context (not extra-linguistic) for a word. A sentence α is more marked than a sentence β if α is acceptable in less contexts than β . It can be determined either by the judgments of the speakers or by extracting the number of possible context types for a sentence [36]. Therefore, a canonical constraint or property ( C α ) is never context dependent. For example, C α ( D E T N O U N ). A borderline constraint or property ( C γ ) is context-dependent if the degree of unacceptability triggered by its violation varies from one context to another. For example, in Spanish, canonically, determiners do not precede proper nouns, that is C α ( P R O P N D E T ). However, such constraint can be violated when what the structure { D E T , P R O P N } is triggering the meaning of a football team, “El ( D E T ) Barcelona ( P R O P N ) era mejor con Messi que ahora” (“The Barcelona was better with Messi than now”). Such a feature is important to characterize and clarify both the referent and the evaluation. Following the last sentence, the evaluation of was better with Messi than now would have a biased referent if the grammar would understand that the Barcelona meant the city of Barcelona, and not the football club.
    In this case, the violation of the property, accepted only in few contexts, displays and characterizes a marked structure ( D E T P R O P N [ s e m : c i t i e s ] ) ( s e m : means semantics), with a marked meaning ( D E T P R O P N [ s e m : f o o t b a l l t e a m ] ), expressed as: I F F C β ( D E T P R O P N ) ; A N D P R O P N [ c i t y ] T H E N C γ ( D E T P R O P N [ s e m : f o o t b a l l t e a m ] ) .
    Additionally, in semantics, structures are marked when they do not allow interchangeability of elements, and yet keep the same meaning. A non-marked structure would be: “Mary’s house, the nice one, is in Elm Street”. I change the elements and yet I can get almost same meaning “My sister’s home, the one when you were last time, is next to mine” (knowing that Mary is my sister, the nice one is the one when you were last time and next to mine is Elm Street)’. However, this cannot happen with idioms, or other marked structures such as “Mark is a pain in the ass” by Mark is a pain in the finger/eye/foot/chest/etc., or “it is raining cats-and-dogs” by it is raining elephants-and-giraffes, etc. Therefore, idioms tend to have marked meanings. Another example: “No shit, Sherlock”, meaning ‘judgment over a statement that was obvious’. It is implausible to exchange such construction for other elements such as: No turd, detective, or No smoke, Einstein. Moreover, FNL should not purse to characterize extra-linguistic extreme-implied (non-)evaluations that demand extreme knowledge of the context, such as the one when you were last time understanding that such a construction is a very polite way of saying I am referring to the nice one, not the ugly one.
  • Semantic Coercion stands for implausibility of meaning. Coercion is a typical effect in defining and identifying metaphors, which non-literal meanings are triggered due to the implausibility of such meaning. To summarize, it is the similar effect of markedness but focused on the semantic domain.
  • Compositionality is typically understood as that the summarization of meanings brings us the final meaning, i.e., “it a strong rain”, meaning: [it + is + a + strong + rain]. However, compositionality does not always work taking into account the previous example it is raining cats-and-dogs. It does not actually mean that animals are falling from the sky!. Therefore, the evaluative sentence it is raining cats-and-dogs does not fall on the logic of compositionality.
  • Lexical unit, in FPGr, stands for those structures that work as a single part-of-speech, that is as a single unit, and, therefore, they only count as a one category. For example, in the chief ( N O U N ) executive ( N O U N ) office ( N O U N ) of Amazon is Jeff Bezos; FPGr would parse it as the [chief executive office ( N O U N )] of Amazon is Jeff Bezos, and not as the [chief ( N O U N ) executive ( N O U N ) office ( N O U N )] of Amazon is Jeff Bezos. Therefore, the chief executive office is working as a single N O U N . In fact, that is why the tendency is to reduce it to CEO, the CEO of Amazon is Jeff Bezos. The same applies to evaluations such as it is raining cats-and-dogs, which stands for it is raining really heavily or it is pelting down, or it is bucketing down, which stands for it is raining very, very heavily. Therefore, the notion of the lexical unit serves both the evaluation and its referent.

4.2. The Concept of Evaluative Expressions

An evaluation is a concept characterizing particularities of a specific object (object understood in a broad sense, as people are objects too). That is the characterization of a referent.
Examples of linguistic evaluations are the following ones:
(1)
I love you very much.
-
Express affection. The verb love is the evaluative word. The referent is ‘you’.
(2)
Plane food is disgusting.
-
Express a judgment. The evaluative word is ‘disgusting’. The referent is ‘plane food’. The referent operates as either a lexicalized N O U N (‘plane food’ is N O U N ), or as a N O U N with an adjective fit which modifies the other N O U N : ‘plane’ ( N O U N [ x A D J ] m o d ) ‘food’ ( N O U N [ s u b j ] ).
(3)
John is not a bad person, but he can be a snake.
-
Express a judgment. ‘Snake’ is fitting a metaphorical meaning, ’treacherous’, and not the animal. The referents are John, and he.
(4)
John looked like this (showing at a tomato) when Jennifer asked him for a date.
-
Express a judgment. ‘this’ is a deictic work. Thus, it is empty of meaning by itself, and the context is needed to grasp the meaning that is intended to be communicated. The context reveals that it probably means John was shy. Probably is stressed since the meaning is not known for sure. Only an implication can be made by linking the linguistic input, its linguistic context, world knowledge, and communicative context.
If we accept that all the above sentences express concepts that are a matter of degrees, then we have a fuzzy set: A : M [ 0 , 1 ] . Every element can be associate with types: M δ stands for ‘John’, M ϵ stands for the concept of ‘being bad’, M θ stands for the sentiment of the concept (positive or negative), and M o express a membership degree, creating a complex set of functions: A : M ϵ × M δ × M θ M o .
Therefore, an evaluative linguistic expression is an evaluative concept expressed through a language system characterizing a referent. The key elements are the evaluator and the object being evaluated.
Additionally, the linguistic evaluative expressions can be understood under the notion of linguistic construction [32,33], which stands that a linguistic construction is a pair of structure and meaning. Therefore, defining both linguistic syntactic and semantic features of an evaluative construction is necessary. An evaluation construction is a linguistic input evaluating a referent. This referent can be found in the linguistic input itself or can be extra-linguistic (found in the communicative context or the common knowledge).

4.3. Semantics of Evaluative Expressions

To express the generation of an evaluative linguistic expression by being the evaluator at the centre of the concept, we have provided a semantic hierarchy of the evaluation represented in Figure 2.
Figure 2 expresses the hierarchy of the different levels in which the semantics of an evaluative concept are found:
  • On the first level, closer to the thought, we will find a semantic prime, always understanding semantic primes as those most abstract concepts which will define a fuzzy set. This idea is inspired in [37]. However, this author considers as semantic primes only the concepts of good-bad. While in this work, we consider primes as all those triplets that create a fundamental trichotomous expression such as [ugly-medium-beautiful], [small- medium-big], [easy-medium-complex], etc. Any prime has to be a prototypical meaning, that is, non-context dependent.
  • On the second level, we would find those evaluations that are sensitive to the linguistic context. Those words that are given a specific structure trigger a non-prototypical meaning. That is the case of Example 3), in which the syntactic relation between the referent he, meaning a person, the verb ‘to be’, and the word snake, trigger a borderline meaning on the word snake due to the implausibility that a person would be an actual animal. This effect of implausibility is called semantic coercion [38,39,40,41]. However, it is clear that meaning is not a matter of probabilities in both the semantic prime, and in the linguistic context stage. The semantics of linguistic inputs related with those stages can be characterized using logic and linguistic features.
  • On the third level, and further from the semantic prime, we would find those evaluations that need extra-linguistic context. Such concepts can only be understood when the communicative context is known or expresses a common knowledge shared by the interlocutors. Example 4) is a clear example of extra-linguistic meaning. Therefore, it is impossible to extract its meaning from the linguistic input’s information to us. These cases are acknowledged as being extremely borderline. We will not formalize this type of evaluative expression here.
Summing up, evaluative expressions are linguistic inputs that express a concept of an evaluation that is a gradient and has a sentiment. The semantic hierarchy gives structure to identify the abstract types of the meaning of an evaluative concept. The meaning of an evaluation can be closer to a prototypical evaluation (a semantic prime), to a borderline evaluation (needing of a characterization of the linguistic context), or closer to an extremely borderline evaluation (needing of the information of the communicative context).

4.4. Syntax of Evaluative Expressions

Recall from (9) that the general form of an evaluative linguistic expression is
linguistic hedge TE - adjective
This form poses two problems in linguistics: the confusion of the term “hedge” and the fact that a TE-adjective is not necessarily an adjective.
The term “linguistic hedge” is used in pragmatics for what is called hedging defined as people use hedging when they try to avoid criticism [42,43,44]. Some other definitions can also be found in [45]. Hassan and Said [45] define hedges as those elements “which serve human communication as more flexible, moderate, and convincing”. Hedges convey “intentional vagueness”,“mitigation”, “tentativeness”, “politeness”, “indirectness”, “possibility”, “evasiveness”, “lack of full commitment”, etc. Hedges can be characterized as expressing the writer’s attitude towards both the propositional information and his/her awareness of the readers. Words that are present in hedging strategies are “I think”, “I don’t know, “a little bit”, “in my view”, “quite”. Some of this hedges can be found also as part of an evaluative expression; however, it is important to differentiate the intensifier strategy than the discourse hedging strategy:
  • I think this is quite right, but…
  • Your paint seems a little bit odd.
The discourse function of the pragmatics of language of the terms “quite” and “a little bit” is different from the intensifying semantic function in terms of widening or stretching the degree of vagueness of the TE-adjective “right”, and “odd”. That is why it would be better to define this first part as simply intensifier .
Regarding the use of the concept of TE - adjective , it is necessary to change it to TE - head . Firstly, not necessarily an adjective will be the head of an evaluative expression itself. For example, in English, we will find differences of such things:
(1)
Mark is very intelligent:
intensifier - very TE - head intelligent ( an adjective ) .
(2)
Mark is an Einstein
intensifier - none TE - head Einstein ( a Proper Noun ) .
(3)
I love Mark’s intelligence very much
intensifier very much TE - head love ( a verb ) .
Example (1) is what we would consider a canonical evaluative expression if we consider that the prototypical evaluative heads in the English language are adjectives. That is because using an adjective as an evaluative head is very productive in English.
Example (2) has a proper noun as an evaluative head, which could be considered a borderline form of the evaluative term “intelligence” since, in English, proper nouns do not usually bear the quality of being evaluative. Additionally, we can satisfactorily apply a constituent replacement test of the word ”Einstein” for ”intelligent”. The test is a success since we keep the same meaning, and the constituents work similarly. Such cases will lead us to distinguish between prototypical and borderline evaluative heads. It must be kept in mind that each language will decide which linguistic categories are prototypical productive or borderline for bearing an evaluative semantic meaning. Therefore, sentences (1) and (2) are synonyms but display different evaluative heads.
Example (3) shows a verb being used as an evaluative head. This case helps us to stress why the formula of an evaluative expression in its general form should bear the notion of TE - head and not TE - adjective . The evaluative verb “love” is not considered a borderline evaluative head here since the verb love is prototypical an evaluative verb even though verbs do not prototypically bare evaluative meanings.
We establish the following rules to determine if a word is a prototypical evaluative head or a borderline one:
  • If the evaluative head is canonically and productively used as an evaluative, the evaluative head is prototypical.
  • If the evaluative head is not canonically and productively used as an evaluative, but it passes a constituency test by being replaced by a prototypical evaluative head, then the evaluative head is borderline.
From the above follows that the general form of an evaluative expression should be:
intensifier TE - head
Figure 3 represents the formal characterization in FPGr of the basic structure of a linguistic construction of an evaluative expression in terms of linguistic constraints.
  • The elements within the construction can be found between brackets [].
  • The first element is the intensifier, defined as optional, that is why it is between parenthesis ().
    -
    It is also defined with a linguistic dependency of modifier; meaning that the intensifier will always be dependent of the evaluative head as a modifier, no matter the language or grammar.
  • The second element is the evaluative head, which is mandatory. That is why it is represented with the constraint of obligation □.
    -
    Additionally, the evaluative head will bare a dependency on its referent, that is, the element which is being evaluated. Because we want to keep this formalization as the most linguistically universal possible, we do not state the dependency that the evaluative bears. That is why it is represented with [ x d e p ] . In English, the most common evaluative construction is found as the object of a copular verb (“to be”, “feel”, “seem”, “become”, etc.) and as the subject as its reference: Subject-Copular Verb-Evaluative Construction. In such a case, we would say that the evaluative head has a dependency on the subject.
    *
    It is specified that the referent R e f experience a pairing with the evaluative head, which states if they are semantically compatible or if any coercion would be involved [ p a i r i n g ] . A clear case of a coerced pairing can be found in the sentence “Mark is a giant”. The pairing between “Mark” and “giant”is implausible since “Mark” bares the quality of being “+ human” and “giants” does not exist, they are mythological creatures. However, the paring is possible, and a coerced meaning is interpreted. If a pairing is not present in the linguistic context, the referent would be found in the communicative context; that is, it will be extra-linguistic [ e x t r a l i n g ] .
The possibility of having different categories depending on the language makes the notion of the evaluative head provide universality to our model.
If we use an intensifier, we can make the evaluative head more precise. FNL does not imply that adjectives are always a typical evaluative head across languages. For example, Korean and Japanese do not have adjectives in the strict sense, but can use verbs as evaluative units. These kinds of structures are also possible in English where an evaluative expression such as “I love your food” has a verb as its evaluative head. The meaning of this structure would be paraphrased as “ Your food is excellent”. Both sentences would be computed with a high value in the degree of judgment. Therefore, with FNL we can characterize the core/depth structure of the evaluation semantics, extracting the exact core meaning of both sentences, independently of the surface structure of the linguistic construction.

4.5. Evaluative Expressions: A Formal Characterization

A formal characterization with the FPGr of the syntactic and semantic properties that a universal construction of an evaluative expression has to fulfill is presented in Figure 4.
As shown in Figure 4, we consider three elements in the formal characterization of evaluative expressions:
  • The insertion of grammar. i n s e r t g r a m m a r stands for inserting an FPGr or other grammar with constraints (a universe), which will define the specific constraints of a language, as mentioned in Equation (4). This grammar will define those satisfied and violated constraints of a particular language. Therefore, FPGr will be able to characterize the degree of grammaticality of an evaluative expression with regards to a specific grammar.
  • The constraints of the intensifier. The intensifier displays the same constraints as in Figure 3.
    -
    X [ I N T ] [ x m o d ] X [ E H ] means any category X as intensifier X [ I N T ] has a dependency as a type of modifier [ x m o d ] towards any category as an evaluative head X [ E H ] .
  • The constraints of the evaluative head. The evaluative head displays some of the same constraints observed, as in Figure 3, such as:
    -
    X [ E H ] meaning any category X as an evaluative head X [ E H ] is mandatory □, leaving the intensifier as optional.
    -
    X [ E H ] [ x d e p ] R e f [ p a i r i n g e x t r a l i n g ] means any category X as an evaluative head X [ E H ] has a non specified dependency [ x d e p ] towards a referent R e f , which can be found in the linguistic context with linguistic pairing or in the extra linguistic context R e f [ p a i r i n g e x t r a l i n g ] .
    -
    Additionally, the evaluative head displays semantic constraints here that are necessary to complete such a universal characterization of the constructions of an evaluative expression.
    *
    X [ E H ] T E v L v S v R L S V . Meaning that any category X as an evaluative head X [ E H ] requires to be part of one of the sets of a trichotomous expression T E , either the left one v L , the one in the middle v S , or the right one, v R . And (∧), the trichotomous expression, requires to be tagged with a Linguistic Semantic Variable ( L S V ). The trichotomous expression can be incorporated in any constraint-based grammar in terms of hyponym and hypernymy (being the hypernym interpreted as our semantic prime) simply by using v L v S v R as a constraint. Therefore, lexical items as “beautiful”, “difficult”, “pig”, and “hate”, can be defined with such constraints as:
    ·
    beautiful”, A D J [ E H ] T E v R L S V b e a u t y .
    ·
    difficult”, A D J [ E H ] T E v R L S V c o m p l e x .
    ·
    pig”, N O U N [ E H ] T E v L L S V P l e a s i n g L i k a b i l i t y .
    ·
    hate”, V E R B [ E H ] T E v L L S V E s t e e m .
    *
    Moreover, it can be specified whether such lexical items are a prototypical or borderline evaluative expression, always with regards to the Linguistic Semantic Variable (the semantic prime). If the evaluative head is prototypical with regards to the prime, it will be defined with the semantic constraint [ P R O T ( L S V α ) ] . If the evaluative head is borderline with regards to the prime, it will be defined with the semantic constraint of [ B O R D ( L S V γ ) ] . For example:
    ·
    beautiful”, A D J [ E H ] T E v R L S V b e a u t y [ P R O T ( L S V α ) ] .
    ·
    pig”, N O U N [ E H ] T E v R L S V c l e a n l i n e s s [ B O R D ( L S V γ ) ] .
    -
    The last constraint is X [ E H ] S e n t i m e n t T E J u d g m e n t v L v S v R . Meaning that any category X as an evaluative head X [ E H ] requires a sentiment to be assigned S e n t i m e n t . Sentiment is characterized as a trichotomous expression of judgment T E J u d g m e n t v L v S v R , such as n e g a t i v e n e u t r a l p o s i t i v e . As with before, the sentiment can be prototypical or borderline, since it is considered that any evaluation can be potentially used in both ways [16]. For example:
    *
    This song is excellent”, it is interpreted as positive because the word “excellent” is prototypical positive: A D J [ E H ] S e n t i m e n t T E J u d g m e n t v R [ P R O T ( S α ) ]
    *
    This song is excellent for deaf people”, it is used as a borderline negative because of the linguistic context: X [ E H ] S e n t i m e n t T E J u d g m e n t v L [ B O R D ( S γ ) ]
Summarizing, in FNL we define an evaluative expression as a linguistic construction characterized by a lexical unit tagged as Evaluative Heading (EH), which is a subjective evaluation of an object. Evaluative heads have the following properties:
  • They occur in each member of a triplet of expressions –an evaluative trichotomy–, such as small-medium-big, short-medium-tall, beautiful-average-ugly. A trichotomous expression is a characterization of a general semantic feature –a semantic prime– called a linguistic semantic value. Examples of these linguistic semantic values are height, beauty, complexity, intelligence, etc. Technically, we can refer to features of a given object.
  • They have a sentiment value expressed in terms of positive, negative degree. Sentiment is computed by ordering all the trichotomous expressions as protoypically negative v L , protoypically medium v S , protoypically positive v R , and converting those to borderline sentiment when needed.
  • FNL can characterize the interpretability of an evaluation even if it is subjective.

5. Towards a Lexicon of Evaluative Expressions

The formal characterization of evaluative expressions we have presented in this paper constitutes a robust mathematical model that is inapplicable without a lexicon of evaluative expressions. One of the main reasons for this is because linguistic categories, or part-of-speech, are not enough to identify an evaluative expression. It is, therefore, necessary to establish a procedure to find and classify evaluative expressions in trichotomous structures by identifying semantic primes and assigning values on a sentiment scale.
Taking into account the need to have a lexicon of evaluative expressions, we have carried out a first experiment, as a proof of concept, to identify trichotomic expressions in Spanish and English. We have performed a linguistic evaluation for FNL over a sentiment analysis corpus entirely build on evaluative words.
For our proof-of-concept we have chosen SO-CAL, lexicon. https://github.com/sfu-discourse-lab/SO-CAL. SO-CAL has two tagged sentiment corpus, one in Spanish, and another one in English. SO-CAL has three different lexicons for each language: (1) a lexicon for evaluative adjectives, (2) a lexicon for adverbs, and (3) a lexicon for intensifiers. SO-CAL lexicon was built through pairing of DSm tools. The corpus merely tagged each word with a value from −5 to +5 without further detail.
SO-CAL [46] was the best option for building up a lexicon with FNL, since it already has a classification of 11-point scale, which fits with the theoretical extension of FNL (Figure 5).
We have taken the SO-CAL lexicon and re-distributed its adjective-lexical units. We were interested in re-building the lexicon for characterizing evaluative adjectives with FNL providing more linguistic information for each EH, and re-distributing the sentiment polarities.
In order to built a lexicon underpinned in SO-CAL lexicon, two researchers elaborate Linguistic Semantic Variables (LSV) scales. One worked with an English Lexicon, and the other worked with a Spanish Lexicon. Without having contact with each other, they had to find the Prime Evaluative Heads (EHs) for each LSV scale. They have also to tag every adjective into an LSV scale, under a Prime EH.
The English adjective lexicon has 2827 lexical units, and the Spanish Adjective lexicon has 2049.
In Figure 6, the x-axis displays the sentiment polarity for each adjective cluster [−5, 5]. At the same time, the y-axis represents the number of words (in proportion from one corpus to the other) for each polarity set in the lexicon. Therefore, thanks to Figure 6, we confirm that SO-CAL does not demonstrate substantial differences in the classification of the adjectives’ polarity in English and Spanish. Moreover, SO-CAL lexicon demonstrates a preference for medium values. Most of the adjectives are classified in the polarity of [+2, +1], and [−2, −1], leaving the 0 value just with one case, with the word better.
On the other hand, we re-classified the SO-CAL lexicon applying the criteria from FNL. We manually re-built half of the lexicon for both languages. We re-tagged 1419 adjectives for the English language, and 1549 adjectives for the Spanish Language. Therefore, it allowed us to evaluate FNL cross-linguistically.
After manually tagging each adjective under an LSV, and under a prime EH, the general result was that the evaluation values’ distribution considerably changed in both FNL lexicons.
In both Figure 7 and Figure 8, the FNL lexicon preferred extreme values rather than middle ground values. However, both classifications coincide in leaving the very mid almost empty, with the exception that in Figure 7, it has a visible small peak in the mid value in contrast with Figure 6 and Figure 8. These results can be explained by taking into consideration the following:
  • Evaluation of the plots;
  • Universality of the mid-value;
  • Struggles with TE-adjective in LSV.
Regarding evaluating the plots, it was more natural to classify the plots roughly either in v L or v R . Acknowledging the context was unnecessary to decide whether the word is placed in v R or v L or v S . We conclude that the universality of FNL resides in the rough understanding of fuzzy prime EHs. Therefore, an EH is usually roughly either a v R or v L . However, when we try to fine-grain a particular value, we do need the linguistic context to clarify whether that word is a prototypical closer to the extremes of either v L or v R , or whether it is a lexical unit closer to the middle plot v S , which intersections with v L or v R .
Regarding the universality of the mid-value v S , it turns out that we could barely classify words in v S . Most of them are general enough to be included in almost every LSV. Some examples are: ordinary, average, average-sized, grey, central, medium, adult, similar, normal. We jump into the following conclusion according to this phenomena. Mid words are the universal words in an evaluation scale. Because of their lack of markedness, they can flexibly be included in most of the scales as v S , and they still keep the mid-meaning. Words in v S can be considered the fuzziest semantics units since they do not show any clear-cut semantic traits, and yet they roughly fit in several LSV. Table 2, shows some TE-adjective with an X in the middle. X means that the mid-plot is still existing; however, there is no specific prime EH for such plot.
Finally, regarding struggles with TE-adjective in LSV, Table 2 displays the tags that have been found and agreed after manually tagging both the Spanish and English corpus.
We have found more satisfying those scales that classifies more specific LSV, such as Fear-Courage, Esteem, Complexity, Capability-Skill, Temperature, or Value (economical) in contrast with these scales with a more general LSV such as Size, Beauty, Pleasing-Likability, Fullness, and so on. Using hyper-generic LSV such as Size, or Pleasing-Likability can bias a fine-grained classification, since it is easy to fit too borderline cases into those scales. Examples of too borderline cases in the LSV of Pleasing-Likability would be: happy, funny, or clean. Assuming that all these words transmit a semantic value of something being pleasing. The LSV of Fullness is a particular one. This LSV is a mishmash of evaluative heads that cannot completely fit in other LSV without creating a specific scale with almost no adjectives. Some of the cases are: informative as (full of information), illogic as (empty of logic), words with the morpheme full like help-ful as (full of help), same with words with -less or -free like painless (empty of pain), and dramafree (empty of drama), among others.
Therefore, one of the significant challenges in the classification of an FNL lexicon is (1) not being too specific without reason during the extraction of LSVs, and (2) not pushing the fit of borderline cases in a hyper-generic LSV.
On the other hand, two other big struggles arise when tagging adjectives:
(1)
It is complex to know under which LSV tag the adjective should be classified if we do not look at the linguistic context of the word,
(2)
Some adjectives might be suspicious of being able to appear in more than one LSV scale. It is the case of beautiful which could be Beauty: she is beautiful, and/or Pleasing-Likability: the food taste beautiful, and/or Judgment: it is a beautiful job.
These two struggles brought us to implement a new feature in the FNL lexicon. We are implanting a sub-tag of prototypical LSV ( L S V α ) and borderline LSV ( L S V γ ). This tag is particular for each lexical unit. Which means that if a prime EH such as beautiful is a tag with L S V α and L S V γ does not necessarily imply that the lexical units under this prime will require the same specific tagging.
Table 3 displays how every particular lexical unit is tagged with prototypes, borderline, and sentiment features. In the case of Table 3, the TE-adjective membership coincides in the three cases. Let us go back to Table 2. We will realize how all the TE are prepared for automatically converging in most cases: v L is usually equivalent for a prototypical negative, v S is usually equivalent for a prototypical neutral, and v R is usually equivalent for a prototypical positive.
However, in some cases, some adjectives do need a dual tag for prototype sentiment ( S α ) and borderline sentiment ( S γ ). These are the example of words such as sick, geek, among others.
Table 4 provides two examples in which its lexical units have borderline cases in LSV and sentiment. Both cases clearly specify that the borderline case is activated only when a specific LSV is triggered. Therefore, within a particular context, sick would trigger S γ in a sentence such as Jordan’s game is sick (skills). Same situation for geek. It would trigger a borderline sentiment in a sentence such as I am a Netflix geek (generates-interest).
Therefore, thanks to FNL, it is only necessary to specify the borderline L S V γ , and in which side of the fuzzy-graph the lexical unit will be in the borderline case ( v L or v R ).
We must mention that the borderline cases are challenging to extract. They arise as words with extreme variability, making it very difficult to know which exact word-pairing are borderline, and which borderline properties are triggering. The first step in our future work is working on an FNL lexicon with tagged-nouns. Therefore, we will be in a better position to evaluate the characterization of borderline cases.
Summing up, with this experimental work, we aimed:
  • To evaluate the model’s advantages and limitations when building a lexicon with FNL—improving the model with new linguistic properties.
  • To check if it is feasible to convert adjectives from the semantics of the evaluative expression into an equivalent positive-negative value.
  • To check in which cases the linguistic context is needed to classify an evaluative adjective in a LSV under a prime EH.

6. Conclusions

The model we have introduced takes into account the four main features when describing evaluative expressions:
  • Firstly, evaluative linguistic expression’s syntax, semantics, and sentiment are vague and gradient. The vagueness and gradience of its syntax and semantics are definable through the concepts of grammaticality and coercion.
  • Secondly, the semantics of an evaluative linguistic expression can be associated with a semantic prime with a sentiment value. A real number cannot represent the extension of this prime most of the time. Therefore, it is necessary to consider an abstract extension representing an intensity similar to a Liker-t scale.
  • Thirdly, every evaluative linguistic expression has a single lexical unit as an evaluative head. Such heads are not necessarily adjectives and have to be defined in terms of prototypical and borderline meanings.
  • Lastly, syntactic structures that violate syntactic constraints detonate semantic intensions equivalent to a semantic prime. Therefore, variability does not compromise the final processing of the meaning of the construction.
Our findings and research provide the basis for creating a lexicon of evaluative expressions for future computational applications. Furthermore, we have expanded the trichotomous expressions, extracting new ones from a sentiment analysis corpus by experts.
Additionally, we outline the following aspects related to some of the theoretical criticism this model might receive.
  • The linguistic mechanisms displayed in this paper to define evaluative expressions are mainly found in Section 3 and Section 4. Those mechanisms are mainly the linguistic constraints from FPGr (Section 3) and selected concepts of Fuzzy Natural Logic (FNL) adapted to FPGr (Section 4), which consist of using the description of the trichotomous expression v L , v S , v R as a constraint to define the semantics of an evaluative head. So, for example, some evaluative heads would have a constraint of V R that define its semantics, such as “big”, “tall”, “complex”, “beautiful”, and so on. In summary, the general approach to the model, its mechanisms such as the linguistic constraints, and its architecture are married to how FNL compute evaluative expressions.
  • Evaluative expressions as a “component” are identified with a lexicon. Therefore, building a lexicon is fundamental for applying both FPGr and FNL to characterize evaluative expressions, mainly because an evaluative head does not have a fixed linguistic category.
  • Our proposal is theoretical, including a proof-of-concept of how to build a lexicon of evaluative expressions. In the future, it is necessary to extend the tags on the evaluative expressions because it does need a lexicon to identify them, so their special fuzzy semantics can be characterized.
  • Regarding whether the system prevents over-generation or not, FPGr and FNL system does not generate language, treebanks, or tree structures; they only characterize language. The generation of constraints or establishing the number of constraints for each language is up to each language’s grammar or the experts configuring such grammar. If there is a problem with the overgeneration, it is not a fault of the system or the architecture. It is a fault of how the system was applied to define a particular language.
  • The grammaticality of syntax is considered, as the gradient is a fundamental part of the FPGr. However, the notion of gradient grammaticality is not as “radical” as it might seem under the architecture of FPGr. First, as observed in Section 3, the grammar is defined as a set of constraints. Thus, the set of constraints represents the knowledge of grammar. These constraints can be prototypical; therefore, they are, for sure, constraints of a specific grammar, or borderline, so they are “sometimes” constraints of specific grammar. That is, they are marked, and they appear in specific contexts that are not prototypical. Therefore, the grammar is fuzzy because it considers both prototypical and borderline constraints and weighs them. Secondly, a natural language input will have a degree of membership of a specific grammar according to the number of satisfied and violated constraints of such specific grammar. Therefore, the degree of grammaticality is defined as the degree of belongingness of an input, with regards to specific natural language grammar. For further information, see Torrens-Urrutia [19].
  • FPGr and FNL do not consider extra-linguistic interpretations, since such a case would entail pragmatics. Therefore, it cannot generate from an input such as “I love your food” an output as: “I love the fact that you cooked for me, regardless of the quality of the food”. Such an interpretation is considered pragmatic due to extra-linguistic reasons, i.e., theory of politeness, context, intention, implicatures, etc. (see Figure 2). Semantics in FPGr only focuses on the essential part of the meaning, the most primitive one, that is why we defend the notion of semantic primes as a hypernym, and the system defines constraints of evaluations regarding a prime. These constraints are v R , v S , v L .

7. Future Work

The future work for Fuzzy Natural Logic involves expanding the fundamental trichotomous expressions and LSVs. For such a task, it will be necessary to either explore computational techniques which can automatically extract the LSV, or to perform a systematisation of surveys in which a significant number of native speakers with linguistic knowledge of their language perform classifications of words. Regardless of the method, the final objective should be the creation of an ontology of the evaluative meaning based on the feature descriptions that this work pointed out.
Figure 9 shows an ontology made ad hoc as a guidance for a future design of a computational one. On the top, at the same level, we find both the Prime Sentiment, and Prime semantics, which will typically share polarities; that is, v L will be prototypically negative. At the same time, v R will be prototypically positive. From the most abstract notion of prime, we will find the next level, at which we find the fundamental trichotomous expressions with their L S V s . The closer we get to the bottom, the more metaphorical evaluations will be found. The prototypical relations are marked with a straight line, while fractured lines denote the borderline relations. The lines will state the semantic distribution between the prime and the related words if they have a relationship to v L , v S , or v R . For example, Einstein has a borderline relationship of being a metaphorical evaluation with a semantic polarity of v R in the LSV prime of intelligence, which means that it will be included in the fuzzy set of the T E E H of intelligent. The ontology must be able to define evaluative words as a combination of several fundamental trichotomous expressions, such as the case of cute. Typically, something cute is referred for something closer to be pretty, but at the same time, it shares traits with something being quite small. We would not say that a big truck is cute, but a cartooney small toy truck could be cute. Independently of how right is this semantic analysis, the main advantage of such an ontology is that it provides ‘explainability’, and the particular cases can be changed and improved. Thus, if a semantic relation is inaccurate, it could be fixed. Some words might have a lot of semantic relationships. It is a case of do-it-yourself. This lexical unit operates as an evaluation, providing a lot of different borderline meanings.
Creating several ontologies would help linguistics evaluate complexity of the syntax and semantics across languages for future experiments and linguistic applications in explainable AI. That is, computing with constraints how similar the meanings between languages are, elaborate ontologies of the history of language through the evolution of importance, and the evaluation of the linguistic universality in the semantic domain, among others.

Author Contributions

All authors have contributed equally. Conceptualization, A.T.-U., V.N. and M.D.J.-L.; Formal analysis, A.T.-U., V.N. and M.D.J.-L.; Writing—original draft, A.T.-U., V.N. and M.D.J.-L.; Writing—review & editing, A.T.-U., V.N. and M.D.J.-L. All authors have read and agreed to the published version of the manuscript.

Funding

This paper has been supported by the project CZ.02.2.69/0.0/0.0/18_053/0017856 “Strengthening scientific capacities OU II”.

Data Availability Statement

https://github.com/sfu-discourse-lab/SO-CAL (accessed on 10 January 2022).

Acknowledgments

We also want to thank Maite Taboada for her collaboration and support during this research and Eduard Mir Neira for his work with the Spanish Lexicon LSV scales.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Novák, V. The Concept of Linguistic Variable Revisited. In Recent Developments in Fuzzy Logic and Fuzzy Sets; Sugeno, M., Kacprzyk, J., Shabazova, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2020; pp. 105–118. [Google Scholar]
  2. Novák, V. Fuzzy Logic in Natural Language Processing. In Proceedings of the International Conference FUZZ-IEEE, Naples, Italy, 9–12 July 2017. [Google Scholar]
  3. Novák, V.; Perfilieva, I.; Dvořák, A. Insight Into Fuzzy Modeling; Wiley & Sons: Hoboken, NJ, USA, 2016. [Google Scholar]
  4. Novák, V. What is Fuzzy Natural Logic. In Integrated Uncertainty in Knowledge Modelling and Decision Making; Huynh, V., Inuiguchi, M., Denoeux, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2015; pp. 15–18. [Google Scholar]
  5. Novák, V. Fuzzy Natural Logic: Towards Mathematical Logic of Human Reasoning. In Fuzzy Logic: Towards the Future; Seising, R., Trillas, E., Kacprzyk, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2015; pp. 137–165. [Google Scholar]
  6. Novák, V. Evaluative linguistic expressions vs. fuzzy categories? Fuzzy Sets Syst. 2015, 281, 81–87. [Google Scholar] [CrossRef]
  7. Novák, V. Mathematical Fuzzy Logic: From Vagueness to Commonsese Reasoning. In Retorische Wissenschaft: Rede und Argumentation in Theorie und Praxis; Kreuzbauer, G., Gratzl, N., Hielb, E., Eds.; LIT-Verlag: Wien, Austria, 2008; pp. 191–223. [Google Scholar]
  8. Lakoff, G. Linguistics and natural logic. Synthese 1970, 22, 151–271. [Google Scholar] [CrossRef] [Green Version]
  9. Novák, V.; Murinová, P.; Boffa, S. On the properties of intermediate quantifiers and the quantifier “MORE-THAN”. In Information Processing and Management of Uncertainty in Knowledge-Based Systems, Part III; Medina, J., Ojeda-Aciego, M., Verdegay, J.L., Pelta, D.A., Cabrera, I.P., Bouchon-Meunier, B., Yager, R.R., Eds.; Springer Nature: Zürich, Switzerland, 2020; pp. 159–172. [Google Scholar]
  10. Murinová, P.; Novák, V. The theory of intermediate quantifiers in fuzzy natural logic revisited and the model of “Many”. Fuzzy Sets Syst. 2020, 388, 56–89. [Google Scholar] [CrossRef]
  11. Novák, V.; Murinová, P. A formal model of the intermediate quantifiers “A few, Several, A little”. In Fuzzy Techniques: Theory and Applications; Kearfott, R., Batyrshin, I., Reformat, M., Ceberio, M., Kreinovich, V., Eds.; Springer: Cham, Switzerland, 2019; pp. 429–441. [Google Scholar]
  12. Nguyen, L.; Novák, V. Forecasting seasonal time series based on fuzzy techniques. Fuzzy Sets Syst. 2019, 361, 114–129. [Google Scholar] [CrossRef]
  13. Novák, V.; Pavliska, V.; Perfilieva, I.; Stepnicka, M. F-transform and Fuzzy Natural logic in Time Series Analysis. In Proceedings of the EUSFLAT Conference, Milano, Italy, 11–13 September 2013. [Google Scholar]
  14. Novák, V. Linguistic characterization of time series. Fuzzy Sets Syst. 2016, 285, 52–72. [Google Scholar] [CrossRef]
  15. Novák, V.; Perfilieva, I.; Dvořák, A.; Chen, G.; Wei, Q.; Yan, P. Mining pure linguistic associations from numerical data. Int. J. Approx. Reason. 2008, 48, 4–22. [Google Scholar] [CrossRef] [Green Version]
  16. Torrens Urrutia, A.; Jiménez-López, M.D.; Novák, V. Fuzzy Natural Logic for Sentiment Analysis: A Proposal. In Proceedings of the International Symposium on Distributed Computing and Artificial Intelligence, Salamanca, Spain, 6–8 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 64–73. [Google Scholar]
  17. Torrens Urrutia, A. Towards a fuzzy grammar for natural language grammars. In Recerca en Humanitats 2018; Publicacions URV: Reus, Spain, 2018; pp. 271–282. [Google Scholar]
  18. Torrens Urrutia, A. An Approach to Measuring Complexity with a Fuzzy Grammar & Degrees of Grammaticality. In Proceedings of the Workshop on Linguistic Complexity and Natural Language Processing, Santa Fe, NM, USA, 25 August 2018; pp. 59–67. [Google Scholar]
  19. Torrens Urrutia, A. A Formal Characterization of Fuzzy Degrees of Grammaticality for Natural Language. Ph.D. Thesis, Universitat Rovira i Virgili, Tarragona, Spain, 2019. [Google Scholar]
  20. Torrens Urrutia, A.; Jiménez-López, M.D.; Brosa-Rodríguez, A. A Fuzzy Approach to Language Universals for NLP. In Proceedings of the IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Virtual, 11–14 July 2021; pp. 1–6. [Google Scholar]
  21. Hemmatian, F.; Sohrabi, M.K. A survey on classification techniques for opinion mining and sentiment analysis. Artif. Intell. Rev. 2019, 52, 1495–1545. [Google Scholar] [CrossRef]
  22. Yadav, A.; Vishwakarma, D.K. Sentiment analysis using deep learning architectures: A review. Artif. Intell. Rev. 2020, 53, 4335–4385. [Google Scholar] [CrossRef]
  23. Taboada, M. Sentiment Analysis: An Overview from Linguistics. Annu. Rev. Linguist. 2016, 2, 325–347. [Google Scholar] [CrossRef] [Green Version]
  24. Baccianella, S.; Esuli, A.; Sebastiani, F. Sentiwordnet 3.0: An enhanced lexical resource for sentiment analysis and opinion mining. In Proceedings of the LREC, Valletta, Malta, 17–23 May 2010; Volume 10, pp. 2200–2204. [Google Scholar]
  25. Mohammad, S.; Dunne, C.; Dorr, B. Generating high-coverage semantic orientation lexicons from overtly marked words and a thesaurus. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, Singapore, 6–7 August 2009; pp. 599–608. [Google Scholar]
  26. Wilson, T.; Wiebe, J.; Hoffmann, P. Recognizing contextual polarity in phrase-level sentiment analysis. In Proceedings of the Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing, Vancouver, BC, Canada, 6–8 October 2005; pp. 347–354. [Google Scholar]
  27. Wiebe, J.; Wilson, T.; Bruce, R.; Bell, M.; Martin, M. Learning subjective language. Comput. Linguist. 2004, 30, 277–308. [Google Scholar] [CrossRef] [Green Version]
  28. Blache, P. Representing syntax by means of properties: A formal framework for descriptive approaches. J. Lang. Model. 2016, 4, 183–224. [Google Scholar] [CrossRef] [Green Version]
  29. Universal Dependency Corpora. Available online: https://universaldependencies.org/ (accessed on 1 September 2021).
  30. Blache, P.; Rauzy, S.; Montcheuil, G. MarsaGram: An excursion in the forests of parsing trees. In Proceedings of the 12th Language Resources and Evaluation Conference, LREC 2020, Marseille, France, 11–16 May 2020; Volume 10, p. 7. [Google Scholar]
  31. Manca, V.; Jiménez-López, M.D. GNS: Abstract Syntax for Natural Languages. Triangle 2012, 8, 55–79. [Google Scholar] [CrossRef]
  32. Fillmore, C.J. The mechanisms of “construction grammar”. In Proceedings of the Annual Meeting of the Berkeley Linguistics Society, 13–15 February 1988; Volume 14, pp. 35–55. Available online: http://journals.linguisticsociety.org/proceedings/index.php/BLS/article/viewFile/1794/1566 (accessed on 10 January 2022).
  33. Goldberg, A.E. Constructions at Work: The Nature of Generalization in Language; Oxford University Press on Demand: Oxford, UK, 2006. [Google Scholar]
  34. Fillmore, C.J.; Baker, C. A frames approach to semantic analysis. In The Oxford Handbook of Linguistic Analysis; Oxford University Press: Oxford, UK, 2010. [Google Scholar]
  35. Goldberg, A. Verbs, constructions and semantic frames. In Syntax, Lexical Semantics, and Event Structure; Oxford University Press: Oxford, UK, 2010; pp. 39–58. [Google Scholar]
  36. Müller, G. Optimality, markedness, and word order in German. Linguistics 1999, 37, 777–818. [Google Scholar] [CrossRef] [Green Version]
  37. Wierzbicka, A. Semantics: Primes and Universals: Primes and Universals; Oxford University Press: Oxford, UK, 1996. [Google Scholar]
  38. Erk, K.; Padó, S.; Padó, U. A flexible, corpus-driven model of regular and inverse selectional preferences. Comput. Linguist. 2010, 36, 723–763. [Google Scholar] [CrossRef]
  39. Baroni, M.; Lenci, A. Distributional memory: A general framework for corpus-based semantics. Comput. Linguist. 2010, 36, 673–721. [Google Scholar] [CrossRef]
  40. Greenberg, C.; Sayeed, A.; Demberg, V. Improving unsupervised vector-space thematic fit evaluation via role-filler prototype clustering. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics Human Language Technologies, Denver, CO, USA, 31 May–5 June 2015; ACL: Denver, CO, USA, 2015; pp. 21–31. [Google Scholar]
  41. Santus, E.; Chersoni, E.; Lenci, A.; Blache, P. Measuring thematic fit with distributional feature overlap. arXiv 2017, arXiv:1707.05967. [Google Scholar]
  42. Crystal, D. On keeping one’s hedges in order. Engl. Today 1988, 4, 46–47. [Google Scholar] [CrossRef]
  43. Layman, L. Reticence in oral history interviews. In The Oral History Reader; Routledge: London, UK, 2015; pp. 234–252. [Google Scholar]
  44. Islam, J.; Xiao, L.; Mercer, R.E. A lexicon-based approach for detecting hedges in informal text. In Proceedings of the 12th Language Resources and Evaluation Conference, Marseille, France, 11–16 May 202; pp. 3109–3113.
  45. Hassan, A.S.; Said, N.K.M. A pragmatic study of hedges in American political editorials. Int. J. Res. Soc. Sci. Humanit. 2020, 10. [Google Scholar] [CrossRef]
  46. Taboada, M.; Brooke, J.; Tofiloski, M.; Voll, K.; Stede, M. Lexicon-based methods for sentiment analysis. Comput. Linguist. 2011, 37, 267–307. [Google Scholar] [CrossRef]
Figure 1. (a) Fuzzy sets modeling extensions of expressions that form the fundamental linguistic trichotomy; (b) Scheme of extensions of evaluative expressions with modifiers.
Figure 1. (a) Fuzzy sets modeling extensions of expressions that form the fundamental linguistic trichotomy; (b) Scheme of extensions of evaluative expressions with modifiers.
Mathematics 10 02760 g001
Figure 2. Semantic hierarchy of evaluative meaning.
Figure 2. Semantic hierarchy of evaluative meaning.
Mathematics 10 02760 g002
Figure 3. Characterization of the universal construction of evaluative expressions.
Figure 3. Characterization of the universal construction of evaluative expressions.
Mathematics 10 02760 g003
Figure 4. Universal construction of evaluative expressions (syntax and semantics).
Figure 4. Universal construction of evaluative expressions (syntax and semantics).
Mathematics 10 02760 g004
Figure 5. Trichotomous Expression of [small-medium-big] in FNL.
Figure 5. Trichotomous Expression of [small-medium-big] in FNL.
Mathematics 10 02760 g005
Figure 6. SO-CAL Adjectives by polarity.
Figure 6. SO-CAL Adjectives by polarity.
Mathematics 10 02760 g006
Figure 7. English Adjectives SO-CAL vs. FNL lexicon.
Figure 7. English Adjectives SO-CAL vs. FNL lexicon.
Mathematics 10 02760 g007
Figure 8. Spanish Adjectives SO-CAL vs. FNL lexicon.
Figure 8. Spanish Adjectives SO-CAL vs. FNL lexicon.
Mathematics 10 02760 g008
Figure 9. Ontology of evaluative expressions.
Figure 9. Ontology of evaluative expressions.
Mathematics 10 02760 g009
Table 1. Sample of semantic orientation values for lexicon-based approaches from Taboada [23].
Table 1. Sample of semantic orientation values for lexicon-based approaches from Taboada [23].
SentiWordNet [24] Baccianella et al. (2010)Macquire Dictionary [25] Mohammad et al. (2009)Subjectivity Dictionary [26] Wilson et al. (2005)SO-CAL [23] Taboada (2016)
goodPositivePositivePositive
(weak)
3
excellentPositivePositivePositive
(strong)
5
masterpiecePositivePositivePositive
(strong)
5
badNegativeNegativeNegative
(strong)
−3
terribleNegativeNegativeNegative
(strong)
−5
disasterNegativeNegativeNegative
(strong)
−4
Table 2. Linguistic Semantic Variable (LSV) and Prime Evaluative Head (EH) tags in Fuzzy Natural Logic (FNL) Lexicon.
Table 2. Linguistic Semantic Variable (LSV) and Prime Evaluative Head (EH) tags in Fuzzy Natural Logic (FNL) Lexicon.
FNL Tags in English and Spanish Lexicon
LSVJudgmentEsteemBeautySizeCapability-Skills
Primes EH〈negative/bad-medium/normal-positive/good〉〈hated-X-loved〉〈ugly-X-beautiful〉〈small/short-medium-big/long〉〈disable-average-capable〉
LSVComplexityFear-CourageFullnessIndeterminacyIntelligence
Primes EH〈simple-normal-complex〉〈scared-X-brave〉〈empty-X-full〉〈blurred-X-clear〉〈stupid-average-intelligent〉
LSVGenerates-InterestPleasing-LikabilityProximityVeridicalitySimilarity-Usual
Primes EH〈boring-X-interesting〉〈disgusting-X-pleasing〉〈far-central-close〉〈fake/false-X-real/truth〉〈different-similar-usual〉
LSVSpeedStrength-IntensityTemperatureTime-LifetimeWorth-Value
Primes EH〈slow-medium-fast〉〈weak/fragile-X-intense/strong〉〈cold-X-hot〉〈life/new/young/
beggining- medium/
adult-death/old/end〉
〈worthless-X-worthy〉
LSVValue (economical)
Primes EH〈cheap-affordable-expensive〉
Table 3. Example of a tagged lexical unit.
Table 3. Example of a tagged lexical unit.
Lexical Unit: Beautiful
L S V α Beauty v R
L S V γ Judgment, Pleasing-Likability v R
S e n t i m e n t positive v R
Table 4. Tagging lexical units with borderline cases.
Table 4. Tagging lexical units with borderline cases.
Lexical Unit: SickLexical Unit: Geek
L S V α Judgment v L L S V α Similarity-Usual v L
L S V γ Pleasing-Likability
Capability-Skills
v L
v R
L S V γ Intelligence v R
S α negative v L S α negative v L
S γ positive
{IFF capability-skills}
v R S γ positive
{IFF generates-interest}
v R
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Torrens-Urrutia, A.; Novák, V.; Jiménez-López, M.D. Describing Linguistic Vagueness of Evaluative Expressions Using Fuzzy Natural Logic and Linguistic Constraints. Mathematics 2022, 10, 2760. https://doi.org/10.3390/math10152760

AMA Style

Torrens-Urrutia A, Novák V, Jiménez-López MD. Describing Linguistic Vagueness of Evaluative Expressions Using Fuzzy Natural Logic and Linguistic Constraints. Mathematics. 2022; 10(15):2760. https://doi.org/10.3390/math10152760

Chicago/Turabian Style

Torrens-Urrutia, Adrià, Vilém Novák, and María Dolores Jiménez-López. 2022. "Describing Linguistic Vagueness of Evaluative Expressions Using Fuzzy Natural Logic and Linguistic Constraints" Mathematics 10, no. 15: 2760. https://doi.org/10.3390/math10152760

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop