Next Article in Journal
CIDOC2VEC: Extracting Information from Atomized CIDOC-CRM Humanities Knowledge Graphs
Next Article in Special Issue
Multimodal Fake News Detection
Previous Article in Journal
Finding Central Vertices and Community Structure via Extended Density Peaks-Based Clustering
Previous Article in Special Issue
Exploring the Impact of COVID-19 on Social Life by Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Automated Semantic Explainability of Multimedia Feature Graphs

1
Faculty of Mathematics and Computer Science, University of Hagen, D-58097 Hagen, Germany
2
Academy for International Science & Research (AISR), Londonderry BT48 7ER, UK
*
Author to whom correspondence should be addressed.
Information 2021, 12(12), 502; https://doi.org/10.3390/info12120502
Submission received: 28 September 2021 / Revised: 22 November 2021 / Accepted: 26 November 2021 / Published: 2 December 2021
(This article belongs to the Special Issue Sentiment Analysis and Affective Computing)

Abstract

:
Multimedia feature graphs are employed to represent features of images, video, audio, or text. Various techniques exist to extract such features from multimedia objects. In this paper, we describe the extension of such a feature graph to represent the meaning of such multimedia features and introduce a formal context-free PS-grammar (Phrase Structure grammar) to automatically generate human-understandable natural language expressions based on such features. To achieve this, we define a semantic extension to syntactic multimedia feature graphs and introduce a set of production rules for phrases of natural language English expressions. This explainability, which is founded on a semantic model provides the opportunity to represent any multimedia feature in a human-readable and human-understandable form, which largely closes the gap between the technical representation of such features and their semantics. We show how this explainability can be formally defined and demonstrate the corresponding implementation based on our generic multimedia analysis framework. Furthermore, we show how this semantic extension can be employed to increase the effectiveness in precision and recall experiments.

1. Introduction and Motivation

Bridging the semantic gap has been a research goal for many years. Narrowing down the gap between detected features from multimedia assets (i.e., images, video, audio, text, and social media) and their semantic representation has led to numerous investigations and research in the field of Multimedia Information Retrieval (MMIR) [1]. With the extraction of MMIR supporting features and the integration of these features into data models, internal representations of these features are created. MMIR applications process these representations e.g., w.r.t. indexing, retrieval, and querying and employ them to analyze relationships between or within assets. In particular, the topic of querying becomes highly important, as e.g., every single minute, 147,000 photos are uploaded to Facebook, 41.6 million WhatsApp messages are sent, or 347,000 stories are posted by Instagram [2] and the users still expect highly accurate query results. Due to higher resolutions of images and video, the Level-Of-Detail (LOD) in multimedia assets has increased significantly. Current professional cameras such as the Sony α7R IV 35 are equipped with a resolution of 61.0 megapixels [3], and smartphones such as the Xiaomi Redmi Note 10 Pro even push that boundary to 108 megapixels [4]. This high LOD is also reflected by other multimedia types, e.g., text, where news agencies maintain huge archives of textual information, enriched by user comments, web information, or social media [5].
The figures given above demonstrate that it has become more and more important to semantically understand MMIR supporting features to increase the precision and recall of quasi-semantic MMIR querying. The Semantic Web [6] and all its related technologies and concepts are a sound basis for knowledge representation, reasoning, inferencing, and truth maintenance. To bridge the gap between detected features and their semantic representation, a machine-readable and formal approach is required, as well as a human-understandable explanation of the corresponding processing steps. However, the requirement of a high LOD, the increasing number of assets, and the demand for fast and accurate semantic querying contradict each other and lead to further challenges in the area of explainability in a human-understandable way.
In this paper, we present a solution for the automated explainability of MMIR processing steps in the form of human-understandable natural language texts based on a semantic modeling, which also supports inferencing and reasoning.

2. State-of-the-Art and Related Work

This section gives an overview of current techniques and standards for semantic indexing and retrieval and discusses related work. We introduce the Multimedia Feature Graph (MMFG) and the Graph Code concept, also discussing semantic analysis and intelligent information retrieval methods.
In previous work, we already introduced the Generic Multimedia Analysis Framework (GMAF) [7,8,9,10] as a unifying framework that can integrate various multimedia features into a single data structure. The GMAF utilizes selected existing technologies as plugins to support various multimedia feature detection algorithms for text (e.g., social media posts, descriptions, tag lines) [11,12,13], images (especially object detection and spatial relationships including the use of machine learning) [11,14,15,16], audio (transcribed to text) [15,17,18], and video including metadata [19] and detected features [18,20,21].
The GMAF produces a Multimedia Feature Graph (MMFG), which is defined in [7] and represents various integrated multimedia features. Within an application, these MMFGs are typically represented as a collection. For the remainder of this paper, some internal structures, sets, and definitions of the MMFG are relevant (see also [7,22]):
  • F V T M M F G = { f t 1 , . . . , f t n } is the feature term vocabulary, i.e., the set of feature term labels, which represent detected features. Elements of F V T M M F G are represented by N o d e s ( n ) in the MMFG graph structure.
  • F V T C o l l = i = 1 n F V T M M F G i is the feature term vocabulary of the collection of MMFGs within a MMIR application.
  • the set F R T M M F G = { c n , s , s r } representing the feature relationship types of an MMFG, where c n represents the “child” relationship, s represents the “synonym” relationship, and “sr” represents the “spacial” relationship between feature vocabulary terms. Elements of F R T M M F G are represented by links between N o d e s in the MMFG graph structure.
As the integration of multimedia features within MMFGs produce a much higher level-of-detail (LOD), effective and efficient algorithms are required to process these feature graphs. Hence, we introduced Graph Codes, which are a 2D projection of MMFGs and corresponding algorithms, performing calculations based on matrices instead of graph traversal and also support the higher LOD of MMFGs [7]. For these Graph Codes, we introduced a metric for similarity calculation, the mathematical concepts of the indexing and retrieval algorithms, including a detailed evaluation regarding performance, precision, and recall. Figure 1 briefly summarizes this concept as a foundation for subsequent sections of this paper. Figure 1a shows a snippet of an exemplary MMFG visualized in a graph editing tool [23]. Figure 1b illustrates a part of the MMFG in an object diagram including various node and relationship types, which can be represented as a Graph Code table (see Figure 1c) based on the graph’s valuation matrix. A Graph Code’s matrix representation is shown in Figure 1d, where the correspondence to mathematical matrix calculations is obvious (It is notable for multimedia that due to the current object detection algorithms, MMFGs and their corresponding Graph Codes contain semantic information (e.g., “is a”), spatial information (e.g., “above”), and also temporal information (by the temporal ordering of sub-collections of Graph Codes)).
To calculate MMIR results based on Graph Codes, a Graph Code Metric is defined, which can be applied for similarity algorithms. In general, every detected feature can be regarded as a multimedia indexing term. The indexing term of any relevant feature thus becomes part of the vocabulary of the overall retrieval index. In Multimedia Indexing and Retrieval (MMIR), typically these terms have structural and/or semantic relationships to each other and represent the basis for semantic query construction and result representation. In [7], we already defined a metric for similarity of Graph Codes, which is a triple M G C = ( M F , M F R , M R T ) containing a feature-metric M F based on the vocabulary, a feature-relationship-metric M F R based on the possible relationships, and a feature-relationship-type-metric M R T based on the actual relationship types. This metric can be applied for result representation and has to be considered when constructing corresponding queries. Querying and result presentation based on Graph Codes is discussed in [24].
Current Graph Code Query Construction technologies employ structured query languages (e.g., SQL, OQL, XML-Query), including Visual Query Languages (VQLs) and Natural Language Querying (NL) [25]. A Meaning Driven Data Query Language (MDDQL) [25,26] can support query construction by system-made suggestions of natural language based terms. In the field of Natural Language Processing (NLP), there have been several approaches to automatically translate natural language into structured queries, e.g., NLP to SPARQL processing [27,28]. Typically, results of these kinds of queries are represented in the form of ranked lists. All these query construction methodologies require semantic modeling.
Semantic Representation is covered by the concepts and standards defined by the Semantic Web [29], where the manual, semiautomatic, and automated generation of annotations is defined. The basis for these representations and annotations is a set of domain-oriented vocabulary terms. Once a basic "subclass" relationship is introduced between vocabulary terms, taxonomies can be built, which structure these terms in the form of class/subclass relationships. Typically, taxonomies also contain a consistent set of predefined textual labels and synonyms. Thesauri can be used to model further relationships (e.g., “broader”, “narrower”) for additional structuring, scoping, and increased expressiveness. Resource Description Framework (RDF) [6] can serve as a foundation. It covers the description of any kind of resource by employing XML Syntax. The Resource Description Framework Schema (RDFS) provides domain specific extension points and a standardized model of exchanging RDF documents. As RDF is based on XML, it can automatically be represented in the form of a graph model, which provides the opportunity to employ a mapping to the MMFG on a structural level. RDF-techniques such as publishing or linking, with a shared data model can act as a base layer for other technologies [6,29,30,31]. Finally, ontologies describe arbitrary relationships between taxonomy terms (now called concepts) going beyond the hierarchical taxonomy structure. OWL [29] represents these concepts and relationships as classes and properties [32]. Once such a well-defined formal semantic model is in place, reasoning and inference algorithms can also be applied to such semantic representations [33].
However, to clearly define how the concepts of an ontology can be combined automatically (e.g., during automatic inferencing), a well-defined grammar [34] is required. Based on such grammars, an algorithmic implementation can distinguish between valid and invalid statements of a given formal or natural language. According to [34], a grammar G = ( V , T , P , S ) for a language L is defined by the tuple of vocabulary terms V, the list of terminal symbols T, which terminate valid sentences of L, production rules P, which describe valid combinations of vocabulary terms and a set of starting symbols S for sentences of L.
In [35], PS-grammars (Phrase Structure grammars) are employed as a specialized form to generate language terms by production rules, in which the left side of the rule is replaced by the right side. If, e.g., α β is a production rule in P, and ϕ , ρ are literals in V, then ϕ α ρ ϕ β ρ is a direct replacement. Rules of PS-Grammars (PSG) are further detailed by four types, 0: unlimited PS-rule, 1: context sensitive PS-rule, 2: context-free PS-rule, 3: regular PS-rule, which denote systematic restrictions of the production rules. These restrictions lead to a hierarchy of formal languages and the corresponding calculation and validation complexities (see Table 1).
Typically, when defining grammars, the set V will contain additional classes to structure the possible production rules (typically defined as Chomsky rules [34]), e.g., classes to describe Nominal Phrases (NP), Verbal Phrases (VP), Prepositional Phrases (PP), or other word types such as Adjectives (ADJ), and their location in validly produced sentences [35]. In many cases, grammars are designed that V T = . As an example, the sentence “The hat is above the head” can be represented by the context-free grammar G e n = ( V e n , T e n , P e n , S e n ) for simple English sentences:
  • V e n = { S e n , N P , V P , V , N , D E T , P R } representing the variables of the grammar.
  • T e n = { t h e , h a t , i s , a b o v e , h e a d } is the set of terminal symbols, with V T = .
  • The production rules for this grammar can be defined as follows:
    P e n = { S e n N P V P , V P V P P , N P D E T N , P P P R N P }
The production rules of P S G r a m m a r s can be visualized in form of so called PS-trees [35]. Figure 2 shows such a PS-tree with the sentence, “The hat is above the head” based on the exemplary MMFG of Figure 1b. In this example, the use of Nominal Phrases (NP), Verbal Phrases (VP), Prepositional Phrases (PP), and the Start Symbol S are also illustrated. For example:
  • The NP, “The hat”, consists of the determiner “The”, and the noun “hat”.
  • The NP, “the head”, is built by the determiner “the”, and the noun “head”.
  • The PP, “above the head”, is constructed by the preposition “above”, and the NP “the head”.
  • The VP, “is above the head”, consists of the verb “is”, and the PP “above the head”.
  • The starting symbol for this sentence is constructed by a NP, “The hat” and the VP, “is above the head”.
By applying these production rules, both construction and the analysis of well-formed sentences can be approved. The introduction of a grammar is also a prerequisite for reasoning [35], where additional semantic features are generated by calculating inferences and conclusions [36]. A formal grammar for MMFGs and Graph Codes and the corresponding concepts based on this grammar, is defined in Section 3 (modeling) and will be the basis for constructing Semantic MMFGs (SMMFG).
To manage and maintain such semantic information based on MMFGs, several more technical Knowledge Representation and Processing Systems have been specified and introduced. For example, the W3C introduced the Simple Knowledge Organization System (SKOS) [37] as a standard way to represent knowledge organization systems with RDF [6]. Reasoning (i.e., the automated extension and maintenance of information), Truth Maintenance (i.e., the automated calculation of information validity), and Inference Systems (i.e., deducting new information based on logical rules) also contribute to improving the semantic information of a MMIR collection [33,38,39].
One common approach for Automated Reasoning and Inferencing is the concept of Non-monotonic Reasoning [38], which is based on justified beliefs and reasonable assumptions. Typically, so called Default Logics are employed to represent knowledge in form of rules. For example, the rule A : b / C is intended to state that, “if A is believed, and each b ∈ B can be consistently believed, then C should be believed”. A is a prerequisite, b a set of justifications [33,38]. For the calculation and representation of Default Logic, two major approaches serve as a foundation and are named after their inventors Reiter [40] and Poole [41]. Both approaches result in the common concept of Knowledge Extensions, which represent the set of possible rules, which are assumed (or calculated) to be believed.
Semantic Querying and Retrieval can be performed by employing, e.g., SPARQL [42], which is a standardized query language that operates on RDF, RDFS, or OWL representations and also supports the inclusion of semantic features. Semantic reasoning, which is applied to the underlying semantic representation of an ontology, is often called, “intelligent retrieval”. This means that automatic reasoning can derive new semantic feature annotations from existing feature representations. This further means that newly derived features have not been detected, but are actually derived by means of reasoning. Thus, the Multimedia object is annotated with additional features that are not a result of extraction or detection, but derived by logical and semantic reasoning. The resulting feature index is extended and retrieval results will be more accurate where additional features will be attributed to MMFGs automatically [36]. In the remainder of this paper, we will call the originally detected or extracted features, Internal Feature ( F I n t ), and the features that have been derived by reasoning External Features ( F E x t ).
However, to follow this systematic approach, typically a substantial manual effort is required to map syntactic feature representations to semantic representations. Further, the introduction of rules, a basic logic, and truth maintenance criteria must be performed manually. Hence, in Section 3 we define a formal, standardized, and automated approach for the integration of these systems.
In summary, we can state that current technologies provide a sufficient set of appropriate algorithms, tools, and concepts for semantic modeling, representation, indexing, and retrieval. However, concepts such as RDF, RDFS, OWL, SPARQL, or the Semantic Web mostly rely on graph-based semantic representation structures and thus underlie similar constraints regarding efficiency as syntactic feature graphs. The introduction of grammars provides a standardized way of constructing and analyzing well formed sentences. As Graph Codes provide a set of algorithms to significantly increase effectiveness, LOD, and efficiency for graph-based IR algorithms; we now present their application and extension into Semantic Graph Codes and the corresponding processing algorithms.

3. Modeling and Design

In this section, we define and introduce several semantic extensions of syntactic MMFGs, which serve as a basis for optimized Semantic Graph Code processing and the corresponding application of semantic concepts such as annotation, topic modeling, reasoning, or inferencing. We also introduce Explainable Semantic Graph Codes, providing a human-readable representation of multimedia feature graphs.
The MMFG has been designed to represent MMIR features on a pure syntactical basis, containing only, Internal Features F I n t . To support semantic extensions for MMFGs, we apply Semantic Web concepts [29]. In addition, we introduce a context-free PS-grammar G M M F G for the construction of human-readable, i.e., valid natural language textual phrases and statements based on MMFG features. This enables the construction of a formal semantic representation on the one hand, and establishes the basis for natural language textual explanations on the other hand. This combination leads to a well-defined semantic representation of MMFGs, Semantic Multimedia Feature Graph (SMMFG), and to the Explainable Semantic Multimedia Feature Graph (ESMMFG). In particular, ESMMFGs can serve as a basis for inferencing and thus support the production of additional External Features F E x t . The employment of a PS-grammar in addition to the semantic extension has the advantage that the representation of such an MMFG is not only machine-readable, but also “human-readable”, i.e., the representation supports transparency, explainability, and reproducibility for humans.
The structure of this section follows the logical sequence of extensions from simple MMFGs to semantic SMMFGs and explainable ESMMFGs. Hence, in Section 3.1, we discuss the initial formal foundation of the chosen approach. In Section 3.2, annotations for MMFGs are introduced, which are then employed to define the semantic extension in the form of SMMFGs in Section 3.3. Finally, explainability is introduced by the definition of a PS-grammar resulting in ESMMFGs in Section 3.4.

3.1. Formal Representation of an MMFG

As shown in the state-of-the-art discussion (see Section 2), MMFGs are purely syntactical structures based on a data model, which forms a multimedia feature graph with nodes and edges. The formal model of such a syntactic MMFG representation is shown in Figure 3a. Here, and in subsequent sections, we further detail and extend the exemplary MMFG from Section 2, which serves as a exemplary syntactic representation of a simple MMFG (see Figure 3b). To support the formal representation of such an MMFG with a formal language (i.e., grammar), an additional element—the root node S M M F G —must be defined, which acts as a starting symbol for valid formal language expressions (see Section Table 1). This is shown in Figure 3b.
A basic grammar G M M F G = ( V M M F G , T M M F G , P M M F G , S M M F G ) for the construction of valid MMFG-sentences (see Section 2) can be formalized as
  • V M M F G = F R T F V T is the set of feature relationship terms F R T and the set of feature vocabulary terms ( F V T ) .
  • T M M F G is a set of textual labels ( L B L ) for elements in F R T and F V T .
  • P M M F G is the set of production rules, which produce sentences based on, F R T and F V T . In its simplest form, P can be defined as:
    P = { S F V T F R T F V T F V T L B L F R T L B L }
  • S M F G is the root node of an MMFG.
When such an initial grammar is applied to the exemplary MMFG of Figure 3c, formal language expressions such as, “person has-child head. head has-child hat. hat is-above head.”, can be produced based on feature vocabulary terms and feature relationship terms. All these sentences are built on the pattern, node-relationship-node. Although this initial grammar could be employed for a basic representation of syntactical MMFGs, further extensions must obviously be constructed, particularly with respect to supporting higher level semantics and improving human readability. This initial grammar leaves the following open challenges:
  • The initial grammar does not yet represent the syntactic structure of MMFGs.
  • The initial grammar is not a context-free PS-grammar.
  • The initial grammar for MMFGs does not reflect the structural elements of MMFGs and their corresponding production rules. For example, the structural element c n (i.e., child node) should be transformed into the grammatical structure, “has a child named”, represented by several textual labels. Another example would be the spacial relationship, s r : a b o v e , which should be represented by a set of textual labels forming the phrase, “is above of”.
  • The initial grammar does not provide a semantic description of the syntactic relationships.
However, this initial definition illustrates the overall approach of representing MMFG structures with a formal language grammar with an approximation approach. Based on this, we now formally model semantic and explainable MMFGs and hence introduce several extensions to MMFGs and the corresponding grammars.

3.2. Enabling Annotation of Formal MMFG Representations

Modeling a formal representation for MMFGs, we initially apply an annotation pattern to support the linking of external semantic annotations with a special node type Annotation Anchor (aa) to represent a link to an external semantic annotation. Such annotation anchors can be linked to syntactic MMFG nodes or syntactic relationships with a special relationship type, the Annotation Relationship (ar). To support this, the representation of all MMFG relationships has also been extended by corresponding annotation anchors to allow their semantic annotation. Thus, any syntactic MMFG node or relationship can be linked with an annotation relationship to an annotation anchor. This means that any syntactic resource within an MMFG can be semantically annotated.
The formal representation of such a basic MMFG including annotation anchors is shown in Figure 4a and is further detailed in the following sections. Figure 4b shows the extension of an exemplary MMFG by Annotation Anchors (aa) and Annotation Relationships (ar). To clearly distinguish between a feature and its textual representation, the introduced variable L B L is employed, which allows the production of human-understandable textual representations of MMIR features.
With these extensions, MMFG nodes and relationships can initially be linked to elements of existing semantic representations of vocabularies, taxonomies, or ontologies, and their corresponding machine-readable representations in the Semantic Web. The grammar G M M F G can be extended to support the construction of formal language expressions including the annotation pattern by adding the elements a r and a a to the set V M M F G , so that V M M F G = F R T F V T a r , a a and by refining the production rules as follows:
P M M F G = { S M M F G F V T F R T F V T , F V T a r , F R T a r , a r a a , a a L B L }
Now, the grammar G M M F G supports the construction of additional language expressions such as: “person is-annotated-with-the-semantic-concept rdf:person. has-child is-annotated-with-the-semantic-concept rdf:related. head is-annotated-with-the-semantic-concept rdf:head”. This annotation pattern will now be further employed as a basis for the semantic representation of MMFGs.

3.3. Semantic Annotation of Formal MMFG Representations

The introduction of annotation anchors and annotation relationships is a purely syntactic extension; however, this syntactic structure must be annotated semantically. Hence, we now introduce a semantic annotation, which means that each syntactical element of an MMFG will be annotated with semantics and the purely syntactic MMFG becomes a semantically annotated MMFG—a Semantic Multimedia Feature Graph (SMMFG). For such an SMMFG, we define the following additional elements or structures:
  • A Semantic Node Representation (snr) for each MMFG node, and a Semantic Relationship Representation (srr) for each MMFG relationship. These elements are required to represent the meaning of both nodes and relationships semantically.
  • A set of Semantic Feature Vocabulary Terms  S F V T S M M F G and a set of Semantic Relationship Vocabulary Terms F R V T S M M F G . In previous work [7,43], we already defined the set F V T M M F G as the representation of all syntactic MMFG vocabulary terms (i.e., the textual labels of detected features). In an SMMFG, the semantic of each syntactic vocabulary term, f v t i F V T M M F G , is now represented by a semantic feature vocabulary term s f v t i S F V T S M M F G . Analogously, the set, S R V T S M M F G , is defined as the set of semantic representations of labels related to syntactic MMFG relationships.
  • While in an MMFG, relationships are simply represented by their relationship type (e.g., cn, sr, s), in SMMFGs, these relationship types are modeled by Semantic Relationship Vocabulary Terms (srvt), which represent the semantics of vocabulary terms describing the relationship type.
  • In an MMFG, each node and each relation is linked by an Annotation Relationship (ar) to an Annotation Anchor (aa), which now represents a Semantic Node Representation (snr) or a Semantic Relation Representation (srr). An Annotation Anchor (aa) is a URI for the node or relation it represents and used to link these MMFG nodes or relations to semantic node representations and semantic relationship representations.
  • In an SMMFG, the hasName relation links Semantic Node Representations with Semantic Feature Vocabulary Terms and Semantic Relation Representations with Semantic Relationship Vocabulary Terms. Each srr is linked via the hasDomainNode and hasRangeNode relations to the corresponding snr’s.
Figure 5a shows how Annotation Anchors can now be linked with snr and srr to semantic concepts described in the sets, S F V T S M M F G , and S R V T S M F G . As already outlined, the semantic representation of the syntax of an MMFG contains relationships itself. For example, each s r r has two relationships, hasDomainNode and hasRangeNode, which link to corresponding s n r elements. They form the 1st level semantic annotation of an MMFG and thus a Semantic Multimedia Feature Graph (SMMFG). The 2nd level semantic annotation is modeled by the elements of s f v t and s r v t , which represent the semantic information of the feature vocabulary terms. Figure 5b shows the semantic extension applied to the above example.
Based on these syntactic extensions, also the formal grammar, G M M F G can be extended, resulting in a formal grammar, G S M M F G for the representation of valid formal language expressions of SMMFG structures. G S M M F G = ( V S M M F G , T S M M F G , P S M M F G , S S M M F G = S M M F G ) , where
  • V S M M F G = V M M F G S N R S R R S F V T S M M F G S R V T S M M F G is the set of semantic representations of descriptions of MMFG nodes and relationships (see also Figure 5).
  • T S M M F G is an extension to, T M M F G , and includes additional textual descriptions of the semantic relationships: “hasName”, “hasDomainNode”, “hasRangeNode”, “describes”.
  • P S M M F G extends the production rules, P M M F G , as follows:
    P S M M F G = P M M F G { s r r s n r s n r S R V T a a s n r S F V T a a a a L B L }
The grammar G S M M F G supports the construction of additional language expressions such as, “the-semantic-concept rdf:person hasName person. the-semantic-concept rdf:head hasName head. there-is-a-semantic-relationship-between rdf:person and rdf:hat which hasDomainNode rdf:person and hasRangeNode rdf:head. the-semantic-relationship between rdf:person and rdf:head is-described-by rdf:relation and hasName rdf:related”.
Although these sentences describe further details of a SMMFG in a formally correct way and increase the readability for machines, human-readability decreases due to the mixture of syntactic, semantic, and structural labels. This means that until now human-readability depends on the selection of adequate readable and understandable textual labels. To eliminate this dependency, an automated construction of human-readable textual expressions must be achieved. Summarizing this, until now we defined a formal way to represent syntactic and semantic elements of MMFGs and SMMFGs by introducing the formal grammars, G M M F G and G S M M F G , with which not only the syntactic structure of an MMFG, but also the semantic enrichment of SMMFGs, can be represented by formal language expressions. To achieve our final goal of human-understandable i.e., natural language expressions, we introduce a PS-grammar in the next section, which transforms these machine-readable expressions into human-readable expressions.

3.4. Explainable SMMFGs

In Section 2 we introduced the concept of context-free PS-grammars. This type of grammar is typically employed for the production of natural languages consisting of e.g., Nominal Phrases (NP), Verbal Phrases (VP), nouns (N), verbs (V), adjectives (A), determiners (DET). The grammar G e n = ( V e n , T e n , P e n , S e n ) from Section 2 can be employed to produce valid English sentences (as illustrated in Figure 2).
Based on the grammars, G e n , G M M F G , and G S M M F G , we can now define such a context-free PS-grammar, G E S M M F G , which transforms any MMFG or SMMFG formally into human-readable (i.e., explainable) natural language expressions. MMFGs or SMMFGs that are extended in this way, become explainable and will be called Explainable Semantic Multimedia Feature Graphs (ESMMFG) in the remainder of this paper. Figure 6 shows the introduction of the PS-grammar and the corresponding schema for the syntactic ESMMFG representation.
Formally, we define, G E S M M F G = ( V E S M M F G , T E S M M F G , P E S M M F G , S E S M M F G ) as follows:
  • The variables, V E S M M F G are based on the variables, V e n , of the English grammar, G e n , and additionally includes the variables of the previously defined grammars:
    V E S M M F G = V e n V M M F G V S M M F G
    It thus represents the union of variables defining the English grammar (i.e., V e n = N P , V P , V , N , D E T , P R ), the syntactic elements of an MMFG (i.e., V M M F G = F R T F V T { a r , a a } ), and the semantic enrichment (i.e., V S M M F G = V M M F G S N R S R R S F V T S R V T ).
  • T E S M M F G V E S M M F G = is the set of terminal symbols and represented by the labels L B L , which can be regarded as any English word of type noun, verb, determiner, adjective, or preposition. The production of these words is based on the semantic feature and semantic relationship vocabulary. The order, in which such L B L s can be arranged to formulate valid expressions, is given by the following production rules.
  • P E S M M F G is the set of production rules and defines how the MMFG and SMMFG structures can be formally transformed into valid natural language expressions. P E S M M F G also contains the simple production rules previously defined in, P M M F G and P S M M F G ; however, the phrase structure of, P E S M M F G , leads to various additional and refining elements:
    P E S M M F G = { S E S M M F G F V T F V T F V T F R T F V T F V T a r S N R F R T a r S R R S R R S R V T S N R S N R S N R S F V T N P V P S R V T N P V P S F V T N P V P a r a a a a L B L V P V P P P P P R N P N P D E T N P R L B L N L B L V L B L D E T L B L }
  • S E S M M F G = S S M M F G = S M M F G is the starting symbol for any valid expression. This means that any natural language representation of an MMFG or SMMFG starts with the processing of the root-element; however, as the root-node of an MMFG is a node itself, G E S M M F G , can also be employed to produce expressions of subgraphs of an MMFG or SMMFG.
The application of the production rules P E S M M F G is shown in Figure 7, Figure 8 and Figure 9. For illustration purposes, some of the production rules representing mostly internal structures have been omitted for readability purposes.
These examples show that the expressiveness of ESMMFGs increased significantly with the introduction of, G E S M M F G , and that natural language sentences can now be built formally based on syntactic and semantic structures of MMFGs.
It is important to note that any natural language expression that is generated based on G E S M M F G is content-wise true (i.e., correct) as it purely represents the original multimedia features in a formal, but human-readable way. In addition, G E S M M F G , provides unlimited options for the production of valid natural language expressions due to its underlying phrase structure. This means that any multimedia feature can now be represented as a natural language, human-readable text. It is up to the application to define, which phrases should be used, which level of abstraction should be applied, or which subset of MMFG-nodes has to be selected for the natural language representation. Examples of the application of, G E S M M F G , to multimedia features of various domains is given in Section 5 (evaluation).
The formal definition of G E S M M F G furthermore guarantees that any element of an MMFG (i.e., any detected multimedia feature) can be structurally and semantically represented. It also ensures that the semantic information of any multimedia feature can be mapped to semantic systems, interpreted, and employed for inferencing and reasoning. Furthermore, any MMFG can now be represented as a syntactically correct and human readable text, which further supports automatic processing by employing a selection of numerous text-bases algorithms, e.g., for argument extraction; however, the generated text is highly dependent on the construction of phrases based on the detected (or calculated) MMIR features in the original MMFG. As this textual representation might be different depending on the MMIR processing step (e.g., the explanation of query construction, result presentation, or the ranking of an element in the result list), also different strategies for the construction of phrases need to be employed. This is reflected by introducing various subclasses for the corresponding processing steps as illustrated by the implementation samples (see Section 4).
Currently, the order of the constructed sentences is based on the order of nodes in the original MMFG data structure. This will produce good results for text-based multimedia documents, as the order of explaining texts will follow the document structure; however, for other multimedia types (e.g., images), the order of the descriptions of detected objects will be random. This can be subject to further improvements in future work. It should also be mentioned that the presented concept is a pure mathematical approach to calculate explaining texts for multimedia features without any need of machine learning tasks as, e.g., in deep LSTM language modeling [44].
In this section, we outline how an MMFG can be semantically represented and extended by a Semantic Multimedia Feature Graph. We define how explainability and transparency can be introduced to syntactic data structures based on MMIR features resulting in explainable multimedia feature graphs. To evaluate the full potential of this semantic extension, we apply this concept to the Generic Multimedia Analysis Framework (GMAF), in which MMFGs and now also SMMFGs, and ESMMFGs can be processed employing Graph Codes, which are particularly optimized for MMIR calculations. Hence, in the following subsection, we show briefly how the concept of Graph Codes can be semantically represented and extended by applying the algorithms of Graph Codes to Semantic Multimedia Feature Graphs.

3.5. Semantic Graph Codes

Graph Codes are a 2D representation of MMFGs, which are computationally optimized particularly when employed for MMIR. They are calculated from MMFGs by employing an encoding function, f e n c , which transforms an MMFG into the Graph Code structure. In the following subsections, we define how Graph Codes can be transformed into Semantic Graph Codes with corresponding operations.
Until now, the dictionary of a Graph Code (GC) is based on the feature vocabulary terms of detected features (i.e., textual labels describing detected features) of the MMFG. The dictionary of Semantic Graph Codes (SGC) is based on the semantic representation of the meaning of such feature vocabulary terms, i.e., unique labels of elements of S N R . Thus, in a S G C , unique identifiers are used to represent the dictionary. However, this representation can lead to ambiguous representation possibilities.
For example, the feature vocabulary term, Jaguar, of a Graph Code, G C J a g u a r , could be connected to the semantic node representation for an animal, but it could also be connected to the semantic node representation for a car. To solve these problems, a function, s q ( v t i ) , is introduced to determine the unambiguous semantic representation of the meaning of a feature vocabulary term. s q ( v t i ) performs a semantic query for each, v t i , to the semantic model and receives either a single unique result, s n r i S N R , representing the meaning of, v t i , or it receives a list of possible results (e.g., the Jaguar-animal or the Jaguar-car semantic nodes). To identify the correct element in this list of possible results, we apply the Graph Code Similarity algorithm to, G C J a g u a r , and each element of the result list. To do this, for each element of the result list, we perform an additional query on the semantic model (e.g., for the Jaguar-car) and represent the result as an MMFG, which is then transformed into a Graph Code, G C R e s u l t i , G C R e s u l t j . The, G C J a g u a r , will not just contain the detected feature vocabulary term Jaguar, but additional information. As Graph Codes support a high Level-Of-Detail (LOD) and are generated by recursive processing, they will also contain numerous additionally detected feature vocabulary terms, such as wheel, road, window, … or whiskers, teeth, furr, …. Our query to the semantic model also returns relationships of the Jaguar to semantic node representations of some of these detected feature vocabulary terms. Thus, if the Jaguar in our MMFG is a car, the similarity to the result of the semantic query for the Jaguar-car will be more similar to, G C J a g u a r , than the result of the Jaguar-animal query. In addition, topic modeling can be applied to further optimize the selection of semantic query results as unambiguous semantic node representations.
Summarizing this, the construction of Semantic Graph Codes utilizes a semantic query, s q ( v t i ) , for each feature vocabulary term, v t i , to identify the unambiguous semantic node representation, s n r i . To construct SGCs, the encoding function, f e n c , has to be modified to employ, s q ( v t i ) , when calculating the dictionary for SMMFGs:
f e n c ( M M F G ) = G C M M F G
f e n c ( S M M F G ) = S G C S M M F G
s n r i S M M F G , v t i M M F G : s q ( v t i ) = s n r i
In addition to the calculation of IDs, f e n c will also eliminate synonym and relationship nodes from a SGC-dictionary, as they are represented in the corresponding semantic model. So, f e n c will return an empty value for MMFG nodes of the type Synonym.
For further illustration of our example (see Figure 1c), we define the function, s q ( v t i ) , in a way that it returns the following values for the vocabulary terms of this example (see Table 2).
For the vocabulary terms “Individual” and “Human Being” representing MMFG nodes of type Synonym, the function, s q ( v t i ) , does not return any value, as these relationships are already represented by the semantic model and hence do not need to be repeated in each individual Semantic Graph Code. Applying, s q ( v t i ) , and f e n c , to our example would result in a compressed Semantic Graph Code S G C e x (see Table 3).
Further concepts of Graph Codes, such the calculation of similarity, recommendations, querying, or result presentation remain unchanged. However, it should be noted that Semantic Graph Codes lead to a further compression of the Graph Code matrix, as synonyms, or common knowledge can be removed from the Graph Code, as it is already represented in the external semantic system. Furthermore, new knowledge that exists in the external system can be employed for Semantic Graph Codes, serving as a basis for inferencing and reasoning. As the initial construction of Graph Codes is purely based on the detected feature vocabulary terms of a given multimedia object, the Graph Code vocabulary is typically very small. For Semantic Graph Codes, these feature vocabulary terms are translated into semantic IDs, or even removed, when they exist in the general semantic model of the application, which leads to a further compression, but also requires a representation of the overall semantic model (i.e., the ontology or taxonomy). This model can be represented in an SKOS, which is queried at runtime to identify such vocabulary terms. It is also possible to represent the complete semantic knowledge in the form of a Semantic Graph Code; however, this would lead to a very high number of feature vocabulary terms (i.e., any term in the ontology), and to very large Semantic Graph Codes. This approach is not recommended, as Graph Codes are optimized for indexing and not for knowledge representation.

3.6. Summary

In this section, we discussed our approach to formally defining natural language expressions from multimedia feature graphs. We showed how a context-free PS grammar can be built to generate human-readable English sentences formally and thus are able to close the gap between the technical representation of a multimedia feature and a human-understandable representation of the meaning of such a feature. We also showed how semantics can be introduced to the GMAF framework in the form of Semantic Graph Codes. Based on this modeling, we now provide further details of the implementation in the next section.

4. Implementation

Basis for the implementation of the semantic extensions and concepts discussed is the current GMAF prototype, which is written in Java and has a Java Swing UI attached. The prototype including the presented code samples of this section is available at [9], and frequently updated according to the ongoing progress of this research. Following the Factory Design Pattern [45], the GMAF has been extended to utilize external semantic representation frameworks. In this section, we discuss exemplary implementation details. First of all, in Section 4.1, the representation of SMMFGs by RDF and RDFS is given.

4.1. RDF and RDFS Representation of MMFGs

As discussed in Section 3 (modeling), we can represent any SMMFG with RDF and RDFS. The implementation of this is shown here:
<?xml version="1.0" encoding="UTF-8"?>
<rdf:RDF xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
xmlns:rdfs="http://www.w3.org/2000/01/rdf-schema#">
 
<rdfs:Class rdf:ID="MMFG">
        <rdfs:comment>MMFG Root Node Class</rdfs:comment>
</rfds:Class>
 
<rdfs:Class rdf:ID="Node">
        <rdfs:comment>MMFG Node Class</rdfs:comment>
</rfds:Class>
 
<rdfs:Class rdf:ID="Annotation_Anchor">
        <rdfs:comment>Annotation Anchor</rdfs:comment>
</rfds:Class>
 
<rdfs:Class rdf:ID="Relationship">
        <rdfs:comment>Relationship</rdfs:comment>
</rfds:Class>
 
<rdf:Property rdf:ID="Feature">
        <rdfs:comment>Feature respresenting nodes of a MMFG
        </rdfs:comment>
        <rdfs:domain rdf:resource="#MMFG"/>
        <rdfs:range rdf:resource="#Node"/>
</rdf:Property>
 
<rdf:Property rdf:ID="cn">
        <rdfs:comment>child relationship
        representing the level of detail
        </rdfs:comment>
        <rdfs:domain rdf:resource="#Node"/>
        <rdfs:range rdf:resource="#Relationship"/>
</rdf:Property>
 
<rdf:Property rdf:ID="sr">
        <rdfs:comment>Spacial relationship</rdfs:comment>
        <rdfs:domain rdf:resource="#Node"/>
        <rdfs:range rdf:resource="#Relationship"/>
</rdf:Property>
 
<rdf:Property rdf:ID="s">
        <rdfs:comment>Synonym relationship</rdfs:comment>
        <rdfs:domain rdf:resource="#Node"/>
        <rdfs:range rdf:resource="#Relationship"/>
</rdf:Property>
 
<rdf:Property rdf:ID="ar">
        <rdfs:comment>Semantic relationship</rdfs:comment>
        <rdfs:domain rdf:resource="#Relationship"/>
        <rdfs:range rdf:resource="#Annotation Anchor"/>
</rdf:Property>
 
<rdf:Property rdf:ID="aa">
        <rdfs:comment>Semantic relationship</rdfs:comment>
        <rdfs:domain rdf:resource="#Node"/>
        <rdfs:range rdf:resource="#Annotation Anchor"/>
</rdf:Property>
</rdf:RDF>
        
This implementation allows the transformation of a pure syntactical vocabulary into a semantic vocabulary, as now the syntactical model of a SMMFG can be represented by a formally well-founded semantic schema. Thus, the syntactic labels can be linked to or represented as semantic concepts. This means that any information of the above example can now be represented with RDF-statements as follows:
<rdf:RDF>
        <rdf:Statement rdf:about="mmfg:Statement">
        <rdf:subject rdf:resource="Node:Hat"/>
        <rdf:predicate rdf:resource="Relationship:above"/>
        <rdf:object rdf:resource="Node:Head"/>
</rdf:Statement>
…
</rdf:RDF>
        
Additionally, any information can now be linked to external semantic concepts:
<rdf:RDF xml:lang="en" ⋯ xmlns:mmfg="mmfg.rdf">
        …
        <mmfg:Node rdf:about="Node">
        <mmfg:name>Hat</mmfg:name>
</mmfg:Node>
             
<mmfg:Hat>
        <mmfg:sr rdf:resource="mmfg:Head" rdf:about="above"/>
        <mmfg:ar rdf:resource="mmfg:rel_3"/>
</mmfg:Hat>
             
<mmfg:Relationship>
        <mmfg:name>rel_3</mmfg:name>
        <mmfg:aa rdf:resource="mmfg:aa_3"/>
</mmfg:Relationship>
             
<mmfg:AnnotationAnchor>
        <mmfg:name>aa_3</mmfg:name>
        <mmfg:comment>
        URI of external semantic representation
        </mmfg:comment>
</mmfg:AnnotationAnchor>
…
</rdf:RDF>
        
As the MMFG is now represented in RDF and RDFS and thus aligned with semantic concepts, a further integration based on the GMAF framework can be implemented. This is described in the next subsection.

4.2. Semantic Extension of the GMAF

For each MMIR application based on the GMAF, an external semantic framework can be attached by implementing (and configuring) an adapter class, which is defined by the interface Semantic Extension (see Figure 10). The detected vocabulary term is then passed to the external semantic framework.
Whenever Semantic Graph Codes must be calculated (i.e., for indexing, retrieval, querying, filtering) in the GMAF, this extension is called to provide information from or to an external semantic representation. In our prototype, we built two implementations for this. One serves as a reference for the connection of external semantic representations and is implemented to employ the Semantic Web (class SemWebExtension), the other serves as an internal default implementation to illustrate and validate the concepts of Section 3, in particular relevance and topic calculation and inferencing (class DefaultExtension).

4.3. Semantic Representation

As external semantic representation, several tools, databases, or services can be applied. In our prototypical implementation, we chose Wikidata [46,47], as it not just serves as a basis for many other MMIR applications, it also provides a fully functioning SPARQL interface [42], which can be utilized for semantic querying. Hence, attaching Wikidata to the GMAF is straightforward as illustrated in Figure 11.
As Wikidata already supports SPARQL, GMAF-SPARQL-Queries can simple be forwarded and fused with already existing Query Graph Codes. The resulting Semantic Graph Codes are displayed in the GMAF-UI as shown in Figure 12 and can be applied to all kind of queries (including Manual Querying, Query By Example, and Query Refinement) [24] and result presentation.
As illustrated in this subsection, connecting external semantic representation systems to the GMAF is quite straightforward. In the case of external systems, the effectiveness of semantic results in terms of precision and recall experiments mainly depends on the external system. In our evaluation, we wanted to compare our internal default representation and the algorithms for topic modeling and intelligent information retrieval with these external systems, as semantic extensions are handled transparently within the GMAF (i.e., the GMAF itself does not perform any enrichment or modification). Hence, in the next subsection, we discuss the implementation of the defaults, which will also be employed later for evaluation purposes.
Summarizing this, we showed in this section that the semantic extension to the GMAF can be implemented based on an interface extension point, which generically provides access to external systems. If no such external system is available, the internal default algorithms for semantic extensions can serve as a good alternative, as Graph Codes and their metrics provide a well-defined mathematical model for intelligent information retrieval. An evaluation of our implementation including experiments is given in Section 5.

4.4. Explainability

For the implementation of Explainable Semantic Multimedia Feature Graphs, we apply the design patterns, Interpreter, Composite, and Facade. The Composite pattern is employed to recursively process ESMMFG nodes and to construct the final human-readable text. The Interpreter pattern is chosen to represent the phrase structure of the underlying grammar, where each existing element (e.g., NP, VP, FVT) is encapsulated by a subclass being responsible for the correct construction of valid expressions. Finally, the Facade pattern serves as a wrapper and provides a simple API, with which explanations of a given MMFG can be generated. A simple call of this module can be made as follows:
public class Explainer {
        public static String explain(MMFG mmfg, int levelOfDetail,
                                    int languageLevel) {
                SMMFG smmfg = new SMMFG(mmfg);
                ESMMFG esmmfg = new ESMMFG(smmfg);
                LanguageModel model =
                       LanguageModel.getInstance(languageLevel);
                String text = model.produceText(esmmfg, levelOfDetail);
                return text;
        }
        …
}
…
public abstract class LanguageModel {
        public static final int SIMPLE = 0;
        public static final int NORMAL = 1;
        public static final int COMPLEX = 2;
	 
        public static LanguageModel getInstance(int languageModel) {
                if (languageModel == SIMPLE)
                        return new SimpleLanguageModel();
                else if (languageModel == COMPLEX)
                        return new ComplexLanguageModel();
                else return new DefaultLanguageModel();
        }
        public abstract PSTree producePSTree(
                ESMMFG esmmfg, int levelOfDetail);
        public abstract PSTree produceQueryPSTree(
                ESMMFG esmmfg, int levelOfDetail);
        public abstract PSTree produceResultPSTree(
                ESMMFG esmmfg, int levelOfDetail);
        public abstract PSTree produceComparisonPSTree(
                ESMMFG esmmfg1, ESMMFG esmmfg2);
			 
        public final String produceText(ESMMFG esmmfg,
                        int levelOfDetail) {
                PSTree ps = producePSTree(esmmfg, levelOfDetail);
                return ps.createSentence();
        }
}
        
The parameter l e v e l O f D e t a i l is used to define the number of recursions that should be applied for the generation of natural language text. This directly corresponds to the level of detail of the detected MMIR features. The parameter l a n g u a g e L e v e l can define the style of the produced natural language text. Currently, there is a selection of simple, medium, and complex. Furthermore, for each step of the MMIR process, a different PS-Tree can be constructed. For example, the method produceComparisonPSTree will calculate a PS-Tree with phrases to explain why esmmfg1 has been ranked before esmmfg2 in a result list. The method produceResultPSTree calculates a PS-Tree with phrases that explain why an element is part of the result list. Further, the method produceQueryPSTree would construct a PS-tree with phrases to outline, which query has been calculated e.g., from a given keyword list or from a query by example pattern. The solution has been implemented in an extendable way, so that further subclasses of LanguageModel can be employed to refine, extend, or newly define human-understandable natural language phrases. Figure 13 shows the result for a given image with settings, l e v e l O f D e t a i l = 2 and l a n g u a g e L e v e l = S i m p l e . Figure 13 also shows the integration of explainability into the GMAF user interface.

5. Evaluation

In this section, we discuss concept and algorithm evaluation. In previous work [7], we evaluated Graph Codes retrieval against existing graph-traversal-based algorithms and were able to prove that their efficiency and effectiveness are superior to graph-based solutions. In the first part of this section, we extend this previous evaluation by experiments based on Semantic Graph Codes based on images from the Flickr30k and the DIV2K dataset. In the second part of this evaluation section, we chose to employ the text sample dataset of the 2021 TREC conference’s News evaluation [48] with 600.000 full-text articles from the Washington Post [5] to illustrate semantic retrieval and inference.

5.1. Semantic Retrieval

To determine the effectiveness of the Graph Code Algorithm, we selected five test queries from the annotations of a random set of 1000 Flickr30k images and calculated values for precision and recall for these [7]. When Graph Codes are transformed into Semantic Graph Codes, the same evaluation employs synonyms and “is-a” relations of the external (or internal) semantic model. In the following experiment, we compared our previous results with results based on external and attached internal semantic models. Table 4 shows the measured results for queries based on data, which contain r e l , relevant results, s e l , matching results. The columns, Precision P B , Recall R B , and the F1-score ( F 1 = 2 P B R B P B + R B ), F 1 B contain values for the basis experiment without any semantic extension. The values in columns P I , R I , and F 1 I are calculated for the internal semantic analysis, and the values in columns P E , R E , and F 1 E are calculated when using an external semantic model, in this case the Wikidata extension.
The findings of this experiment can be summarized as follows:
  • Any semantic enrichment increases the values for precision and recall (summarized by their F1 value) by 18% (see bold in Table 4).
  • An additional 4% increase can be achieved, when an external semantic system is connected.
These results for effectiveness are currently applied to an image dataset [49]; however, in many MMIR applications, text retrieval rates of effectiveness are also important. Hence, in the next subsection, we discuss the evaluation of our algorithms based on text datasets.

5.2. Text Retrieval and Inference

For the evaluation of text retrieval, we employed the TREC2020 dataset of the Washington Post News Archive [5,48], and followed the evaluation criteria of the TREC2021 challenge, which is based on Similarity (Top-10) and Recommendations (Top-10). For both tasks, the calculation of a semantic model of each text is required, in our case an MMFG and the corresponding SGC, including the application of a metric. As discussed in [24], similarity can be calculated by applying, M F , based on semantic vocabulary terms, and recommendations are calculated by applying, M R T . For this evaluation, we measured effectiveness based to the published results of previous years. In the first test scenario, we applied a standard “Bag-Of-Words” algorithm without any semantic enhancement. The second test-scenario then employs the full semantic analysis and features described in Section 3, but does not yet include reasoning and inferencing. This is added in the third experiment. In the fourth scenario, we attached an external framework (Wikidata) and compared all the results to the TREC reference results. We measured the values, P S i m , and R S i m , as Precision and Recall of Similarity, i.e., if the retrieved documents are in the Top-10-List, P R e c , and R R e c , as Precision and Recall for the Recommendation Top-10, and the corresponding F 1 S i m , and F 1 R e c values.
This experiment shows that the introduction of semantics to text analysis provides an increase in effectiveness of 150% (see in Table 5). The difference between an external and internal implementation exists, but is not very significant and highly dependent on the dataset and the external system. A more detailed evaluation of external systems can provide further insight; however, in the context of this paper, we are able to prove that the concepts of Section 3 are also valid for text retrieval and provide a significant increase in effectiveness.

5.3. Explainability

In addition, to these results in the area of efficiency and effectiveness, Explainability has also been evaluated. For this evaluation, we generated various texts describing a number of typical MMIR asset types and their corresponding features. An exemplary text is shown in Figure 13.
Thus, in this evaluation section, we were able to show that semantic enrichment of MMIR applications provides a significant increase in retrieval effectiveness. We also showed that explainability provides a huge potential, in particular, in combination with Semantic Graph Codes, topic modeling, and intelligent information retrieval, improvements of up to 150% can be achieved. Finally, in the next section, we summarize our results.

6. Conclusions and Future Work

In this paper, we discussed concepts and algorithms for narrowing the semantic gap [1], i.e., the gap between detected features in multimedia objects and the meaning of these features. We introduced a well-defined semantic representation of the MMFG and enhanced the concept of Graph Codes to fully support semantic querying, filtering, reasoning, and inferencing. Both external and internal implementations of the semantic model can be attached to the GMAF, which opens a varied range of extensions to existing MMIR applications and standards. In addition, we showed that our internal default algorithms are also highly effective, and can serve as a solid basis for further implementations. In our evaluation of both image- and text-based datasets, the results of our Semantic Graph Code algorithms and the corresponding concepts give evidence for our modeling approach. Hence, Semantic Graph Codes are an effective and efficient foundation for automated reasoning and inferencing for any MMIR application.
However, there are still some remaining challenges: the presentation of inferencing conflicts to the user, the implementation of additional integration with existing SKOS systems, further implementation of language models, and the evaluation of our implementation with further datasets. These challenges will be part of our ongoing and future work, which is also frequently updated in our GitHub repository [9].

Author Contributions

Conceptualization and methodology: S.W. and M.H. Software, validation, formal analysis, investigation, resources, data curation, and writing: S.W. Review, editing, and supervision: P.M.K. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in [9].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kwasnicka, H.; Jain, L.C. Bridging the Semantic Gap in Image and Video Analysis; Springer: Berlin, Germany, 2018; ISBN 978-3-319-73891-8. [Google Scholar]
  2. Clement, J. Social Media—Statistics & Facts. 2020. Available online: https://www.statista.com/topics/1164/social-networks/ (accessed on 23 August 2020).
  3. Sony Electronics Inc. a7R IV 35 mm Full-Frame Camera with 61.0 MP. 2021. Available online: https://www.sony.com/electronics/ interchangeable-lens-cameras/ilce-7rm4 (accessed on 22 November 2021).
  4. Xiaomi. Redmi Note 10 Pro—The 108MP Voyager. 2021. Available online: https://www.mi.com/global/product/redmi-note-10-pro/overview (accessed on 22 November 2021).
  5. The Washington Post. Washington Post Archives. 2021. Available online: https://www.washingtonpost.com (accessed on 22 November 2021).
  6. Domingue, J. Introduction to the Semantic Web Technologies. In Handbook of Semantic Web Technologies; Springer: Berlin/Heidelberg, Germany, 2011; pp. 1–41. [Google Scholar]
  7. Wagenpfeil, S.; Engel, F.; Kevitt, P.M.; Hemmje, M. AI-Based Semantic Multimedia Indexing and Retrieval for Social Media on Smartphones. Information 2021, 12, 43. [Google Scholar] [CrossRef]
  8. Wagenpfeil, S. GMAF Prototype. 2020. Available online: http://diss.step2e.de:8080/GMAFWeb/ (accessed on 23 August 2020).
  9. Wagenpfeil, S.; Engel, F.; Kevitt, P.M.; Hemmje, M. Github Repository of GMAF and MMFVG. 2021. Available online: https://github.com/stefanwagenpfeil/GMAF/ (accessed on 28 November 2021).
  10. Wagenpfeil, S.; Hemmje, M. Towards AI-bases Semantic Multimedia Indexing and Retrieval for Social Media on Smartphones. In Proceedings of the 2020 15th International Workshop on Semantic and Social Media Adaptation and Personalization, Zakynthos, Greece, 29–30 October 2020; pp. 1–9. [Google Scholar] [CrossRef]
  11. Beyerer, J.; Richter, M.; Nagel, M. Pattern Recognition—Introduction; Walter de Gruyter GmbH & Co KG: Berlin, Germany, 2017; ISBN 978-3-110-53794-9. [Google Scholar]
  12. Manotumruksa, J.; Macdonald, C.; Ounis, I. A contextual attention recurrent architecture for context-aware venue recommendation. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 555–564. [Google Scholar]
  13. Leveling, J. Interpretation of coordinations, compound generation, and result fusion for query variants. In Proceedings of the 36th International ACM SIGIR Conference on Research and Development in Information Retrieval, Dublin, Ireland, 28 July 2013; pp. 805–808. [Google Scholar]
  14. Bhute, A.N.; Meshram, B.B.; Bhute, H. Multimedia Indexing and Retrieval Techniques: A Review. Int. J. Comput. Appl. 2012, 58, 35–42. [Google Scholar]
  15. Lew, M.S.; Sebe, N.; Djeraba, C.; Jain, R. Content-based multimedia information retrieval: State of the art and challenges. ACM Trans. Multimed. Comput. Commun. Appl. 2006, 2, 1–19. [Google Scholar] [CrossRef]
  16. Hernández-Gracidas, C.; Juárez, A.; Sucar, L.E.; Montes-y-Gómez, M.; Villaseñor, L. Data Fusion and Label Weighting for Image Retrieval Based on Spatio-Conceptual Information. Adapt. Pers. Fusion Heterog. Inf. 2010, 76–79. [Google Scholar]
  17. Dufour, R.; Esteve, Y.; Deléglise, P.; Béchet, F. Local and global models for spontaneous speech segment detection and characterization. In Proceedings of the Workshop on Automatic Speech Recognition and Understanding, Merano, Italy, 13–17 December 2009; pp. 130–135. [Google Scholar]
  18. Subrahmanian, V. Principles of Multimedia Database Systems; Morgan Kaufmann: San Francisco, CA, USA, 1998; ISBN 978-1-558-60466-7. [Google Scholar]
  19. Chang, S.F.; Sikora, T.; Purl, A. Overview of the MPEG-7 standard. IEEE Trans. Circuits Syst. Video Technol. 2001, 11, 688–695. [Google Scholar] [CrossRef]
  20. FFMpeg.org. ffmpeg Documentation. 2020. Available online: http://ffmpeg.org (accessed on 23 August 2020).
  21. Mu, X. Content-based video retrieval: Does video’s semantic visual feature matter? In Proceedings of the 29th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seattle, WA, USA, 6–11 August 2006; pp. 679–680. [Google Scholar]
  22. Wagenpfeil, S.; Engel, F.; McKevitt, P.; Hemmje, M. Graph Codes—2D projections of Multimedia Feature Graphs for Fast and Effective Retrieval. 2021. Available online: http://164.177.153.44/pubs/icivr2021.pdf. (accessed on 26 August 2020).
  23. yWorks GmbH. yEd Graph Editor. 2020. Available online: https://www.yworks.com/products/yed (accessed on 23 August 2020).
  24. Wagenpfeil, S.; Engel, F.; McKevitt, P.; Hemmje, M.L. Semantic Query Construction and Result Representation based on Graph Codes. 2021. Available online: http://ceur-ws.org/Vol-2863/paper-06 (accessed on 28 July 2021).
  25. Kapetanios, E.; Groenewoud, P. Query Construction through Meaningful Suggestions of Terms. In Flexible Query Answering Systems; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2002; pp. 226–239. ISBN 978-3-540-36109-1. [Google Scholar]
  26. Kapetanios, E.; Baer, D.; Groenewoud, P.; Mueller, P. The design and implementation of a meaning driven data query language. In Proceedings of the 14th International Conference on Scientific and Statistical Database Management, Aalborg, Denmark, 30 June–2 July 2014; pp. 20–23. [Google Scholar]
  27. Shaik, S.; Kanakam, P.; Hussain, S.M.; Suryanarayana, D. Transforming Natural Language Query to SPARQL for Semantic Information Retrieval. Int. J. Eng. Trends Technol. 2016, 41, 347–350. [Google Scholar] [CrossRef]
  28. Jung, H.; Kim, W. Automated conversion from natural language query to SPARQL query. J. Intell. Inf. Syst. 2020, 1–20. [Google Scholar] [CrossRef]
  29. W3C.org. W3C Semantic Web Activity. 2020. Available online: http://w3.org/2001/sw (accessed on 23 August 2020).
  30. Wong, R.; Leung, C.H. Automatic Semantic Annotation of Real-World Web Images. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1933–1944. [Google Scholar] [CrossRef] [PubMed]
  31. Ni, J.; Qian, X.; Li, Q.; Xu, X. Research on Semantic Annotation Based Image Fusion Algorithm. Res. Semant. Annot. Based Image Fusion Algorithm 2017, 945–948. [Google Scholar]
  32. Asim, M.N.; Wasim, M.; Khan, M.U.G. The Use of Ontology in Retrieval: A Study on Textual. IEEE Access 2019, 7, 21662–21686. [Google Scholar] [CrossRef]
  33. Beierle, C.; Kern-Isberner, G. Methoden Wissensbasierter Systeme–Grundlagen; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 2019; ISBN 978-3-658-27084-1. [Google Scholar]
  34. Aho, A. Compilerbau; Oldenbourg Wissenschaftsverlag: Munich, Germany, 1999; ISBN 9783486252941. [Google Scholar]
  35. Hauser, R. Principles of Computer Linguistics; Springer Publishing: Berlin/Heidelberg, Germany, 2010; ISBN 3-540-67187-0. [Google Scholar]
  36. Evans, J. Thinking and Reasoning: A Very Short Introduction; Oxford University Press: Oxford, UK, 2017; ISBN 0198787251. [Google Scholar]
  37. W3C. SKOS Simple Knowledge Organisation System. 2021. Available online: https://www.w3.org/2004/02/skos/ (accessed on 28 May 2021).
  38. Bochman, A. Nonmonotonic Reasoning. 2007. Available online: https://www.sciencedirect.com/science/article/pii/ S1874585707800124 (accessed on 24 October 2021).
  39. Das, A. Knowledge Representation. 2003. Available online: https://www.sciencedirect.com/science/article/pii/B0122272404001027 (accessed on 28 October 2021).
  40. Reiter, R. A Logic for Default Reasoning. 1980. Available online: https://www.sciencedirect.com/science/article/pii/ 0004370280900144 (accessed on 20 September 2021).
  41. Poole, D. A Logical Framework for Default Reasoning. 1988. Available online: https://www.sciencedirect.com/science/article/pii/ 000437028890077X (accessed on 14 July 2021).
  42. W3C.org. SPARQL Query Language for RDF. 2013. Available online: https://www.w3.org/TR/sparql11-overview/ (accessed on 23 August 2020).
  43. Wagenpfeil, S.; Vu, B.; Kevitt, P.; Hemmje, M. Fast and Effective Retrieval for Large Multimedia Collections. Big Data Cogn. Comput. 2021, 5, 33. [Google Scholar] [CrossRef]
  44. LSTM Model. Long Short-Term Memory. 2021. Available online: https://en.wikipedia.org/wiki/Long_short-term_memory (accessed on 28 November 2021).
  45. Gamma, E.; Helm, R.; Johnson, R.; Vlissides, J.; Patterns, D. Design Patterns—Elements of Reusable Object Oriented Software; Addison Wesley: Boston, MA, USA, 1994; ISBN 0-201-63361-2. [Google Scholar]
  46. Wikidata.com. Wikidata—The Free Knowledgebase. 2021. Available online: https://www.wikidata.org/wiki/Wikidata:Main_Page (accessed on 24 October 2021).
  47. Vrandečić, D.; Krötzsch, M. Wikidata: A free collaborative knowledgebase. Commun. ACM 2014, 57, 78–85. [Google Scholar] [CrossRef]
  48. Text Retrieval Conference, Datasets. 2021. Available online: https://trec.nist.gov/data.html (accessed on 14 September 2021).
  49. Young, P.; Lai, A.; Hodosh, M.; Hockenmaier, J. From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. Trans. Assoc. Comput. Linguist. 2014, 2, 67–78. [Google Scholar] [CrossRef]
Figure 1. Exemplary M M F G and its related representations.
Figure 1. Exemplary M M F G and its related representations.
Information 12 00502 g001
Figure 2. Production rules for the example “The hat is above the head” visualized as Phrase-Structure tree (PS-tree).
Figure 2. Production rules for the example “The hat is above the head” visualized as Phrase-Structure tree (PS-tree).
Information 12 00502 g002
Figure 3. MMFG overview: (a) formal model of the syntactic MMFG representation; (b) formal syntactic schema of the MMFG representation of M M F G e x ; (c) additional starting symbol S M M F G .
Figure 3. MMFG overview: (a) formal model of the syntactic MMFG representation; (b) formal syntactic schema of the MMFG representation of M M F G e x ; (c) additional starting symbol S M M F G .
Information 12 00502 g003
Figure 4. MMFG overview including Annotation Anchors (aa) and Annotation Relationships (ar): (a) formal model; (b) formal syntactic schema of the MMFG representation M M F G e x .
Figure 4. MMFG overview including Annotation Anchors (aa) and Annotation Relationships (ar): (a) formal model; (b) formal syntactic schema of the MMFG representation M M F G e x .
Information 12 00502 g004
Figure 5. SMMFG overview: (a) formal model of the syntactic SMMFG; (b) formal syntactic schema of the SMMFG representation S M M F G E x .
Figure 5. SMMFG overview: (a) formal model of the syntactic SMMFG; (b) formal syntactic schema of the SMMFG representation S M M F G E x .
Information 12 00502 g005
Figure 6. ESMMFG overview: (a) formal model of the syntactic ESMMFG representation; (b) formal schema of the syntactic ESMMFG representation.
Figure 6. ESMMFG overview: (a) formal model of the syntactic ESMMFG representation; (b) formal schema of the syntactic ESMMFG representation.
Information 12 00502 g006
Figure 7. PS-tree with production rules, P E S M M F G , of, G E S M M F G , applied to M M F G e x .
Figure 7. PS-tree with production rules, P E S M M F G , of, G E S M M F G , applied to M M F G e x .
Information 12 00502 g007
Figure 8. PS-tree with production rules, P E S M M F G , of, G E S M M F G , applied to S M M F G e x including annotation anchors.
Figure 8. PS-tree with production rules, P E S M M F G , of, G E S M M F G , applied to S M M F G e x including annotation anchors.
Information 12 00502 g008
Figure 9. Snippet of the PS-tree with production rules, P E S M M F G , of G E S M M F G , applied to S M M F G e x , including semantic feature vocabulary terms.
Figure 9. Snippet of the PS-tree with production rules, P E S M M F G , of G E S M M F G , applied to S M M F G e x , including semantic feature vocabulary terms.
Information 12 00502 g009
Figure 10. Factory implementation to attach external semantic frameworks.
Figure 10. Factory implementation to attach external semantic frameworks.
Information 12 00502 g010
Figure 11. Querying an external semantic system for unique IDs with SPARQL.
Figure 11. Querying an external semantic system for unique IDs with SPARQL.
Information 12 00502 g011
Figure 12. GMAF-UI displaying image: (a) Graph Codes; (b) the corresponding Semantic Graph Code.
Figure 12. GMAF-UI displaying image: (a) Graph Codes; (b) the corresponding Semantic Graph Code.
Information 12 00502 g012
Figure 13. GMAF-UI displaying a human-readable explanation of an image.
Figure 13. GMAF-UI displaying a human-readable explanation of an image.
Information 12 00502 g013
Table 1. PS-grammar (Phrase Structure grammar) hierarchy of formal languages according to [35].
Table 1. PS-grammar (Phrase Structure grammar) hierarchy of formal languages according to [35].
RestrictionPSG ClassLanguage ClassComplexity
type 3regular PSGregular languagelinear
type 2context free PSGcontext free languagepolynomial
type 1context sensitive PSGcontext sensitive languageexponential
type 0unrestricted PSGrecursive enumerable languageunsolvable
Table 2. Exemplary function, s q ( v t i ) , and values for G C e x .
Table 2. Exemplary function, s q ( v t i ) , and values for G C e x .
vt i sq ( vt i )
Person101
Head102
Hat103
above104
Individual-
Human Being-
Table 3. Table representation of, S G C e x 2 , including external unique identifiers as row and column descriptors.
Table 3. Table representation of, S G C e x 2 , including external unique identifiers as row and column descriptors.
SGC ex 101102103104
1011300
1020130
1030025
1040006
Table 4. Effectiveness based on the Flickr30k dataset.
Table 4. Effectiveness based on the Flickr30k dataset.
DogManGolfGuitarBicycleAvg
Basic Experiment
t p B 18811951754
t n B 28521324
P B 0.980.530.710.560.690.71
R B 0.910.250.450.890.940.69
F 1 B 0.940.350.550.690.800.67
Attachment of External Framework
t p E 18830962663
t n E 21042619
P E 0.981.510.850.860.801.00
R E 0.910.670.541.361.100.92
F 1 E 0.940.920.661.060.930.90
Table 5. Result of text analysis based on the Washington Post News dataset.
Table 5. Result of text analysis based on the Washington Post News dataset.
Experiment P Sim R Sim F 1 Sim P Rec R Rec F 1 Rec
Bag-Of-Words0.30.30.30.20.20.2
Internal Impl.0.70.70.70.60.60.6
Inferencing0.80.80.80.80.80.8
External Impl.0.70.80.70.50.50.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wagenpfeil, S.; Mc Kevitt, P.; Hemmje, M. Towards Automated Semantic Explainability of Multimedia Feature Graphs. Information 2021, 12, 502. https://doi.org/10.3390/info12120502

AMA Style

Wagenpfeil S, Mc Kevitt P, Hemmje M. Towards Automated Semantic Explainability of Multimedia Feature Graphs. Information. 2021; 12(12):502. https://doi.org/10.3390/info12120502

Chicago/Turabian Style

Wagenpfeil, Stefan, Paul Mc Kevitt, and Matthias Hemmje. 2021. "Towards Automated Semantic Explainability of Multimedia Feature Graphs" Information 12, no. 12: 502. https://doi.org/10.3390/info12120502

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop