Next Article in Journal
Conductivity Enhancement of Graphene and Graphene Derivatives by Silver Nanoparticles
Next Article in Special Issue
Improving Abstractive Dialogue Summarization Using Keyword Extraction
Previous Article in Journal
Analysis and Test of the Tillage Layer Roll-Type Residual Film Recovery Mechanism
Previous Article in Special Issue
“Standard Text” Relational Classification Model Based on Concatenated Word Vector Attention and Feature Concatenation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sentiment Analysis of Semantically Interoperable Social Media Platforms Using Computational Intelligence Techniques

1
Department of Networks and Communications Engineering, College of Computer Science and Information Systems, Najran University, Najran 61441, Saudi Arabia
2
Department of Data Science, School of Science, Engineering and Environment, University of Salford, Salford M5 4WT, UK
3
Department of Electrical and Computer Engineering, Lebanese American University, Byblos 13-5053, Lebanon
4
Department of Computer Science, College of Computer Science and Information Systems, Najran University, Najran 61441, Saudi Arabia
5
Department of Computer Science and Information, College of Science, Majmaah Univesity, Al Majma’ah 11952, Saudi Arabia
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7599; https://doi.org/10.3390/app13137599
Submission received: 15 April 2023 / Revised: 12 June 2023 / Accepted: 13 June 2023 / Published: 27 June 2023
(This article belongs to the Special Issue Application of Machine Learning in Text Mining)

Abstract

:
Competitive intelligence in social media analytics has significantly influenced behavioral finance worldwide in recent years; it is continuously emerging with a high growth rate of unpredicted variables per week. Several surveys in this large field have proved how social media involvement has made a trackless network using machine learning techniques through web applications and Android modes using interoperability. This article proposes an improved social media sentiment analytics technique to predict the individual state of mind of social media users and the ability of users to resist profound effects. The proposed estimation function tracks the counts of the aversion and satisfaction levels of each inter- and intra-linked expression. It tracks down more than one ontologically linked activity from different social media platforms with a high average success rate of 99.71%. The accuracy of the proposed solution is 97% satisfactory, which could be effectively considered in various industrial solutions such as emo-robot building, patient analysis and activity tracking, elderly care, and so on.

1. Introduction

Machine learning not only influences the social media market but, simultaneously, it is highly capable of tracking the so-called unpredictable real-time matrices of growth, needs, results, and features. Machine learning is one of the most powerful tools to control the human mind′s transition through machine interpretation [1,2]. Individuals as well as sets of individuals using social media platforms are targeted by business media, multistoried companies, product sellers, and influencers. Facts and figures show that more than 56% of eighth-graders are additionally unhappy because their weekly social media involvement is over 10 h. Spending over 3 h per day on social media presents a high risk to the mental health of adolescents. According to the National Center for Health Research, 32% and 13% of children aged 12–17 suffer from depression and anxiety because of the inappropriate use of social media, respectively. Furthermore, 25% of 25-year-old adults have a mental illness, and they belong to the highest usage group [3]. Apart from individual consumption, teenagers are experiencing crimes related to online harassment and cyberbullying. However, the numbers are only based on reported and known occurrences. Tracking particular algorithms behind the consumer attraction strategies of each social media platform is inconceivable and unrealistic [4]. Still, in recent years, researchers have identified possible measurement factors to monitor the mental health of social media users using machine learning techniques [5,6]. Existing machine-learning-based solutions can predict a product value after studying historical sales-related data. For example, it can indicate the increase and decrease in the share market curve, and it can even measure the happiness index using behavioral analytics [7,8,9].

1.1. Motivation

To analyze human emotion through a smart device or machine is challenging in reality. The method of measuring a person’s feelings using machine learning is known as sentiment analysis [6]. There are several machine learning algorithms and classification models in development in data science for the application of the sentiment analysis method [7,8]. This subsection focuses on some of the relative research gaps and challenges we encountered in the field of sentiment analysis using machine learning approaches.
We found three fundamental difficulties in the sentiment analysis of an individual social media user, which are as follows:
  • A person is a user of different social media platforms;
  • Every social media platform offers diverse user access possibilities;
  • Different social media platforms have distinct emotional expression structures.
These three real-time complications lead to a lack of consistent interpretation among various social media accounts handled by a single person. Therefore, we incorporated social media accounts containing ontology with the help of data science to improve the sentiment analysis algorithm by making it more pertinent.

1.2. Contributions

Considering the survey reports and existing machine learning techniques, we propose an improved approach integrating the following features:
  • We propose an improved social media sentiment analytics technique through pleasure factor measuring to predict the individual state of mind of social media users and the ability of users to resist profound effects.
  • We suggest an integrated sentiment analysis machine learning model to compute the next best solution to a significant emotion value.
  • We consider dividing and seaming for two different but complementary estimation sentimental states, namely, happiness level counting and neutral expression level counting of each inter- and intra-linked expression.
  • We consider the integration of data through semantic-level interoperability from heterogeneous consequences.
  • With a high success rate, the proposed algorithm can track down more than one semantically linked activity from diverse social media platforms via a single user ID.
The rest of this paper is organized as follows: The related work section consists of existing research and describes the state-of-the-art in machine learning, sentiment analysis, information systems, and social media analytics. Section 2 defines the problem statement behind our research goal. Section 3 illustrates the proposed solution with algorithm selection and methodologies. Next, Section 4 displays the outcomes of our proposed solution. Finally, Section 5 concludes the paper with a description of the future aspects of the research area.

2. Literature Review

This section highlights the corresponding domains and acquaintances among the buzzwords in the field of information systems. Information system itself is an autonomous research domain in this era of Industry 4.0. This field of research deals with information gathering, big data analysis, and possible utilization to transfer it into knowledge.

2.1. Sentiment Analysis and Social Media

Social media platforms have been working as marketing and interpreting platforms for the last few years. Social media could manipulate, amplify, and reinforce one′s personal and social rank. Individuals and organizations mostly connect to a social media platform based on their choice of exposure [10]. Though very few platforms charge a certain subscription amount, most of them are free of cost if you have a connecting device like a smartphone, desktop, or laptop with internet connectivity. Hence, engagement in a specific social media platform is directly related to human interest. Feedback procedures vary from one platform to another in the form of like, comment, post, tag, send, share your feelings and moods, emoji, hashtag, and many more. Social media engagement and pleasure are directly influenced by human emotion, and the feedback through these platforms also affects human sentiment [10,11]. Most likely, the inventors of social media invented these platforms to reach out to specific communities in the shortest and fastest time. However, it can help product marketing through consequence penetration, consumer attractions, live promos, and easy money transaction in sales [11]. Additionally, positive ratings could be key growth factors in terms of customer support, bargaining, immediate response, and feedback. Similarly, these platforms help to monitor customer loyalty and public relation [12]. Besides this business growth, individual interests are also encouraged and entertained by millions of connected people worldwide through these social networks [13]. Whilst traders analyze customer feedback and sentiments using machine learning as their brand momentum drives, we intend to explore the distinct proposition of an individual social media handler through machine learning.

2.2. Information Systems and Machine Learning

In information systems, artificial intelligence applications and systems can perform their defined task without any human interaction. Furthermore, without being explicitly programmed, a computer could be equipped with machine learning with the capacity to comprehend real-time inputs [14]. Machine learning is a kind of artificial intelligence that can program a computer to adjust when exposed to exotic data. There are three types of machine learning: supervised, unsupervised, and reinforcement to make a computer intelligent. Additionally, supervised learning could be applied to two categories: regression problems and classification problems. In comparison, unsupervised learning deals with clustering and dimensionality reduction [15].

2.3. Machine Learning Models and Sentiment Analysis

Supervised ML holds two segments of attributes: independent features and dependent or derived features [16]. Whenever a continuous variable in supervised learning represents the output, it is a regression problem. On the other hand, if the ML model has a fixed number of categories as the output, it becomes a classification problem. If there are two categories as the output for a classification problem, then it is called a binomial classification problem. If the model has more than two categories as the output, it is called a multi-class classification problem. Similarly, a model could be learned even if the dependent variable is absent in the available dataset. This kind of machine learning procedure is known as unsupervised machine learning. The train dataset could be comprised in two ways into an unsupervised ML model: clustering and dimensionality reduction [17]. The clustering procedure makes a group concerning similar conditions or likenesses factors if there is no dependent variable in the available dataset. Dimensionality reduction converts higher dimensional data to its lower dimensions for an implemented ML model. Linear regression, logistic regression, decision tree, AdaBoost, random forest, xgboost, Naive Bayes, SVM, and KNN are popular supervised algorithms in information systems. Similarly, K-Means, DB Scan, Hierercial clustering, K-nearest neighbor clustering, PCA, and LDA are a few unsupervised learning algorithms [18]. Based on some real-time situations, the parameter settings could be changed for a designed ML model to achieve a better performance [16,17,18].

2.4. Natural Language Processing

Several supervised and unsupervised models are available for natural language processing problems in data science [19]. Support vector machine (SVM), logistic regression (LR), gradient boosting (GB), and neural networks (NNs) are a few known machine learning (ML) algorithms used for both classification and regression challenges [20,21]. An ML model′s accuracy and success rate determine the model selection based on the target dataset. Basic feature engineering reflects the basic modeling of tabular format data (basic features) for the training and validation datasets. Data extend further for logistic regression. The datasets can reach for extreme gradient boosting (XGB) in support of pushing the computational resource limit.
We could conduct an approximate string matching operation to improve the accuracy of the discussed machine learning algorithm altogether [21,22]. The string matching operation in data science is known as Fuzzy features inclusion. Fuzzy features contribute a positive push to the comprehensive computational process and improve the accuracy of the absolute model. The most traditional way to handle text data has four pathways: hashing of words, count vectorization, term frequency–inverse document frequency (TF-IDF), and singular-value decomposition (SVD).

3. Proposed Solution

We intended to design an ML model to perform NLP, an improved social media sentiment analytics technique to predict the individual state of mind. The proposed estimation function tracks the counts of the aversion and satisfaction levels of each inter- and intra-linked expression through semantic level interoperability. It collects data from several social media posts, feedback, and reactions handled and expressed by a single user through the self-learning pipelining mode. Therefore, we aim to traverse peculiar sentiment reflections by a single user into different social media handles.
This section discusses how the pipeline of the proposed model helps to resolve existing difficulties. As shown in Figure 1, the proposed ML model first semantically connects multiple (up to four) social media handles of a single person. Then, it collects data from semantically related classes into a single database. After that, the emotion score is counted for each input string with the help of an advanced NLP (natural language processing) strategy [19,20,21,22]. In this way, the model helps continuously monitor a person’s emotional outbursts from a neutral to a happy state and vice versa. The semantically interoperable property makes the model more competent [23,24,25,26]. It helps to analyze the overall activity of a person in their several social channels. It helps to estimate the satisfaction levels of each inter- and intra-linked expression.
The following steps to build such a model are shown in Algorithm 1. And the descriptions of each possible effort toward making this ontology are also defined in this section.
Algorithm 1: Establishment of semantic level interoperability through ontology design.
Input: Data from different social media platforms of a person
Output: Semantically interoperable data
Initialization and declaration:
     i. L1, L2: Linear expressions;
    ii. T1, T2: Subjective terms;
   iii. Y1, Y2: Variables;
   iv. xsd: decimal, integer, long, short, byte;
    v. ObjectProperty: OP
   vi. InverseObjectProperty: IOP
  vii. DataIntersectionOf: DIO
Start
   Step 1: Entity selection:
   (DataComparison(Arguments(Y1, Y2) comprel(Y1, Y2)))xsd
   Step 2: Class identification:
      i. ClassAxiom:= SubClassOf | EquivalentClasses | DisjointClasses | DisjointUnion
     ii. ClassAssertion (DataHasValue (Y1R1”^^xsd:decimal) Y2)
    iii. ClassAssertion(DataHasValue(Y1R2”^^xsd:decimal) Y2)
    iv. EquivalentClasses(NormalSubstance DataAllValuesFrom(Y1 DataComparison(Arguments(T1 Y1) leq(T2 Y2))))
   Step 3: Object properties define:
      i. IOP:= ′ObjectInverseOf′ ′(′OP′)′
     ii. ObjectPropertyExpression:= OP | IOP
    iii. ′DataComparisonDefinition′ ′(′axiomAnnotations IRI DataRange′)′
    iv. ObjectAllValuesFrom: = ′ObjectAllValuesFrom′ ′(′OP/IOP ClassExpression′)′
   Step 4: Data properties define:
      i. Variable: = NCName
     ii. Rational: = Integer/NonZeroInteger
    iii. Arguments: = ′Arguments′ ′(′NCName{NCName}′)′
    iv. DIO: = ′DIO ′(′DataRange DataRange{DataRange}′)′; DIO (xsd:nonNegativeInteger xsd:nonPositiveInteger)
   Step 5: Annotation properties define:
     i. Axiom: = Declaration | ClassAxiom | ObjectPropertyAxiom | DataPropertyAxiom | DatatypeDefinition | HasKey |
        Assertion | AnnotationAxiom
    ii. axiomAnnotations: = {Annotation}
        *Annotation: ObjectIntersectionOf, ObjectUnionOf, ObjectComplementOf, ObjectOneOf, ObjectSomeValuesFrom,
        ObjectAllValuesFrom, ObjectHasValue, ObjectHasSelf, ObjectMinCardinality, ObjectMaxCardinality,
        ObjectExactCardinality, DataSomeValuesFrom, DataAllValuesFrom, DataHasValue, DataMinCardinality,
        DataMaxCardinality, DataExactCardinality
   Step 6: Individual definition establishment
     i. ObjectHasValue: = ′ObjectHasValue′ ′(′ObjectPropertyExpression Individual′)′
    ii. ObjectHasSelf: = ′ObjectHasSelf′ ′(′ObjectPropertyExpression′)′
   iii. ObjectMinCardinality: = ′ObjectMinCardinality′ ′(′nonNegativeInteger ObjectPropertyExpression [ClassExpression]′)′
   iv. ObjectMaxCardinality: = ′ObjectMaxCardinality′ ′(′nonNegativeInteger ObjectPropertyExpression [ClassExpression]′)′
   Step 7: Define annotations;
   Step 8: Ontology documentation:
   ontologyDocument: = {prefixDeclaration} Ontology
   prefixDeclaration: = ′Prefix′ ′(′prefixName ′=′ fullIRI′)′
   Ontology: = ′Ontology′ ′(′[ontologyIRI [versionIRI]] directlyImportsDocuments; ontologyAnnotations axioms ′)′
   ontologyIRI: = IRI
   versionIRI: = IRI
   directlyImportsDocuments: = {′Import′ ′(′IRI′)′}
   ontologyAnnotations: = {Annotation}
   axioms: = {Axiom}
End

3.1. Semantic Interoperability

We choose web ontology language (OWL) [25] to define possible semantic interoperability among the social media handlings by a single user id. It primarily helps to scale down the meaning of word expressions from comments, feedback, posts, and reactions on different social media platforms. It defines the equivalency and domain properties of distinct objects under similar classes. The consequence of the proposed ontology design has been described through Algo-rithm1, which carries the following steps:

3.1.1. Entity Selection

The entity selection defines the closure property of a set of declarations. Each IRI (internationalized resource identifier) utilized in an OWL entity is an axiom property that speaks about the class identification, object property, and datatype declaration. For the first-time entity declaration, it has to be defined which class data belong to whether they are individual or in combination. Most importantly, it enters the ontology vocabulary. The utilization of an existing entity could be possible simply by selecting or re-using it.
Declaration: = ′Declaration′ ′(′axiomAnnotations Entity′)′.
Entity: = ′Class′ ′(′Class′)′|′Datatype′ ′(′Datatype′)′|′ObjectProperty′ ′(′ObjectProperty′)′|′DataProperty′ ′(′DataProperty′)′|′AnnotationProperty′ ′(′AnnotationProperty′)′|′NamedIndividual′ ′(′NamedIndividual′)′.

3.1.2. Class Identification

Class and property construction could be complex as it represents a set of formal and individual special conditions. Instances of a particular class expression satisfy the requirements of a unique object property. ′ClassExpression′ represents class expressions in OWL, used to construct class identity. There are few inbuild or predefined class identifiers, like ′ObjectIntersectionOf′, ′ObjectUnion-Of′, ′ObjectComplementOf′, and ′ObjectOneOf′.

3.1.3. Object Properties Define

Object property expressions are a method to restrict the class expressions in the OWL platform. Some predefined object properties help express an object regionally. The ′ObjectSomeValuesFrom′ helps quantify an object property and holds at least one individual connection to the primary class property. In contrast, a universal quantification is conducted by the ′ObjectAllValuesFrom′ class expression. It joins the individuals linked via an object′s belongings countenance exclusively to instances of a disseminated class expression. An object property expression connects many distinct variables to a particular individual under the class expression ′ObjectHasValue.′ Ultimately, the ′ObjectHasSelf′ class declaration holds the individuals linked by an object property indication to themselves.

3.1.4. Data Properties Define

Restrictions on data property expressions help to define unique class expressions in OWL. Similar to limiting object property expressions, the leading contrast is that the existential and omnipresent quantification declarations permit n-ary data coverage. The ′DataSomeValuesFrom′ class declaration qualifies for a specified existential quantification from a checklist of facts-related expressions. It includes the individuals united via the data property countenances to the smallest literal within a given data field. There are two more data-restricted class expressions, like ′DataAllValuesFrom′ and ′DataHasValue.′ Data property cardinality restrictions also could be conducted using class expressions like ′DataExactCardinality′, ′DataMaxCardinality′, and ′DataMinCardi-nality′. Data types with range and domain are also specified during object property definition, like DataPropertyDomain: = ′DataPropertyDomain′ ′(′axi-omAnnotations DataPropertyExpression ClassExpression′)′, to explain the domain of a distinct data variable. Similarly, DataPropertyRange: = ′DataPropertyRange′ ′(′axiomAnnotations DataPropertyExpression DataRange′)′ declares the range of a certain variable.

3.1.5. Annotation Define

Applications of OWL frequently require tracks to associate supplementary knowledge with axioms, entities, and ontologies. Annotation defines this additional knowledge as axioms, entities, and ontology annotation. It is as simple as a general sentence that declares the domain and range of intercommunicated variables. However, the annotation syntax and structure of the comment and expression is defined by the ontology designer, independent of an ontology′s structural grammar.

3.1.6. Axiom

A bunch of hypotheses is the principal feature of an OWL ontology called axiom which looks like a statement that depicts the actual existence true to the domain. Axiom class in the structural specification in OWL is an extended comprehensive set of axioms. OWL axiom can be declarations about the classes, objects, or data properties. A class expression like class1 is a subclass of another class expression, such as class2, which could be expressed as a subclass axiom SubClassOf (class1 class2). Similarly, if a class expression like class1 is pairwise disjointed to class2, class3 and classn could be defined as:
DisjointClasses (Class1 … Classn). If a class, class1, is a disjoint union of the class declarations classi, 1 ≤ in, it could be defined by the disjoint union axiom:
DisjointUnion (class1 … classn).

3.1.7. Individual Definition Establishment

In a domain, an individual represents the existence of an actual object through OWL syntax. Anonymous individuals and named individuals are two types of individuals in OWL syntax. Same objects are referred to in any ontology through an explicit name given by a named individual. In an OWL ontology, there are restrictions on identifying named individuals; reserved vocabulary cannot be used. However, anonymous individuals are regional to an ontology as they do not have a global name. In a resource description framework (RDF), anonymous or unidentified individuals are analogous to empty nodes. The representation of an individual is:
Individual: = AnonymousIndividual | NamedIndividual

3.1.8. Ontology Documentation

A notional vision could be represented in phrases of structural specification using OWL ontology. An ontology document is associated with each ontology and physically stores the ontology in a specific way.
Free OWL tools offer OWL syntax that helps to store ontologies in a written text document format. The structural specification does not represent the ontology documents altogether. In contrast, OWL specification makes these two assumptions regarding their nature:
  • Every well-defined ontology converts into an ontology document that is a structural specification and illustrates the ontology UML class.
  • Using the appropriate protocol, an IRI helps to access individual ontology documents.

3.1.9. Semantic Interoperability Establishment

Semantic interoperability is not only concerned with the syntactical packaging of data but also the concurrent transmission of the semantics of those data (meaning). This procedure is performed by assessing metadata (data about the data) and merging the apiece data component into a shared and controlled vocabulary. Semantic interoperability makes a system understandable to an unknown system by taking the benefit of structural data exchange and vocabulary, including data codification. Therefore, the data interpretation becomes more accessible through information technology to the receiving end. The highest level of interoperability among the systems is represented by semantic interoperability. Semantic interoperability deals with the messaging format as well as concerns with the message content. Figure 2 shows the resultant ontology graph view, for social media activity by single user ID, defined by OWL language.

3.2. Pre-Processing of the Text Data

In this phase, the machine learning process discovers pre-defined categories like abbreviations, unintentional characters, symbols, emoji, etc. Depending upon the use cases, there are several steps to pre-process the text data. We followed the sentiment classification standard by applying the natural language processing (NLP) model. Depending upon the subject of the consequence, we adopt seven sub-steps to pre-process our data, as follows in the following sections.

3.2.1. Remove Unconventional Space

Weird spaces can easily be removed from an input expression using a pre-defined natural language toolkit (NLTK) function in python. The operation removes unnecessary spaces from a sentence and returns join (text).

3.2.2. Tokenization

Tokenization of a sentence helps us reduce the number of words required to train a machine learning model. Tokenization not only converts words into tokens but also deals with the spaces into a phrase. From the NLTK library, we can refer to the ′word_tokenizer′ function or write our own. There are several tokenizer func-tions like ′word_tokenize,′ ′wordpunct_tokenize′, ′sent_tokenize′, etc. The ′word_tokenize′ function is based on treebank tokenizer, whereas the ′word-punct_tokenize′ function is based on simple regular expression.

3.2.3. Spelling Correction

The success of a real-time sentiment analyzer highly depends on the spelling correction phase. This particular level comprises two sub-layers: the embedding layer and bidirectional stacked char–long short-term memory (LSTM) network; it holds a collection of words or a language-based dictionary. Peter Norvig′s spelling correction method is highly preferred for social media data analysis among several spelling correction methods.

3.2.4. Contraction Mapping

The next crucial step is contraction mapping. A contraction is a squeeze or an abbreviated form of a succession of words. A computer considers a word and its contraction as two different expressions because it does not understand that contractions stand for abbreviations for a sequence of words and have a precisely similar meaning. As a result, it makes computation more expensive while raising the dimensionality of the composition span matrix by proffering two different columns. Contraction mapping replaces the abbreviation form with its valid syntax.

3.2.5. Stemming

Stemming is another essential step used in advanced machine learning procedures to reduce a word to its root form. Stemming and lemmatization are used alternatively in information systems because both the operations convert a term into its base form. But the fundamental difference between them is that lemmatization always reduces a word to a foundation form with linguistic meaning. In comparison, stemming may not be able to recede a word to its elemental composition with reasonable linguistic meaning. A different language is supported by distinct stemming or the lemmatization function in information systems. A linguistic specialist could develop a Lemmatizer, whereas we commonly use a stemming algorithm for data science purposes. NLTK supports several stemmers, like Porter stemmer, Snowball stemmer, Lancaster stemmer, etc. Ultimately, this sub-process reduces the training data volume needed to train our machine learning model.

3.2.6. Emoji Handling

Emoji handling is another essential step in pre-processing the data phase to analyze text data from multi-social media platforms. Here, we used a pre-defined emoji library to regulate the emoji.

3.2.7. Stop Words Handling

The final sub-step is to operate with stop words. This procedure helps to prioritize and filter out the words which could be considered as individual features in the case of NLP. In our case, the procedure also includes the cleaning of hypertext markup language (HTML).

3.3. Feature Extraction and Selection

Feature extraction and selection is a process to reduce data dimensionality. It helps to choose or select a more suitable input dataset (a subset of the available dataset) for a machine learning algorithm. If a feature set is defined by F(i) = {f1, …, fi, …, fn}, then the expected subset (F(s)) of selected features after this particular method could be expressed as F(s) F(i). However, this subset F(s) maximizes the learner′s proficiency by maximizing some mathematically defined scoring function. The feature selection mechanism fulfills two primary goals:
  • Dimensionality reduction is essential when there are many variables in the input dataset.
  • Dimensionality reduction is essential when many variables exist in the input dataset.
Dimensionality reduction occurs intending to encounter an optimal feature subset able to maximize the scoring function. In most cases, theoretically, Bayes error rate reduction is the best; however, often, the performance term is associated with the expectation by the chosen classifier model.

3.3.1. Bag of Words

Using a bag of words (BoW), a machine converts the word frequency matrix into its vector representation. The BoW document matrix could be of two types. One is BoW, and the other is binary BoW. For BoW, the machine is concerned with the frequency of a particular feature within each sequence. In binary BoW, the model is only concerned with the presence of the component in an input sequence. If a specific feature is present in the sequence, it is considered ′1′ or substituted by ′0′ if it is absent. The BoW procedure gives equal weightage to each present feature, providing poor semantic representation. We need another TF-IDF technique (term frequency and inverse document frequency) to solve this deficiency.

3.3.2. Part of Speech

Each word belongs to a general grammar category and is called a part of speech (PoS). Therefore, each order of a feature has its weightage towards the sentiment analysis of an input string.

3.3.3. TF-IDF and SVD

The most traditional way to handle text data has four pathways: hashing of words, count vectorization, term frequency–inverse document frequency (TF-IDF), and singular-value decomposition (SVD).
T F t = H o w   m a n y   t i m e   a   t e r m   t   a p p e a r s   i n   a   d o c u m e n t N u m b e r   o f   t o t a l   t e r m s   p r e s e n t   i n t o   t h e   d o c u m e n t
T F I D F = L o g ( N u m b e r   o f   t o t a l   d o c u m e n t s D o c u m e n t   n u m b e r s   h a v i n g   t e r m   t   i n   i t )
T F I D F t = T F t I D F ( t )
Likewise, in SVD, one can follow further steps: latent semantic analysis and scikit learn version of SVD for n numbers of components. It has already been verified that by applying TF-IDF and SVD together, a model can improve the result of logistic regression and XGB.

3.3.4. Word Embedding

A word-embedding layer is efficiently used in the practical implementation field of long short-term memory (LSTM), like sentiment analysis from text data. This layer is available in Keras. While coming to the word to vector (Word2Vec) feature extraction procedure, there are two-layered neural networks; they give a multi-dimensional word vector for every word in the dictionary or corpus a language has. Word2vec can represent a word or even a whole sentence in its vector form. It takes all the words in a sentence and tokenizes them, using only alpha-numeric words and removing the stop words. The process eventually appends everything to a list vector and normalizes it by the square root function of the sum over the columns.

3.4. Classification Model

We followed the proposed ML model, inspired by Algorithm 2, applying the advanced natural language processing (NLP) method. Depending upon the subject of the consequence, the steps of the algorithm are as follows in the next section.

3.4.1. Data Balancing

A deal imbalanced dataset is quite common in real-time classification problems. Sometimes the number of levels in one observation class is comparatively less than other target class levels. This kind of imbalanced dataset gives significantly inaccurate results in the classification method. There is a solution to this problem called resampling (oversampling and undersampling). The synthetic minority oversampling technique, or SMOTE, is an appropriate technique to oversample the minority class in our case.

3.4.2. Imputation

Imputation is a strategy that assists in retaining most of the data/information by substituting the missing data with alternate values in a dataset. Removing data is not feasible as it can lead to inappropriate size reduction, resulting in biasing and incorrect analysis by a model.
Most Python libraries do not have an inbuilt missing data handling strategy, and the ML model can lead to errors if the problem is not fixed. We can make an approach made up of several imputations for different classification model variables like numerical, date–time, categorical, and mixed (both for numerical and categorical) datasets.
Algorithm 2: Proposed advanced NLP-based ML model.
Input:  i = 1 n j = 1 n D i , j .
Output: Best fit solution for the proposed ML model.
Start
   Step 1: Import libraries
   Step 2: If count null from ( D i , j ): df.isnull().sum = 0;
   Step 3: Calculate correlation:
r ( c ) = i = n ( x i x ) ( y i y ) i = n ( x i x ) 2 ( y i y ) 2
   where ( x i , y i = x and y variable sample,  x , y  = means of values in x and y variables.
   Step 4: Else
      i. Remove null, replacing by most frequent occurrence;
     ii. Repeat step 4;
   Step 5: If ( r c i , j 0.01   o r   0.01 )
   Remove feature ( D j ) : from ( D i , j ) ;
   Step 6: Calculate feature importance:
E n t r o p y S = i = 1 n p i l o g 2 ( p i )
   where S: Unique class label numbers;  p i : proportion between rows and output label.
   Select top k features;
   Step 7: Split train-test data sets;
   Step 8: Best fit calculation:
       i. ′n_estimators′: 1000,
      ii. ′min_samples_split′: 2,
     iii. ′min_samples_leaf′: 1,
     iv. ′max_features′: ′sqrt′,
      v. ′max_depth′: 25
   Step 9: rf_random.best_score calculation
   Step 10: MAE, MSE, RMSE, R2 Score calculation
End

3.4.3. Cross-Validation

Cross-validation is a technique in ML to test a model′s stability in terms of efficiency. Certifying an ML model is quite impossible, depending only upon the training dataset. Therefore, the cross-validation procedure first reserves a sample dataset from the primary dataset called the validation set; it is a subset of the initial dataset. Before deployment, an ML model is tested with this validation dataset, apart from the testing one. If the model performs well, then the model goes for the next step. This procedure checks the model′s efficiency by preparing the model with an input subset and testing the model on a prior unseen and independent input subset.

3.4.4. Hyper-Parameter Tuning

Some of the parameters in an ML model are predefined and not to be learned through training. These essential and fixed parameters, known as hyper-parameters, usually express several properties like the model′s speed or complexity.

4. Resultant Outcomes

The data from different social media platforms are diverse in type and length. Also, various expressions and actions define similar information. Because of this, a machine is uncertain about collecting each word′s emotional status and activity in an identical score bucket. We introduced a pipeline model able to include multiple hyper-parameters. Encountering the most satisfactory combination of such parameters is a searching procedure known as hyper-parameter tuning. GridSearchCV and RandomizedSearchCV are alternatively used for this hyper-parameter tuning process as demanded. First, we collected four datasets from four social media platforms with different patterns to design the model. We aimed to measure and analyze the interoperability value among those four social media platforms handled by a single user. Primarily the merged dataset includes the diverse level of emotion values for 1104 rows × 11 columns.
The dataset contains several numerical and categorical features like emotion values for a particular word, the number of accounts held by a single user, age group, hours of usage group, etc. As per the numerical computation, a specific word′s sentiment value is divided into essential parts like happiness, anger, disgust, surprise, neutral, sad, and fear. Apart from these, we considered up to four different accounts held by a single user id into four different social media applications. We also considered five diverse age groups of the users and up to 10 h of active usage. Next, we estimated the correlations among the features to choose the more effective variables, not the aggregate. Figure 3 shows the heat map of the primarily considered features. We followed the Pearson’s Correlation coefficients as the correlation measurement function.
Like line graphs, two-dimensional scatter plots show the impact of one variable upon another with the help of dots against horizontal and vertical axes. We observed data visualization by ′multivariate′ or ′multi-variable′ scatter plot visualization. Precisely, the relationships between numerous dataset attributes are comprehended by multivariate visualization. Figure 4 depicts the relation between happiness values concerning total hour-wise activity on a number of social media accounts.
A density plot is another straightforward and convenient approach for obtaining each distributed attribute in one frame. Density plotting is most likely a histogram, connecting each domain′s top by a smooth curve, also known as abstracted histograms. Figure 5 is the plotted density plot concerning happiness significance. The range (with α value 0.12) and domain of the curve are significantly satisfactory.
We decided to choose the random forest regressor to split the dataset into train and test sub-datasets. And, in the case of the tuning of hyper-parameters, we set randomized search CV. We selected the random forest regressor function in our case as we have mixed dataset features (both numerical and categorical). This particular regression provides limited explaining ability with a high prediction accuracy for integrated datasets.
Primarily, we experimented with two ways to train our model: one considering all primary features and the other after discarding the features based on the correlation factor. Figure 6 shows the correlation heat map among the effectively correlated features. Furthermore, we continued with the effectiveness calculation after pitching the less convincing variables.
From the learning algorithm′s perspective, the relevancy of a feature could be defined in many ways. It may intend to determine the relevance of a particular feature; it could be the overall relevancy among all the features. Two degrees of significance are mostly counted when the machine searches for strong and weak relevance. A distinct feature is trusted as relevant if it has a strong or weak relevance score; otherwise, it is discarded from the dataset. An optimal Bayes classifier determines the strong relevance of a feature or variable in such a way that the resultant performance of the overall ML model could be deteriorated depending upon one variable/feature vi.
Where the set of features/variables Fi = {v1, …, vi − 1, vi + 1, … vn}, except vi. vi is considered a strongly relevant feature if p(vi = xi, Fi = fi) > 0 for some existence of xi, y, and fi. A value-assignment fi denotes over each feature from Fi, such that:
p(Y = y | vi = xi; Fi = fi) ≠ p(Y = y | Fi = fi). p stands for the performance variable of an optimal Bayes classifier. Figure 7 shows the relevancies of different features during model training.
Similar to the previous case, we followed the random forest regressor [27] to split the dataset into train and test sub-datasets. And, in the case of hyper-parameters tuning, we set randomized search CV. Concerning the neutral emotion values, the density curve shows how much the range is relatively less than in the last case. The confidence region has reduced significantly as the arc is sharper in shape, and the alpha value is now 0.02, which is pretty much effective in the case of emotion detection.
Figure 8 is the ultimate density curve for our improved ML model concerning the most stable emotional state, i.e., the neutral state. Figure 9 shows the scatter plot of the test dataset after the model′s training with best-parameter tuning and selection. Furthermore, MAE, MSE, RMSE, and R2 values [28] have evaluated the measures of the ultimate model performance [3], shown in Table 1. All the measurements support the success of our proposed model. Like the R2 value assured, 97% of the model′s inputs can explain the observed variation, which is favorable and satisfactory.
Semantic level interoperability through defined ontology has made it possible to collect data from distinct social media platforms; hence, we have integrated and illustrated training datasets. The model has been further improved compared to the previous case of feature selection. Consequently, we enhanced the proposed model′s efficiency by selecting effective features from the collected dataset.

5. Conclusions and Future Work

Considering a few difficulties in the emotion detection and sentiment-analysis-based ML models, we propose an improved social media sentiment analytics technique using advanced NLP to predict the individual state of mind to resist some profound events. The solution connects several social media platforms used by a single user id and observes their posts, comments, and feedback throughout the day. We considered dividing and seaming for two different but complementary counting of each inter- and intra-linked expression through happiness and a neutral state of mind. Using semantic level interoperability, we removed inter-operability issues among distributed user servers of other but dependent social media accounts. In brief, the proposed algorithm can track down more than one semantically linked activity from diverse social media platforms by a distinct person with a high success rate. The proposed model could have future scope in different industrial concerns, like emo-robot [29], patient care, elderly care [30], and drug recommendation system [6], etc., successfully after the following modifications:
  • We should train the model for more categorical combinations of features in the field of industrial IoT [31].
  • We must improve the model to incorporate multi-level gateways to collect from more than four accounts and user ids [32].
  • Currently, the data collection method is manual; we must automate it [33].
  • Industrial IoT users could comprise the model with IoT modules via a similar user ID for better human-like assistance [34].
  • Furthermore, blockchain security could be incorporated for secured multi-level data collection and, hence, no data loss [35,36].

Author Contributions

Conceptualization, A.A. and S.B.K.; methodology, S.B.K., S.A. and J.A.; validation, S.B.K. and F.A., formal analysis, A.A., S.B.K., J.A., S.A. and F.A.; data curation, S.B.K. and A.A.; writing—original draft preparation, A.A. and S.B.K.; writing—review and editing, J.A., S.A. and F.A.; visualization, F.A.; project administration, A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Deputy for Research and Innovation—Ministry of Education, Kingdom of Saudi Arabia for this research through a grant (NU/IFC/2/SERC/-/9) under the Institutional Funding Committee at Najran University, Kingdom of Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to acknowledge the support of the Deputy for Research and Innovation-Ministry of Education, Kingdom of Saudi Arabia for this research through a grant (NU/IFC/2/SERC/-/9) under the Institutional Funding Committee at Najran University, Kingdom of Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kushwah, S.; Kalra, B.; Das, S. Sentiment Analysis of Healthcare Big Data: A Fundamental Study. In Computationally Intelligent Systems and Their Applications; Springer: Singapore, 2021; pp. 53–70. [Google Scholar]
  2. Naresh, A.; Venkata Krishna, P. An efficient approach for sentiment analysis using machine learning algorithm. Evol. Intell. 2021, 14, 725–731. [Google Scholar] [CrossRef]
  3. Kinyua, J.D.; Mutigwe, C.; Cushing, D.J.; Poggi, M. An analysis of the impact of president trump’s tweets on the djia and S&P 500 using machine learning and sentiment analysis. J. Behav. Exp. Financ. 2021, 29, 100447. [Google Scholar]
  4. Jindal, K.; Aron, R. A systematic study of sentiment analysis for social media data. Mater. Today Proc. 2021, in press. [CrossRef]
  5. Van Atteveldt, W.; van der Velden, M.A.; Boukes, M. The validity of sentiment analysis: Comparing manual annotation, crowd-coding, dictionary approaches, and machine learning algorithms. Commun. Methods Meas. 2021, 15, 121–140. [Google Scholar] [CrossRef]
  6. Hiremath, B.N.; Patil, M.M. Enhancing optimized personalized therapy in clinical decision support system using natural language processing. J. King Saud Univ. Comput. Inf. Sci. 2020, 34, 2840–2848. [Google Scholar] [CrossRef]
  7. Eng, T.; Nawab, M.R.I.; Shahiduzzaman, K.M. Improving accuracy of the sentence-level lexicon-based sentiment analysis using machine learning. Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol. 2021, 3307, 57–68. [Google Scholar] [CrossRef]
  8. Alawneh, E.; Al-Fawa’reh, M.; Jafar, M.T.; Al Fayoumi, M. Sentiment analysis-based sexual harassment detection using machine learning techniques. In Proceedings of the 2021 International Symposium on Electronics and Smart Devices (ISESD), Bandung, Indonesia, 29–30 June 2021; pp. 1–6. [Google Scholar]
  9. Yadav, N.; Kudale, O.; Rao, A.; Gupta, S.; Shitole, A. Twitter sentiment analysis using supervised machine learning. In Intelligent Data Communication Technologies and Internet of Things; Springer: Singapore, 2021; pp. 631–642. [Google Scholar]
  10. Bhatia, S. A comparative study of opinion summarization techniques. IEEE Trans. Comput. Soc. Syst. 2020, 8, 110–117. [Google Scholar] [CrossRef]
  11. Zhang, X.; Huang, D.; Li, H.; Zhang, Y.; Xia, Y.; Liu, J. Self-training maximum classifier discrepancy for EEG emotion recognition. CAAI Trans. Intell. Technol. 2023. [Google Scholar] [CrossRef]
  12. Wang, H.; Cui, Z.; Liu, R.; Fang, L.; Sha, Y. A Multi-type Transferable Method for Missing Link Prediction in Heterogeneous Social Networks. IEEE Trans. Knowl. Data Eng. 2023, 2023, 3233481. [Google Scholar] [CrossRef]
  13. Wu, B.; Liu, Z.; Gu, Q.; Tsai, F. Underdog mentality, identity discrimination and access to peer-to-peer lending market: Exploring effects of digital authentication. J. Int. Financ. Mark. Inst. Money 2023, 83, 101714. [Google Scholar] [CrossRef]
  14. Huang, C.; Han, Z.; Li, M.; Wang, X.; Zhao, W. Sentiment evolution with interaction levels in blended learning environments: Using learning analytics and epistemic network analysis. Australas. J. Educ. Technol. 2021, 37, 81–95. [Google Scholar] [CrossRef]
  15. Saab, S., Jr.; Fu, Y.; Ray, A.; Hauser, M. A dynamically stabilized recurrent neural network. Neural Process. Lett. 2022, 54, 1195–1209. [Google Scholar] [CrossRef]
  16. Ni, Q.; Guo, J.; Wu, W.; Wang, H.; Wu, J. Continuous Influence-Based Community Partition for Social Networks. IEEE Trans. Netw. Sci. Eng. 2022, 9, 1187–1197. [Google Scholar] [CrossRef]
  17. Behera, R.K.; Jena, M.; Rath, S.K.; Misra, S. Co-LSTM: Convolutional LSTM model for sentiment analysis in social big data. Inf. Process. Manag. 2021, 58, 102435. [Google Scholar] [CrossRef]
  18. Saab, S., Jr.; Saab, K.; Phoha, S.; Zhu, M.; Ray, A. A multivariate adaptive gradient algorithm with reduced tuning efforts. Neural Netw. 2022, 152, 499–509. [Google Scholar] [CrossRef]
  19. Qiu, X.; Sun, T.; Xu, Y.; Shao, Y.; Dai, N.; Huang, X. Pre-trained models for natural language processing: A survey. Sci. China Technol. Sci. 2020, 63, 1872–1897. [Google Scholar] [CrossRef]
  20. Tiwari, A.K.; Bathia, D.; Bouri, E.; Gupta, R. Investor sentiment connectedness: Evidence from linear and nonlinear causality approaches. Ann. Financ. Econ. 2021, 16, 2150016. [Google Scholar] [CrossRef]
  21. Al-Shehhi, A.; Grey, I.; Thomas, J. Big data and wellbeing in the Arab world. In Positive Psychology in the Middle East/North Africa: Research, Policy, and Practise; Springer: Berlin/Heidelberg, Germany, 2019; pp. 159–182. [Google Scholar]
  22. Barr, J.R.; Shaw, P.; Abu-Khzam, F.N.; Chen, J. Combinatorial text classification: The effect of multi-parameterized correlation clustering. In Proceedings of the 2019 First International Conference on Graph Computing (GC), Laguna Hills, CA, USA, 25–27 September 2019; pp. 29–36. [Google Scholar]
  23. Gupta, R.K.; Gurumoorthy, B. Feature-based ontological framework for semantic interoperability in product development. Adv. Eng. Inform. 2021, 48, 101260. [Google Scholar] [CrossRef]
  24. Adamczyk, B.S.; Szejka, A.L.; Júnior, O.C. Knowledge-based expert system to support the semantic interoperability in smart manufacturing. Comput. Ind. 2020, 115, 103161. [Google Scholar] [CrossRef]
  25. Szejka, A.L.; Canciglieri, O., Jr.; Panetto, H.; Rocha Loures, E.; Aubry, A. Semantic interoperability for an integrated product development process: A systematic literature review. Int. J. Prod. Res. 2017, 55, 6691–6709. [Google Scholar] [CrossRef]
  26. Veltman, K.H. Syntactic and semantic interoperability: New approaches to knowledge and the semantic web. New Rev. Inf. Netw. 2001, 7, 159–183. [Google Scholar] [CrossRef]
  27. Bashir, M.F.; Javed, A.R.; Arshad, M.U.; Gadekallu, T.R.; Shahzad, W.; Beg, M.O. Context aware emotion detection from low resource urdu language using deep neural network. Trans. Asian Low Resour. Lang. Inf. Process. 2022, 22, 1–30. [Google Scholar] [CrossRef]
  28. Shaw, P.; Barr, J.R.; Abu-Khzam, F.N. Anomaly detection via correlation clustering. In Proceedings of the 2022 IEEE 16th International Conference on Semantic Computing (ICSC), Laguna Hills, CA, USA, 26–28 January 2022; pp. 307–313. [Google Scholar]
  29. Ling, H.Y.; Bjorling, E.A. Sharing stress with a robot: What would a robot say? Hum.-Mach. Commun. 2020, 1, 133–159. [Google Scholar] [CrossRef] [Green Version]
  30. Alexiou, K.I.; Roushias, A.; Varitimidis, S.E.; Malizos, K.N. Quality of life and psychological consequences in elderly patients after a hip fracture: A review. Clin. Interv. Aging 2018, 13, 143. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Chakraborty, K.; Bhatia, S.; Bhattacharyya, S.; Platos, J.; Bag, R.; Hassanien, A.E. Sentiment Analysis of COVID-19 tweets by Deep Learning Classifiers—A study to show how popularity is affecting accuin social media. Appl. Soft Comput. 2020, 97, 106754. [Google Scholar] [CrossRef]
  32. Haraty, R.A.; Boukhari, B.; Kaddoura, S. An Effective Hash-Based Assessment and Recovery Algorithm for Healthcare Systems. Arab. J. Sci. Eng. 2021, 147, 1523–1536. [Google Scholar] [CrossRef]
  33. Zhang, D.; Shen, Y.; Huang, Z.; Xie, X. Auto machine learning-based modelling and prediction of excavation-induced tunnel displacement. J. Rock Mech. Geotech. Eng. 2022, 14, 1100–1114. [Google Scholar] [CrossRef]
  34. Alshamrani, M. IoT and artificial intelligence implementations for remote healthcare monitoring systems: A survey. J. King Saud Univ. Comput. Inf. Sci. 2021, 34, 4687–4701. [Google Scholar] [CrossRef]
  35. Omran, A.; Abouyoussef, M.; Ismail, M.; Bhatia, S. Sharded Blockchain-based Online Diagnostic System for Suspected Patients During Pandemics. In Proceedings of the 2022 IEEE Wireless Communications and Networking Conference (WCNC), Austin, TX, USA, 10–13 April 2022; pp. 2715–2720. [Google Scholar]
  36. Kumar, R.; Sharma, R. Leveraging blockchain for ensuring trust in IoT: A survey. J. King Saud Univ. Comput. Inf. Sci. 2021, 34, 8599–8622. [Google Scholar] [CrossRef]
Figure 1. Work flow diagram of the proposed prototype model.
Figure 1. Work flow diagram of the proposed prototype model.
Applsci 13 07599 g001
Figure 2. Resultant ontology graph view, defined by OWL language.
Figure 2. Resultant ontology graph view, defined by OWL language.
Applsci 13 07599 g002
Figure 3. Heat map of correlation matrix representation between primarily considered features.
Figure 3. Heat map of correlation matrix representation between primarily considered features.
Applsci 13 07599 g003
Figure 4. The relation between happiness values concerning total hour-wise activity on a number of social media accounts.
Figure 4. The relation between happiness values concerning total hour-wise activity on a number of social media accounts.
Applsci 13 07599 g004
Figure 5. Density plot concerning significant happiness value.
Figure 5. Density plot concerning significant happiness value.
Applsci 13 07599 g005
Figure 6. Heat map of correlation matrix representation between effective features selection.
Figure 6. Heat map of correlation matrix representation between effective features selection.
Applsci 13 07599 g006
Figure 7. Relevancies of different features during training.
Figure 7. Relevancies of different features during training.
Applsci 13 07599 g007
Figure 8. Density plot concerning neutral state of mind.
Figure 8. Density plot concerning neutral state of mind.
Applsci 13 07599 g008
Figure 9. Scatter plot of the test dataset after the model′s training with best-parameter tuning and selection.
Figure 9. Scatter plot of the test dataset after the model′s training with best-parameter tuning and selection.
Applsci 13 07599 g009
Table 1. Model′s efficiency calculation factors with values.
Table 1. Model′s efficiency calculation factors with values.
Value ofAbbreviation FormCalculated Value
Mean absolute errorMAE0.0015339035
Mean squared errorMSE0.0000081479
Root mean squared error on predictionRMSE0.0028544564
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alqahtani, A.; Khan, S.B.; Alqahtani, J.; AlYami, S.; Alfayez, F. Sentiment Analysis of Semantically Interoperable Social Media Platforms Using Computational Intelligence Techniques. Appl. Sci. 2023, 13, 7599. https://doi.org/10.3390/app13137599

AMA Style

Alqahtani A, Khan SB, Alqahtani J, AlYami S, Alfayez F. Sentiment Analysis of Semantically Interoperable Social Media Platforms Using Computational Intelligence Techniques. Applied Sciences. 2023; 13(13):7599. https://doi.org/10.3390/app13137599

Chicago/Turabian Style

Alqahtani, Ali, Surbhi Bhatia Khan, Jarallah Alqahtani, Sultan AlYami, and Fayez Alfayez. 2023. "Sentiment Analysis of Semantically Interoperable Social Media Platforms Using Computational Intelligence Techniques" Applied Sciences 13, no. 13: 7599. https://doi.org/10.3390/app13137599

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop