Next Article in Journal
Statistical Inference of Inverted Exponentiated Rayleigh Distribution under Joint Progressively Type-II Censoring
Next Article in Special Issue
Learned Practical Guidelines for Evaluating Conditional Entropy and Mutual Information in Discovering Major Factors of Response-vs.-Covariate Dynamics
Previous Article in Journal
A Word-Granular Adversarial Attacks Framework for Causal Event Extraction
Previous Article in Special Issue
Categorical Nature of Major Factor Selection via Information Theoretic Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unraveling Hidden Major Factors by Breaking Heterogeneity into Homogeneous Parts within Many-System Problems

1
Department of Statistics, National Chengchi University, Taipei 11605, Taiwan
2
Institute of Statistical Science, Academia Sinica, Taipei 11529, Taiwan
3
Department of Statistics, University of California at Davis, Davis, CA 95616, USA
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(2), 170; https://doi.org/10.3390/e24020170
Submission received: 29 November 2021 / Revised: 5 January 2022 / Accepted: 19 January 2022 / Published: 24 January 2022
(This article belongs to the Special Issue Information Complexity in Structured Data)

Abstract

:
For a large ensemble of complex systems, a Many-System Problem (MSP) studies how heterogeneity constrains and hides structural mechanisms, and how to uncover and reveal hidden major factors from homogeneous parts. All member systems in an MSP share common governing principles of dynamics, but differ in idiosyncratic characteristics. A typical dynamic is found underlying response features with respect to covariate features of quantitative or qualitative data types. Neither all-system-as-one-whole nor individual system-specific functional structures are assumed in such response-vs-covariate (Re–Co) dynamics. We developed a computational protocol for identifying various collections of major factors of various orders underlying Re–Co dynamics. We first demonstrate the immanent effects of heterogeneity among member systems, which constrain compositions of major factors and even hide essential ones. Secondly, we show that fuller collections of major factors are discovered by breaking heterogeneity into many homogeneous parts. This process further realizes Anderson’s “More is Different” phenomenon. We employ the categorical nature of all features and develop a Categorical Exploratory Data Analysis (CEDA)-based major factor selection protocol. Information theoretical measurements—conditional mutual information and entropy—are heavily used in two selection criteria: C1—confirmable and C2—irreplaceable. All conditional entropies are evaluated through contingency tables with algorithmically computed reliability against the finite sample phenomenon. We study one artificially designed MSP and then two real collectives of Major League Baseball (MLB) pitching dynamics with 62 slider pitchers and 199 fastball pitchers, respectively. Finally, our MSP data analyzing techniques are applied to resolve a scientific issue related to the Rosenberg Self-Esteem Scale.

1. Introduction

A Many-System Problem (MSP) is a scientific study involving many complex systems [1,2,3]. Such systems are basically governed by the same dynamic principles with many complex tuning parameters that embrace hardly measured idiosyncratic characteristics. This nature of MSPs can be seen in almost all scientific areas and real-world research involving any sort of heterogeneity, such as study subjects of different ages and genders in psychology and neuroscience [4], many distinct households in microeconomics [5], and the well-known many-body problem in physics [6], just to name a few. It is notable that the name MSP is inspired by the many-body problem.
Data observed from each individual system, in general, might not be large enough in size to sustain an insightful study into its individual governing principles and characteristics. Therefore, we need to aggregate all data from all systems in order to understand the systems’ fundamental dynamics. However, as we pool all individual ID-marked data sets into a huge data set, this data pooling generates many confounding effects due to existing diverse aspects of dissimilarity across all systems. Such confounding effects are collectively termed as heterogeneity. That is, the effects of governing principles are expected to interact with the effects of the heterogeneity contained in the data. This is the dilemma facing every MSP, which is the topic of this paper.
In the literature of complex system study [3,6], as well as statistical physics [7], we could not find effective approaches to coherently resolve such a dilemma in any MSPs. One chief difficulty underlying any MSP rests on the categorical nature of data in both response and covariate sides of system dynamics. Certainly, the system-ID is categorical. Another chief difficulty is heterogeneity among all member systems. These two difficulties are linked. This categorical feature naturally gives rise to multiscale issues that are closely tied to the heterogeneity potentially existing among all member systems. First, systems’ individual and distinct characteristics as fine-scale manifestations of heterogeneity can hardly be accommodated into any analytic equation or modeling structure in a systematic fashion. Secondly, the effects of interactions between the system-ID and all other features as mid-scale manifestations of heterogeneity are nearly impossible to capture precisely in man-made system descriptions. Thirdly, some individual systems are similar, while being dissimilar from other systems. Such similarity and dissimilarity among all involving systems can cause large-scale manifestations of heterogeneity.
In this paper, we first design an artificial MSP that captures the above multiscale structures and then reveals permeating effects of heterogeneity. In this example, we clearly show how heterogeneity constrains and hides essential mechanisms, and then how homogeneity can open up and reveal the true mechanisms. Then we turn to the real MSPs of interest in this paper; we consider physical pitching dynamics in Major League Baseball (MLB) in the US. A healthy starting pitcher throughout the entire season could pitch up to or even over 3000 pitches of several pitch types. Therefore, there are only several hundred pitches belonging to each pitch type for each pitcher. This amount of data is unlikely to sustain a detailed study on pitch-type specific pitching dynamics. In fact, pitching dynamics are known to be governed by the biomechanics of musculoskeletal construction and the Magnus effect of spinning baseballs in aerodynamics [8]. There are many parameters for each of these two governing principles.
Even though the biomechanics derived from musculoskeletal construction is more or less the same across all MLB pitchers, slight differences via many tuning parameters collectively make up large differences. Likewise, the Magnus effect is a universal physical phenomenon when spinning a baseball, but there are fine-scale differences in creating spins that will result in significantly distinct Magnus effects. This is why all pitchers are different. Some pitchers’ pitching dynamics are closer to those of some pitchers than others. Further, different pitch types require slightly distinct biomechanical gestures and spinning mechanisms. As such, studying a collective of pitchers’ pitching dynamics for biomechanical and physical governing principles is rather complicated. In this paper, we only focus on pitch-type specific MSPs in MLB.
In this paper, we demonstrate a computational approach for effectively studying an ensemble of pitchers’ pitch-type specific dynamic systems through two stages. In the first stage, we resolve the following essential questions. Where are the effects of heterogeneity in this MSP example? How do these effects come about? How do they interfere with the effects of governing Magnus and biomechanical effects? We explicitly explore and confirm the answers to these questions with explicit details. In a nutshell, all answers are tied to the categorical feature of system-IDs because it is highly associative with almost all response and covariate features.
In the second stage, after exploring and confirming the effects of heterogeneity, it is essential to mitigate all aforementioned effects in order to unravel critical pattern information on governing principles. To achieve this goal, ideally, we need to partition the ensemble of pitchers by grouping similar pitchers into a collection of homogeneous groups, only within which we are able to retain a large enough data set to study group-specific pitching dynamics. Consequently, the differences among these homogeneous groups are surely large-scale effects of heterogeneity.
With the resulting manifestations in the two stages, we can see and appreciate the effects of heterogeneity. These effects weave magical fabrics contained in data’s information content. This magic phenomenon mirrors what was described by Nobel laureate physicist P.W. Anderson [9] in his 1972 Science paper titled “More is Different”: from more data to more relevant scales of patterns to more surprising discoveries. In this Big Data era, MSPs are and will become even more ubiquitous in the internet. The merits of such two-stage data analysis endeavors could be critical when studying a large collection of many complex systems. We demonstrate such merits by resolving an interesting scientific problem related to the Rosenberg Self-Esteem Scale in psychology and sociology [10] by analyzing a real data set collected online.

2. Structural Formation and Major Factors in MSP

In this section, we postulate a generic structural formation for any MSP setting. From each single complex system member of an MSP, a data point is measured and collected in a L + K D vector format. Let the first L ( = 2 ) components be measurements or categories of the response features denoted as Y = ( Y 1 , , Y L ) and the rest of K ( = 10 ) components be measurements or categories of K one-dimensional covariate features denoted as { V 1 , , V K } . It is essential to note that one covariate feature, say V K , is the categorical feature of system labels or IDs.
An unspecified complex structural relation between Y and { V 1 , , V K } consists of a collection of M unknown constituent mechanisms defined by M major factors { F m { A m * } | m = 1 , , M } , in the following fashion:
Y = Y 1 Y 2 Y L G ( F 1 { A 1 * } , F 2 { A 2 * } , , F M { A M * } ) + ε .
Here, we do not have any prior knowledge or assumptions of M, functional forms of F m { · } , the random noise ε , nor the governing structural function G ( · ) . We only focus on identifying feature memberships of each A m * ( { V 1 , , V K } ) with m = 1 , , M . By acquiring these memberships, the layouts or patterns of constituent mechanisms within dynamics underlying Y should be visible and explainable through contingency tables of A m * -vs- Y and m S A m * -vs- Y .
Therefore, our computational task can be simply described as: to discover the collection of major factors { A m * | m = 1 , , M } . This computational task of major factor selection primarily relies on information theoretical measures, as will be seen in the next section.

3. Methods

In this section, we first briefly review the selection protocol for the major factors underlying Y as proposed and illustrated in [11]. The basic foundation is the recently developed computational paradigm called Categorical Exploratory Data Analysis (CEDA) [12,13]. The name EDA was originally coined by John Tukey [14]. The fundamental idea behind CEDA is to let all features’ natural categories assemble freely in order to shed light on the true pattern information contained in data. Then, we present a new efficient algorithm for reliability checking and a generic plan for studying any Many-System Problem.

3.1. Developments of CEDA

The first step of CEDA is to categorize each response and covariate feature via its histogram, which can be properly built using an effective algorithm developed in [15]. This step serves to reduce the noise inherent in all measurements in order to reveal the intrinsic categorical structure of the histogram. The second step employs a contingency table for all developments involving information theoretical measures. The contingency table is used as a platform for coupling multiple categorized features together to form and define a new composite variable. This contingency table platform also serves as a platform for visualizing and evaluating possibly non-linear associations between any two variables. The two directional associations are numerically evaluated through conditional (Shannon) entropy. By properly rescaling with respect to corresponding marginal (Shannon) entropy, a Mutual Conditional Entropy (MCE) [16] is calculated. This association measure is the correct one when considering categorical features.
The third step of CEDA computes the conditional entropies of Y given all possible covariate feature combinations or feature sets, since the structural categories of features are allowed to reassemble freely in these response-vs-covariate contingency tables without being subject to man-made constraints. This collection of all possible directional associations should ideally contain all vital associative patterns that indicate all constituent mechanisms underlying the designated response variable Y within an MSP. Our selection protocol identifies exactly such a collection of major factors. Each major factor is subject to reliability checks.

3.2. Major Factor Selection Based on Information Theoretical Measurements

In this subsection, we briefly review the approach to major factor selection based on the information theoretical measurements proposed in [11].
All features in Y and all covariate features, denoted as V k with k = 1 , , K , are either categorical or categorized 1D covariate features. Let capital letters A or B denote different subsets of covariate features. Within the categorical nature of all features and the contingency table as a synthesizing platform, Y , A, and B can be treated as 1D composite categorical variables in the fashion of each occupied hypercube in their corresponding contingency table as a category.
Furthermore, any pair of 1D categorical features defines a contingency table, as do any pair of 1D composite categorical variables such as ( Y , A ) , ( Y , B ) , and ( A , B ) . For a contingency table, information theoretical measurements are natural tools for discovering associative patterns. Take ( Y , A ) as an example. Let all categories of Y be arranged along the column-axis, while all categories of A are arranged along the row-axis. Then, the resultant contingency table, denoted as < Y , A > , is constructed as a rectangle array of cell counts. With suitable permutations on column- and row-axes, by aggregating unoccupied zero cells as much as possible, associative patterns and relations between Y and A become graphically visible. All information theoretical measurements used here are invariant with respect to row and column permutations. Using a contingency table, we can still visualize the global and large-scale pattern formations contained in < Y , A > .
The aforementioned associative patterns can in fact be numerically evaluated via various versions of conditional entropies (CEs) by basically treating < Y , A > as a 2D histogram of bivariate ( Y , A ) . Given a column, say Y = y , we define a discrete conditional variable. Its Shannon entropy is calculated on this column’s vector of proportions, i.e., cell counts divided by its column sum, and is denoted as H [ A | Y = y ] . Across all rows, we calculate the weighted sum of H [ Y | A = a ] with respect to the weighting scheme of row sum proportions. This is the conditional entropy (CE), denoted as H [ Y | A ] . Likewise along the column-axis, we calculate the expected CE H [ A | Y | A ] .
The intuitive meanings of H [ A | Y ] and H [ Y | A ] are evidently visible through their contingency tables. This CE H [ Y | A ] conveys the amount of expected remaining uncertainty in Y after knowing A. The CE drop H [ Y ] H [ Y | A ] indicates the information explained by A. Therefore, it is natural to select the major factor based on the CE drop. It is worth emphasizing the fact that the conditional entropy drop indicates the shared amount information between A and Y ; see also the review paper [17]:
H [ Y ] H [ Y | A ] = H [ A ] H [ A | Y ] = H [ A ] + H [ Y ] H [ A , Y ] = I [ Y ; A ] .
where H [ A , Y ] is the joint Shannon entropy of bivariate variable ( Y , A ) , while I [ Y ; A ] is the mutual information between Y and A.
Next, consider the CE drop of bivariate ( A , B ) from the CE of Y . It is shown that
H [ Y ] H [ Y | ( A , B ) ] = { H [ Y ] H [ Y | A ] + H [ Y ] H [ Y | B ] } + { I [ A ; B | Y ] I [ A ; B ] } .
This simple difference { I [ A ; B | Y ] I [ A ; B ] } indeed conveys the essence of interpretable meaning of conditional mutual information and plays a key role in the feature selection protocol. Considering the case of A and B being marginally stochastically independent, which means I [ A ; B ] = 0 , the positivity of I [ A ; B | Y ] indicates that the CE drop of A and B jointly is larger than the sum of the marginal ones. Similarly, if the difference { I [ A ; B | Y ] I [ A ; B ] } is significantly larger than zero, this positiveness of I [ A ; B | Y ] I [ A ; B ] acts like the so-called “ecological effect”; the whole is larger than the sum of its parts. The ecological effects are essential in the process of identifying a vital collection of major factors of Y , which offers an avenue for understanding mechanisms underlying Y .
On the other hand, this difference I [ ( A , B ) | Y ] I [ A ; B ] could be nearly zero or even negative when A and B are highly associative. In this case, at most, either A or B may be a candidate of major factors of Y , but not both. This choice of major factor is a conservative way of decision-making.
Based on the concepts above, the following two criteria, “confirmable” and “irreplaceable”, are proposed to identify a major factor of Y :
[C1: confirmable]: 
A feature-set A is confirmable if a feature-set A ˜ is obtained by substituting any one of the feature members of A with a feature that is completely independent of Y and A; we have I [ Y ; A ] significantly larger than I [ Y ; A ˜ ] .
[C2: irreplaceable]: 
A feature-subset A is replaceable if I [ Y ; A ] I [ Y ; A 1 ] + I [ Y ; A 2 ] for any compositions of A, i.e.,  A = A 1 A 2 and A 1 A 2 = . For A to be declared irreplaceable, we require that A is not replaceable and simultaneously satisfies the following two extra conditions: (a) its CE drop is larger the sum of the top ranked CE drop and at least | A | -times its complementary feature-subset CE drop; (b) the candidate A joined with any already identified major factor A m * must achieve I [ Y ; A A m * ] I [ Y ; A ] + I [ Y ; A m * ] .
The criterion [C1: confirmable] is mainly used as a reliability check. This is carried out by Algorithm 1.
Algorithm 1: Simulate a contingency table with addition of a random noise feature.
Entropy 24 00170 i001
As for criterion [C2: irreplaceable], the condition (a) ensures that some kind of structural dependency among all subsets of A is embraced under the constraints imposed by Y , not just the occurrence of ecological effects. This condition allows the following to happen: A m * A m * for ( m , m ) with | A m * | < | A m * | and A m * A m * = B , in the sense that B is the complement of A m * in A m * . This selection of A m * is realistic only when the CE drop of A m * minus the CE drop of A m * is many times of B’s CE drop. That is, A m * and B have to form some “strong” bonds under the conditioning of Y in order to jointly become a higher-order major factor.
For the convenience of checking this condition, the difference between the CE drop of A and the top ranked CE drop of its feature-subsets is routinely calculated. It is called “successive CE drop” or denoted as “SCE drop” in tables.
The condition (b) again ensures the ecological effect among identified major factors. Though obvious, it is worth mentioning the difference between conditions (a) and (b). Condition (a) sets a very high bar for building up any high-order major factors, such as order-3 or higher, while condition (b) only requires the fulfillment of an ecological effect for any two major factors to coexist. That is, there still exists potential for a union of two identified major factors to become a high-order major factor, but condition (a) is rather hard to fulfill.

3.3. Reliability Checking Algorithm

To check the condition [C1: Confirmable], Algorithm 1 simulates a contingency table by expanding the covariate feature-set by including an extra random-noise feature. However, when the size of the covariate feature-set is large, the number of rows of the contingency table becomes very large. This largeness renders the algorithm inefficient. In this subsection, we propose a new algorithm to resolve this computing issue. This algorithm directly estimates the mean and variance of the conditional entropies (CEs) when expanding a covariate feature-set with an extra random-noise feature.
To achieve the goal of Algorithm 1, it suffices to estimate a CE in the case of breaking a single row into χ rows, which is the number of bins of the random-noise feature. The final CE estimate would be the weighted sum of the CE estimates across all the rows in the expanded contingency table. Consider a contingency table n i j χ × P as a random matrix. Each column is a multinomial random variable with a uniform probability { 1 / χ , , 1 / χ } . Let n · 1 , , n · P be the vector of original row sums. We break each of them into χ row sums. We have
n · j = i = 1 χ n i j .
Let sample size n
n = j = 1 P n · j
and row sums n i ·
n i · = j = 1 χ n i j .
The goal is to obtain the distribution of the empirical conditional entropy H ^ ( Y | R ) :
H ^ ( Y | R ) = i = 1 χ n i · n j = 1 P ( 1 ) n i j n i · log ( n i j n i · ) .
with some calculations below:
H ^ ( Y | R ) = i = 1 χ n i · n j = 1 P n i j n i · log ( n i j n i · ) = 1 n i = 1 χ j = 1 P n i j log ( n i j n i · ) = 1 n i = 1 χ j = 1 P n i j log ( n i j ) + 1 n i = 1 χ j = 1 P n i j log ( n i · ) = 1 n j = 1 P i = 1 χ n i j log ( n i j ) + 1 n i = 1 χ n i · log ( n i · ) = j = 1 P n · j n i = 1 χ n i j n · j log ( n i j n · j ) j = 1 P n · j n log ( n · j n ) + i = 1 χ n i · n log ( n i · n )
The first term of the last equation, j = 1 P n · j n i = 1 χ n i j n · j log ( n i j n · j ) can be viewed as the weighted empirical conditional entropy j = 1 P n · j n H ^ ( m u l t i ( n · j , [ 1 χ , , 1 χ ] ) ) . The second term j = 1 P n · j n log ( n · j n ) is the empirical conditional entropy of { n · 1 , , n · P } , which is a fixed number. The last term i = 1 χ n i · n log ( n i · n ) is the empirical conditional entropy H ^ ( m u l t i ( n , [ 1 χ , , 1 χ ] ) ) .
In practice, it is easier to use (2) to understand H ^ ( Y | R ) . The mean of H ^ ( Y | R ) equals the mean of j = 1 P i = 1 χ n i j log ( n i j ) + 1 n i = 1 χ n i · log ( n i · ) , where each term can be numerically computed separately. For a positive integer N, let x = ( x 1 , , x χ ) be the random variable from m u l t i ( N , [ 1 χ , , 1 χ ] ) . Define a function h ( x ) as
h ( x ) = i = 1 χ x i log ( x i ) .
We can numerically compute the mean and variance of h ( x ) . Denote m ( N ) and v ( N ) as the mean of h ( x ) for x from m u l t i ( N , [ 1 χ , , 1 χ ] ) . From (2), we have
E [ H ^ ( Y | R ) ] = 1 n m ( N ) j = 1 P m ( n · j ) .
For the variance, there is no exact and simple form, since the row sums n i · ’s are dependent on columns. However, since the entropy of the row sums is positively correlated to the sum of the entropies of the columns, it is safe to have j = 1 P v ( n · j ) / n 2 as the upper bound of the variance of H ^ ( Y | R ) :
V a r [ H ^ ( Y | R ) ] 1 n 2 j = 1 P v ( n · j ) .
In practice, when there are many rows in a contingency table, the number in each cell of the contingency table will not be large. That is, the unique numbers in the contingency table are relatively few. Therefore, it will become efficient to estimate the mean and variance using (3) and (4). Here we remark that the estimate of the mean by (3) is in fact more accurate than that from Algorithm 2, and the variance estimation is usually within 1.2 times the true value from our experimental experience.
Algorithm 2: Estimating the mean and the variance of the conditional entropy with a random noise feature included.
Entropy 24 00170 i002

3.4. A Generic Plan for Studying Any Many-System Problem

How can we effectively deal with complexity and heterogeneity embedded within any MSP? We propose the following study plan for MSPs under a framework with a designated response variable Y and a collection of covariate features { V 1 , , V K 1 } across all the individual systems.
MSP-1: 
Make the system-ID into an extra categorical feature and denote it as V K ;
MSP-2: 
Treat the MSP as a system described by Equation (1);
MSP-3: 
Perform the CEDA-based feature selection protocol to identify a vital collection of major factors, each of which is a feature-subset of { V 1 , , V K } that satisfies both criteria [C1: confirmable] and [C2: irreplaceable];
MSP-4: 
Apply the hierarchical clustering algorithm to partition the collection of systems into homogeneous system-clusters. This clustering is undertaken based on measurement vectors of major factors without including V K .
Here, a homogeneous system-cluster means that we find no major factors containing V K when a system-cluster is treated and studied as a small version of MSP by going through the steps MSP-1 to MSP-3. The collection of homogeneous system-clusters explicitly reveals the systemwise heterogeneity. Through computational results of MSP-2 and MSP-3, we will evidently see the interacting relationships between complexity and heterogeneity through memberships of major factors within the original MSP under study.
The above study plan for MSPs is proposed based on experience. As expected, and as will be demonstrated in the MLB examples, the complexity and heterogeneity of any MSP interact with each other in many unknown ways across multiple scales. Recognizing this leads us to adopt Equation (1) by having all the functional structures be completely unspecified and unknown. Without knowing the global functional structure G ( . ) and the component-wise structures { F m ( . ) | m = 1 , , M } , our feature selection protocol can freely discover “major factors of various orders” by being free from man-made structural constraints. It is worth reiterating that this study plan works equally well for all data types.
Being free from structural and data-type constraints is critical and essential to understand and resolve real-world problems reliably. That is, we can not only build understanding of any complex system with a large amount of available data, but also realistically resolve MSPs in their original forms with reliability. Especially in this Big Data era, nearly all researchers admit that no modeling can stand true when faced with large amounts of data. Although the phrase: “All models are wrong, but some are more useful than others”, still rings true, we do have one viable alternative: CEDA-based selection for major factors.
Certainly, such freedom of functional structures and universal validity for all data types come at the cost of facing the “finite sample phenomenon”, since reliable contingency tables can only be constructed with a limited number of covariate features, which mirror the “curse of dimensionality”. Such a cost could result in missing high-order major factors in MSPs. However, all data sets are finite. On the other hand, high-order major factors are harder to confirm. Nevertheless, vital and reliable collections of low- and median-order major factors constitute authentic intelligence when they are fully supported by visible and explainable patterns using contingency table platforms.

4. A Designed MSP Example

In this section, we design an MSP to demonstrate the complicated effects of heterogeneity. It is intuitive that such effects are especially profound when many complex systems are involved within an MSP, since a large number of constituent systems means that data points within each category of response features are spread thinner across all system-IDs, coupled with other covariate features. Consequently, all conditional entropy computations within such an MSP study are prone to the finite sample phenomenon. With such intuition in mind, we design our experimental MSP as follows.
Let the V 1 be a categorical feature with 80 categorical-IDs digital coded as: 1 to 80. These 80 ID-codes are equally divided into eight homogeneous groups defined as follows:
G 1 .
Y = V 2 + s i n ( 2 π × ( V 4 + V 5 + V 6 ) ) / 2 + N ( 0 , 1 ) / 10 ,
G 2 .
Y = V 2 + s i n ( 2 π × ( V 3 + V 5 + V 6 ) ) / 2 + N ( 0 , 1 ) / 10 ,
G 3 .
Y = V 2 + s i n ( 2 π × ( V 3 + V 4 + V 6 ) ) / 2 + N ( 0 , 1 ) / 10 ,
G 4 .
Y = V 2 + s i n ( 2 π × ( V 3 + V 4 + V 5 ) ) / 2 + N ( 0 , 1 ) / 10 ,
G 5 .
Y = V 2 + s i n ( 2 π × U ( 5 ) [ 0 , 1 ] ) / 2 + N ( 0 , 1 ) / 10 ,
G 6 .
Y = V 2 + s i n ( 2 π × U ( 6 ) [ 0 , 1 ] ) / 2 + N ( 0 , 1 ) / 10 ,
G 7 .
Y = V 2 + s i n ( 2 π × U ( 7 ) [ 0 , 1 ] ) / 2 + N ( 0 , 1 ) / 10 ,
G 8 .
Y = V 2 + s i n ( 2 π × U ( 8 ) [ 0 , 1 ] ) / 2 + N ( 0 , 1 ) / 10 ,
where, beside V 1 , all 11 covariate features { V 2 , , V 12 } are mutually independent and distributed according to U n i f o r m [ 0 , 1 ] . There are the four unobserved hidden covariates { U ( k ) [ 0 , 1 ] | k = 5 , , 8 } contained in G 5 through G 8 . That is, the six observed covariate features { V 7 , , V 12 } have nothing to do with the response variable Y and are independent of other covariate features.
For each ID via V 1 , we simulate 500 ( Y , V 2 , , V 12 ) data points. Each group contains 10 IDs, so there are 5000 data points for each of the eight groups. Based on the whole set of 40,000 data points, we compute all conditional entropy of Y given all possible feature subsets of { V 1 , , V 12 } . In particular, we also compute and report the successive conditional entropy (SCE) drops of such feature subsets. Here we reiterate once more, for a feature-set A, its SCE drop is defined as: CE drop of A ( H [ Y ] H [ Y | A ] ) minus the maximum of CE drops of all subsets of A. We report SCEs in Table 1 across one-feature to three-feature settings. All feature-sets within all settings with more than three features do not meet criterion [C1:confirmable]. We summarize the results derived from our major factor selection protocol as follows.
Selection of Major Factors in the Designed MSP
1 
In the one-feature setting of Table 1, V 2 achieves the smallest CE with the largest CE drop: 0.3332 . We surely recognize V 2 as the potential candidate for order-1 major factor. In contrast, V 1 achieves the second ranked CE with a CE drop of 0.0106 , which is significantly less than that of V 2 , but it is almost 10 times the CE drop of any of the rest of the 10 features, including the four features { V 3 , , V 6 } involving G 1 through G 4 , all of which achieve tiny CE drops around 0.001 . In fact, V 1 does not meet criterion [C1: confirmable], since its CE, 2.4231 , is beyond the center of CE distribution of ε (having 80 categories) with mean 2.4229 and s d = 5.2654 × 10 4 . Therefore, V 1 is not a candidate of order-1 major factor. Furthermore, no individual feature of { V 3 , , V 12 } meets the criterion [C1: confirmable].
2 
In the two-feature setting, we can see that the first 11 pairs of the 12 ranked SCEs are made of V 1 coupling with { V 2 , , V 12 } , and their SCEs are all many times larger than their corresponding individual CE drops. Therefore, all these feature-pairs satisfy the criterion [C2: irreplaceable]. However, even the top four pairs, V 1 _ V 3 , V 1 _ V 4 , V 1 _ V 5 and V 1 _ V 6 , do not meet criterion [C1: confirmable]. Through Algorithm 2, the distribution of CEs of ( V 1 , ε ) is found to have a mean of 2.2894 and an SD of 0.0019 . The CE of V 1 _ V 3 , 2.2895 , is slightly larger than the mean. This is partly due to the fact that there are 80 categories of V 1 . Thus, there are no confirmed candidates of order-2 major factors. It is notable that the pair ( V 1 , V 2 ) achieves a SCE of 0.1035 , which is about 10 times as large as the CE drop achieved by V 1 . That is, by satisfying the ecological effect, V 2 and V 1 can be order-1 major factors simultaneously.
3 
In the three-feature setting, the top six feature triplets are V 1 coupling any pairs of { V 3 , V 4 , V 5 , V 6 } . They achieve SCEs that satisfy criterion [C2: irreplaceable], but they do not fulfill the criterion [C1: confirmable]. The CEs of these triplets, which range between 1.2146 and 1.2198 , are all larger than the mean ( 1.2051 ) of the distribution of the CEs of ( V 1 , V 3 , ε ) . Therefore, there are no confirmed candidates for order-3 major factors.
4 
In the four-feature setting, the four feature quartets, V 1 coupling any triplets of { V 3 , V 4 , V 5 , V 6 } , do not meet the criterion [C1: confirmable].
From the above summary, it becomes evident that the effects of heterogeneity of V 1 permeate and blur candidacies of order-2 and order-3 major factors. Such blurred pictures are primarily due to the finite sample phenomenon. That is, the heterogeneity causes the loss of full views of the true underlying mechanisms. Accordingly, based on results from our major factor selection protocol, we conclude that the chief collection of major factors is { V 2 } and there are no alternative collections.
Next, we proceed with the decomposition of the whole data set’s heterogeneity into homogeneous parts. Indeed, as shown in Figure 1, G1 to G8 are correctly identified on a heatmap clustering 80IDs’ vectors of means pertaining to the 11 features. When the exclusive data of G 1 through G 8 are analyzed separately, we can precisely confirm major factors of various orders underlying each of the eight groups. The order-1 major factor V 2 is confirmed, but no candidates of order-2 major factors are confirmed across the eight groups. As for potential candidate order-3 major factors, for instance in G 1 , the feature-triplet ( V 4 , V 5 , V 6 ) achieves a CE of 1.0432, satisfying the criterion [C2: irreplaceable], and meets [C1: confirmable] in the following fashion. By applying Algorithm 2, this triplet’s CE is more than 10 SD below the CE distribution of ( V 4 , V 5 , ε ) with mean 1.1094 and SD 0.0058 , where ε is a uniform noise feature. Likewise, we confirm the order-3 major factors in G 2 through G 4 , respectively. We summarize the results of our major factor selection within this designed MSP in Table 2. It is noted that the results of G 6 , G 7 , and G 8 are identical to those of G 5 .

5. MLB’s Two MSPs: Sliders and Fastballs

In this section, we study Major League Baseball (MLB) pitching dynamics from an MSP perspective.
MLB owns two databases, namely PITCHf/x and Statcast, that record every pitch delivered in all MLB games since 2006. In this paper, we study two large real-world MSPs of slider and fastball pitching dynamics in the 2017 season. From the PITCHf/x database, we select all pitchers who have pitched more than 500 sliders or fastballs, respectively. There are 62 slider pitchers and 199 fastball pitchers that meet this 500 pitches criterion. The pitch-ID is denoted as p i t N .
In a structured data format, biomechanical and physical features include a pitched baseball’s releasing coordinates { x 0 , z 0 } , speeds { v X 0 , v Y 0 , v Z 0 , } , and accelerations { a X , a Y , a Z } along the horizontal ( X ), vertical ( Z ), and pitcher-to-catcher ( Y ) directions. The feature of starting speed denoted by s t a r t S p is very close to v Y 0 . Two features are measured for each pitch’s spin information: spin direction s p i n D and spin rate s p i n R . The aforementioned features are the covariate features of pitching dynamics measured around the pitcher mound, while the two features measured around the home plate, horizontal and vertical movements denoted by p f x x and p f x z , are designated as a 2D response variable Y = ( p f x x , p f x z ) . In both pitch-type specific MSPs, each pitching dynamic of ( p f x x , p f x z ) is found to be dominated by ( a X , a Z ) under the presence of p i t N . Therefore, we use ( a X , a Z ) = Y as a 2D response variable against the remaining covariate features to further explore the effects of p i t N . All aforementioned features, except p i t N , are quantitative and undergo categorization. All categorized features retain the same names throughout this paper.
MLB pitching dynamics are basically governed by biomechanical forces in the X , Y , and Z directions, coupled with the Magnus effect of the “spinning baseball” created by the friction between the pitcher’s fingers and the baseball surface, and the gravity of earth. The Magnus effect is a 3D force constantly acting on the baseball with a 2D direction on the 2D disk that is perpendicular to the baseball’s trajectory at all time. Its 2D direction is measured by spin direction ( s p i n D ) , while its stability is determined by spin rate ( s p i n R ) and various directional speeds, accelerations, and other elements. It is not an easy task to clearly separate the force due to the Magnus effect from the biomechanical ones in pitching dynamics.
Different pitch types require different kinds of spinning. Back-spinning in fastballs persistently shows four or two seams rotating backward when viewed from the catcher or batter’s perspective. This is a natural type of spin that can be easily created. Its Magnus effect points upward against gravity. The top-spinning in curveballs shows seams rotating forward instead. Its Magnus effect points downward to the ground and adds to gravity. Sliders are slower than fastballs, but much faster than curveballs in speed. Their spinning direction has a much wider range than fastballs and the curveballs. That is, the Magnus effect makes sliders have wider ranges of vertical movement (measured and denoted by p f x Z ) and horizontal movement (measured and denoted by p f x X ) than that of fastballs and curveballs.
These two measurements measured around the home plate are critical to pitchers when trying to effectively deal with batters. Therefore, we first designate ( p f x Z , p f x X ) as the 2D response variable. Then, the MSP of interest can be very precisely depicted as the fundamental question: what are the major factors underlying ( p f x Z , p f x X ) among many pitchers? This description does not mean that the identified collection of major factors would manifest and separate Magnus effects and biomechanical forces in a clear-cut fashion. On the contrary, the collective of major factors only reveals major factors that most directly cause the designated response variable. That is, major factors are very sensitive to the choice of response variables.
For the two focal pitch types in the 2017 MLB season, there are 62 slider pitchers and 199 fastball pitchers who pass the 500 pitches threshold. We have a 2D response variable Y = ( p f x Z , p f x X ) and 12 covariate features that all are measured around the pitcher’s mound.

5.1. Sliders MSP

A heatmap and a network of MCE-based associative patterns of these 14 features, responses, and covariates are given in panels (A) and (B) of Figure 2. It is clear that feature-pairs ( p f x X , a X ) and ( p f x Z , a Z ) are highly associative, while a X and a Z are not. Based on the heatmap and network, we clearly see diverse associative patterns among these 14 features.
We computed conditional entropies for all possible combinations of 12 covariate features. The resultant table is too large to be included entirely. Thus, we report only top-ranked CEs up to five-feature setting in Table 3 and SCE drops up to four-feature setting in Table 4. These two formats of tables are used throughout the three MLB-related MSPs for expositional efficiency of showing potential candidates of major factors. However, our search for feature selection of major factors is conducted based on all possible feature-sets. We summarize patterns from our feature selection protocol and list our reasoning based on the criteria [C1:confirmable] and [C2:irreplaceable] across all feature settings. The CE of Y is calculated as 4.6890.
1: 
In the one-feature setting, we see four features— { s p i n D , a X , p i t N , a Z } —that achieve CE drops more than 1.1000. They are potential candidates of order-1 major factors. Among these four features, { s p i n D , a X , p i t N } are moderately associated based on Figure 2A,B. This fact would significantly reduce their potentials to simultaneously be order-1 major factors.
2: 
Considering the two-feature setting, the feature-pair ( a X , a Z ) achieves the lowest CE, 0.9264, and meets the criterion [C1: confirmable], see Figure 3A. However, its CE drop, 3.7626, is almost equal to the sum of individual CE drops of a X and a Z : 3.8032. Even though, based on Figure 2A,B, a X and a Z are not moderately dependent to each other, this observation of I [ a X ; a Z | Y ] I [ a X ; a Z ] strongly indicates that both can not be order-1 major factors simultaneously. Certainly, the feature-pair ( a X , a Z ) is not an order-2 major factor because of the criterion [C2:irreplaceable]. In fact, the rest of five feature-pairs derived from { s p i n D , a X , p i t N , a Z } share the same situations, such as ( a X , a Z ) .
Four feature-pairs ( a Y , p i t N ) , ( z 0 , p i t N ) , ( s p i n R , p i t N ) and ( v Z 0 , p i t N ) , which are formed by coupling p i t N with one member of feature-set { a Y , v Z 0 , s p i n R , z 0 } , are potential candidates for order-2 major factors. In fact, they meet the criterion [C1:confirmable], and their SCE drops from p i t N , respectively, are six and three times the individual CE drops of a Y , z 0 , s p i n R , and v Z 0 . We also find the 10 pairs { a Y , z 0 } × { x 0 , v Y 0 , v X 0 , v Z 0 , s p i n R } and ( a Y , z 0 ) are also candidates of order-2 major factors, satisfying the two criteria. Here we use the threshold of “3 times” in condition (a) of criterion [C2:irreplaceable], while the two pairs ( s p i n R , v Y 0 ) and ( s p i n R , s t a r t S p ) are excluded, since their SCE drops are 0.3788 and 0.3772 from either v Y 0 or s p a r t S p , which are at most 2.5 times the individual CE drop of s p i n R .
If we accept any of these 15 as order-2 major factors, then we consequently decide which one { s p i n D , a X , p i t N , a Z } is an order-1 major factor subject to the criterion [C2:irreplaceable]. For example, if we pick ( a Y , v Z 0 ) as an order-2 major factor, then only a X , a Z , and p i t N are qualified to be order-1 major factors. The collections { a X , ( a Y , v Z 0 ) } and { a Z , ( a Y , v Z 0 ) } achieve CE values 2.2452 and 2.3466, respectively. Both collections meet the [C1:confirmable] criterion.
In contrast, it is interesting and important to note that, though the collection { p i t N , ( a Y , v Z 0 ) } achieves a CE value 1.9449, which is smaller than the aforementioned collections, it does not meet the [C1:confirmable] criterion. By applying Algorithm 2, the simulated CE distribution of ( p i t N , v Z 0 , ξ ) has a mean value 1.7586 and an SD of 1.9830 × 10 3 . The underlying reason is the finite sample phenomenon caused by the 62 categories of p i t N . The effect of this finite sample phenomenon becomes less when the ensemble of 62 pitchers is partitioned into groups.
If we pick ( v Y 0 , z 0 ) as an order-2 major factor, then only a Z and p i t N are qualified to be order-1 major factors. The collection { a Z , ( v Y 0 , z 0 ) } achieves a CE value 2.1633, while the collection { p i t N , ( v Y 0 , z 0 ) } achieves a CE value 2.6462.
If we pick either ( v Z 0 , z 0 ) or ( s p i n R , z 0 ) as an order-2 major factor, then a X , a Z , p i t N , and s p i n D are qualified to be order-1 major factors. The two collections { a X , ( v Z 0 , z 0 ) } and { a Z , ( s p i n R , z 0 ) } , respectively, achieve the smallest CE values of 2.2224 and 2.2534 within the two packs of the four collections.
3: 
In the 3-feature setting, the top 10 feature triplets have achieved rather uniform CEs, while the 12 ranked triplets have rather uniform SCE drops. Such uniformness is likely a sign of the finite sample phenomenon. Hence, it becomes necessary to test whether our feature selection should stop before the three-feature setting. This testing is first performed on the triplet ( a Z , v Y 0 , a X ) . It indeed meets the [C1:confirmable] criterion; see panel (B) of Figure 3. However, it is far from meeting the criterion [C2:irreplaceable].
As for triplet ( a Y , v Z 0 , p i t N ) , it achieves the largest SCE drop and seemingly has potential to be order-3 major factor. However, it is not, due to the criterion [C2:irreplaceable]. The reason is this is that triplet ( a Y , v Z 0 , p i t N ) is a union of ( a Y , p i t N ) and ( v Z 0 , p i t N ) with individual CE drops of 1.5820 and 1.5908, respectively. The sum of both CE drops, 3.1728, is much larger than the CE drop of the triplet: 2.7441 ( = 4.6890 1.9449 ) . Despite this, as indicated above, the collection { p i t N , ( a Y , v Z 0 ) } is a valid one. That is, it takes much more to become a high order major factor.
In contrast, the remaining 11 triplets achieve rank 2 to 12 successive CE drops in Table 4, such as { a Y , v Z 0 , z 0 } , do not meet the criterion [C1:confirmable]; see Figure 3C. In conclusion, there are no order-3 major factors confirmed.
4: 
In four-feature setting, the top two quartets ( a Z , a Y , a X , p i t N ) and ( a Z , v Z 0 , a X , p i t N ) achieve the lowest CE. They consist of three potential order-1 major factors, a X , a Z , p i t N , so they do not meet the criterion [C2:irreplaceable]. Similar reasoning is applicable to the quartets ( a Z , v Y 0 , a X , p i t N ) , ( a Z , s t a r t S p , a X , p i t N ) , ( a Z , a Y , a X , s t a r t S p ) , and ( a Z , z 0 , a X , v Y 0 ) .
5: 
In the 5-, 6- and 7-feature settings, they fail to meet the criterion [C1:confirmable] due to the finite sample phenomenon. For instance, if we collect members of confirmed potential major factors of all orders, we arrive at a seven-feature set ( a Z , a X , v Y 0 , v Z 0 , p i t N , s p i n R , a Y ) . THis indeed achieves the lowest CE in the seven-feature setting. However, it does not meet the criterion [C1:confirmable], since, according to Algorithm 2, its CE of 0.0236 is far beyond the range of the simulated CE distribution of ( a Z , a X , v Y 0 , v Z 0 , p i t N , s p i n R , ξ ) , with mean 8.9602 × 10 3 and SD 5.8219 × 10 4 .
From the above feature-selection, we identify { a Z , ( v Y 0 , z 0 ) } as the chief collection of major factors of Y = ( p f x X , p f x Z ) , and two alternative collections: { a X , ( v Z 0 , z 0 ) } and { a X , ( v Z 0 , a Y ) } . These three collections contain one order-1 and one order-2 major factor. We found no order-3 or higher orders. Each order-1 major factor, either a Z or a X , within any of these three collections contributes significantly more than its corresponding order-2 major factor. If the fourth alternative collection is needed, then it would be { a Z , ( s p i n R , z 0 ) } .
It is also evident that the effect of heterogeneity is not explicitly present because of p i t N is not selected as an order-1 major factor, while the Magnus effect is also not visible, since s p i n D is not selected as one of order-1 major factors in the three selected collections. However, collections such as { s p i n D , ( v Z 0 , z 0 ) } and { s p i n D , ( s p i n R , z 0 ) } are not selected because their CE values are much higher than the selected ones.
Finally, all these collections of major factors of Y = ( p f x X , p f x Z ) involve two features of four aspects of the pitching dynamics: (1) the acceleration in the Y-direction ( a Y ), (2) the vertical releasing speed ( v Z 0 ), (3) spin rate s p i n R , (4) the vertical coordinate of the releasing point z 0 . This fact collectively indicates that our understanding of the underlying dynamics of Y = ( p f x X , p f x Z ) should be derived from multiple perspectives. It is also evident that the effect of s p i n R is not ignorable; nonetheless, it is definitely not the primary effect.
On the other hand, a X and a Z seem to play key roles in Y = ( p f x X , p f x Z ) . This pair achieves a CE drop of 3.7626 from a CE of Y : 4.6890. In contrast, the chief and alternative collections of major factors of Y = ( p f x X , p f x Z ) have CE drops of about or less than 2.6890. That is, we need to further investigate by using Y = ( a X , a Z ) for the 25 % unexplained CE beyond our collections of major factors.
Another motivation for looking at the dynamics of Y = ( a X , a Z ) is the fact that the significance of the Magnus effect is not evidently reflected in the collections of identified major factors of Y = ( p f x X , p f x Z ) . In fact, the spin direction ( s p i n D ) is not a primary member of the identified collections of major factors. Obvious reasons might be that this 2D response variable is highly and directly associated with a X , a Z and the heterogeneous effect via p i t N . That is, the Magnus effect and the heterogeneity effect have likely been overshadowed by a X and a Z . To further investigate this speculation, we need to conduct another study via feature selection for major factors of response variable Y = ( a X , a Z ) .
There are 10 covariate features for this investigation in the same MSP of 62 slider pitchers. Likewise we compute all CEs for all possible combinations of 10 features. We report up top 10 CE values across five feature-settings in Table 5 and the top 10 SCE drops in Table 6 across four feature-settings. Computational results of feature selection for major factors are summarized below starting from one-feature setting to five-feature setting. The CE of Y is 4.7213.
1 
In the one-feature setting, s p i n D and p i t N achieve the top two lowest CEs. Their CE drops, 1.6912 and 1.1127, respectively, are many time larger than the rest of eight features’ CE drops. However, we do not expect them to become two separate order-1 major factors due to their moderate mutual association. This can be seen through the fact that the feature-pair ( s p i n D , p i t N ) achieves a SCE drop of 0.5410 from s p i n D , which is less than half the CE drop of p i t N . Again, the selection of an order-1 major factor is highly dependent on the selection of order-2 major factors.
2 
Considering the two-feature setting, the three feature-pairs: p i t N coupled with one member of feature-set { a Y , v Z 0 , z 0 } have SCE drops at least three times the CE drop of a Y , z 0 , and v Z 0 , respectively. They satisfy the criterion [C2: irreplaceable] and meet the criterion [C1:confirmable]. We also have 10 pairs, { a Y , z 0 } × { x 0 , v Y 0 , v X 0 , v Z 0 , s p i n R } , and ( a Y , z 0 ) , that fulfill the two criteria. They are potential candidates for order-2 major factors. Any of them could become an order-2 major factors depending on which order-1 major factor is selected.
With respect to all order-1 and order-2 candidates for major factors, we select the collection { s p i n D , ( v Z 0 , z 0 ) } as the chief collection of major factors of Y = ( a X , a Z ) . We also select two alternative collections: { s p i n D , ( a Y , v Z 0 ) } and { s p i n D , ( a Y , z 0 ) } . All candidate collections involve p i t N , and the aforementioned candidates of order-2 major factors can not meet the criterion [C1:confirmable] under the three-feature setting.
3 
In the three-feature setting, all top 10 triplets of CEs involve p i t N , while seven out of ten involve the s p i n D . These 10 triplets are not the top 10 of CE drops and they achieve rather uniform CEs. According to Algorithm 2, they all fail to meet the criterion [C1:confirmable]. For example, the triplet ( v Z 0 , p i t N , a Y ) achieves the lowest CE of 1.9577 in this three-feature setting. This CE value is far beyond the range of the simulated CE distribution of ( v Z 0 , p i t N , ξ ) , with mean 1.7702 and SD 1.9173 × 10 3 .
4 
Regarding the four- and five-settings, they achieve rather uniform CEs. This is a sign of the effect of the finite sample phenomenon. That is, based on Algorithm 2, they all fail to meet the criterion [C1:confirmable]. For example, ( v Z 0 , s p i n R , p i t N , a Y ) has a CE of 0.9624, far beyond the range of the simulated CE distribution of ( v Z 0 , s p i n R , p i t N , ξ ) , with mean 0.7718 and SD 1.172166 × 10 3 . Furthermore, ( v Y 0 , v Z 0 , s p i n R , p i t N , a Y ) has a CE of 0.3837, far beyond the range of the simulated CE distribution of ( v Y 0 , v Z 0 , s p i n R , p i t N , ξ ) , with mean 0.2649 and SD 7.378123 × 10 4 .
Our conclusion in this investigation where Y = ( a x , a z ) is that the fact that the three collections have s p i n D as the order-1 major factor is rather natural based on general knowledge of pitching dynamics. The three members of order-2 major factors are pairs from { z 0 , a Y , v Z 0 } . That is, the biomechanical features do play some important roles underlying the dynamics of Y = ( a x , a z ) . Further, the effects of heterogeneity within this collective of 62 slider pitchers systems seem evidently overshadowed by the finite sample phenomenon in the three-feature setting and beyond, though these effects are reflected through presences of multiple candidates of order-2 major factors.

5.2. Fastballs: 199 Pitchers

The MSP of fastballs in the 2017 MLB season consists of 199 pitchers with 14 features. In the two panels of Figure 4, the heatmap and the network based on the 14 × 14 MCE matrix reveal again that feature-pairs ( p f x X , a X ) and ( p f x Z , a Z ) are highly associative, but that the pair ( a X , a Z ) is not.
Again, we first consider the 2D bivariate response variable Y = ( p f x X , p f x Z ) . The CE of Y is calculated as 4.7167. According to the two criteria [C1: confirmable] and [C2: irreplaceable], we summarize our feature selection for major factors of Y = ( p f x X , p f x Z ) based on the CEs of all the possible feature-sets from the 12 covariate features. Only the top ranked CEs and SCE drops are reported in Table 7 and Table 8.
1: 
With regard to the one-feature setting, we see four features, { s p i n D , a X , p i t N , a Z } , that achieve CE drops larger than 1.2000. They are clearly potential candidates for order-1 major factors, since s p i n D and a X are highly associated and both are moderately associated with p i t N . Such associative relations among these three features keeps them from being order-1 major factors simultaneously. In contrast, a Z is a strong potential candidate for an order-1 major factor.
2: 
In the two-feature setting, the feature-pair ( a X , a Z ) achieves the lowest CE of 1.7424, with a CE drop of 2.9763, which is almost equal to the sum of the individual CE drops of a X and a Z : 2.9197; that is, I [ a X ; a Z | Y ] I [ a X ; a Z ] . Together with the fact that, based on Figure 4A,B, a X and a Z are independent; only one of them can be an order-1 major factor in a collection of major factors. We also identify six feature-pairs, ( a Z , v Y 0 ) , ( a Y , v Y 0 ) , ( z 0 , v Y 0 ) , ( v Z 0 , p i t N ) , ( v Y 0 , p i t N ) , and ( a Y , p i t N ) , that satisfy the two criteria [C1: confirmable] and [C2: irreplaceable].
Therefore, we select the chief collection { a X , ( v Z 0 , a Z ) } because it achieves the lowest CE value in the three-feature setting, and just barely satisfies the criterion [C2: irreplaceable]. Two alternative collections are: { a Z , ( v Y 0 , a Y ) } and { a Z , ( v Y 0 , z 0 ) } . On the other hand, collections { a Z , ( v Z 0 , p i t N ) } , { a Z , ( v Y 0 , p i t N ) } , and { a Z , ( a Y , p i t N ) } fail to meet the criterion [C1: confirmable].
3: 
In the three-feature setting, the triplets ( a X , a Z , v Y 0 ) and ( a X , a Z , s t a r t S p ) (see Figure 5A pass the test of [C1:confirmable]. They achieve the top two ranked CEs with their CE drops of 0.6536 and 0.6378, respectively, from feature-pair ( a X , a Z ) . On the other hand, if we take ( a Z , v Y 0 ) as a major factor, then triplet ( a X , a Z , v Y 0 ) is a union of two major factors: a X and ( a Z , v Y 0 ) . The SCE drop of this triplet from ( a Z , v Y 0 ) is calculated as 1.8724, which is very close to the CE drop 1.7232 of a X . That is, these two triplets do not meet the [C2:irreplaceable] criterion, so these two triplets are not order-3 major factors.
The largest CE drop among all feature-triplets is achieved by ( a Y , v Z 0 , p i t N ) . Its successive CE drop from ( a Y , p i t N ) is 1.0297 ( = 3.0768 2.0471 ) , which is very significant compared to the CE drop of v Z 0 . However, if we take both ( a Y , p i t N ) and ( v Z 0 , p i t N ) as two order-2 major factors, then triplet ( a Y , v Z 0 , p i t N ) is a union of ( a Y , p i t N ) and ( v Z 0 , p i t N ) . The CE drop of ( a Y , v Z 0 , p i t N ) is calculated as 2.6697, which is less than the sum of the two pairs’ individual CE drops: 3.2402 ( = 1.6399 + 1.6003 ) . For the same arguments, the following four feature-triplets: ( a Y , p i t N , s p i n R ) , ( v Z 0 , p i t N , s p i n R ) , ( a Y , p i t N , s a r t S p ) , and ( v Z 0 , p i t N , s t a r t S p ) , are not potential order-3 major factors. These five triplets do not pass meet the criterion [C1:confirmable]; see Figure 5B.
4: 
If we consider the four-feature-setting, all top 12 quartets on CEs involve p i t N , while all 12 quartets of top CE drops do not involve p i t N , but 10 out of 12 involve s p i n R . The 12 quartets on the top CE list all fail to meet the [C2:irreplaceable] criterion, primarily because they contain multiple major factors: either of the two order-1 or three order-2 or four order-3 major factors if we replace features with their highly associated ones, such as s t a r t S p for v Y 0 and s p i n D for a X , etc. All quartets on the top list of CE drops also fail to meet the criterion [C1:confirmable] by applying Algorithm 2.
5: 
Five-feature and greater feature settings share the same characteristics with feature-sets in the four-feature setting. They all fail to meet either criterion [C1:confirmable] via Algorithm 2 or criterion [C2: irreplaceable] for the same reasons.
For the response variable Y = ( p f x X , p f x Z ) , we can confidently identify the following three collections of major factors: { a X , ( a Z , v Y 0 ) } , { a Z , ( v Y 0 , a Y ) } , and { a Z , ( v Y 0 , z 0 ) } . We clearly see the effect of heterogeneity, through p i t N is overshadowed by the finite sample phenomenon in the three-feature setting within this MSP of 199 fastball pitchers. No order-3 or higher order major factors are confirmed. Again, the joint effects a X and a Z still overshadow the Magnus effect and the effects of heterogeneity. Therefore, we deepen our investigation with the response variable Y = ( a X , a Z ) .
Likewise we report top ranked CEs and SCE drops in two tables, Table 9 and Table 10, respectively: that is, our results of feature selection based on CEs of all possible feature-sets among the 10 covariate features.
1: 
In the one-feature setting, s p i n D and p i t N achieve top two ranked CE values. It is not surprising that s p i n D plays the sole role of order-1 major factor in the identified collections of major factors. It is surprising that, as seen below, we need to bring in a Y and v Z 0 , two separate order-1 major factors, for the sake of improving the CE performances beyond s p i n D , to construct collections of major factors.
2: 
In the two-feature setting, the top nine feature pairs achieving the lowest CEs involve s p i n D . However, none of these nine pairs fulfill the criterion [C2: irreplaceable]. On the other hand, the feature-pair ( p i t N , a Y ) and ( p i t N , v Z 0 ) ranked second and third on the list of top 10 SCE drops, both fail to meet the criterion [C2: irreplaceable], while ( p i t N , s p i n R ) satisfies both criteria. We also report no candidates for order-2 major factors that do not include p i t N . This is a rather surprising observation.
However, collections { v Z 0 , ( s p i n R , p i t N ) } and { a Y , ( s p i n R , p i t N ) } fail to satisfy the criterion [C1: confirmable]. As such, these collections evidently reflect the effects of heterogeneity being overshadowed by the finite sample phenomenon.
3: 
Considering the three-feature setting, the top 10 feature triplets on CEs are primarily extended from feature pair ( s p i n D , p i t N ) . This is not surprising due to the dominant effect of s p i n D . They satisfy the criterion [C1: confirmable] (see panel (C) of Figure 5) but not [C2: irreplaceable]. In contrast, the top 10 feature triplets on SCE drops are all involved p i t N . They satisfy the criterion [C2: irreplaceable], but not [C1: confirmable].
4: 
As for the four-feature and greater feature settings, they all fail to meet the criterion [C1:confirmable] based on Algorithm 2. The same example as in sliders, ( v Z 0 , s p i n R , p i t N , a Y ) , has its CE of 0.9875 far beyond the range of the simulated CE distribution of ( v Z 0 , s p i n R , p i t N , ξ ) , with mean 0.7949 and SD 6.4739 × 10 4 . Furthermore, ( v Y 0 , v Z 0 , s p i n R , p i t N , a Y ) has its CE of 0.3918 far beyond the range of the simulated CE distribution of ( v Y 0 , v Z 0 , s p i n R , p i t N , ξ ) , with mean 0.2619 and SD 3.7392 × 10 4 .
The identified three collections of major factors of Y = ( a X , a Z ) are then (1) { s p i n D , a Y } , (2) { s p i n D , v Z 0 } , and (3) { s p i n D } . We cannot find any collections of major factors consisting of three features.

5.3. Comparisons of Summarized Results from Slider and Fastball

We collect and compare major identified factors of various orders via our feature selection protocol on both pitch types, sliders and fastballs, with respect to two different kinds of response variables, Y : ( p f x X , p f x Z ) and ( a X , a Z ) , in Table 11. Such comparisons reveal fundamental differences of these two pitching dynamics, and at the same time, shed light on the effects of heterogeneity within both MSPs.
When considering Y = ( p f x X , p f x Z ) as the response variable, both pitch types select three collections of triplets in an identical format: one order-1 major factor, either a X or a Z , and one biomechanical order-2 major factor. Newton’s second law of force dictates that a X and a Z as directional forces govern the two directional horizontal and vertical movements: Y = ( p f x X , p f x Z ) . In both pitch types, we also see the heterogeneity effects via p i t N commonly interacting with biomechanical features in a format of candidates of order-2 major factors such as, ( a Y , p i t N ) , ( z 0 , p i t N ) , ( s p i n R , p i t N ) , and ( v Z 0 , p i t N ) . From a biomechanical perspective, we find that slight differences rest on the candidates for order-2 major factors: ( s p i n R , v Y 0 ) for sliders, and ( a Z , v Y 0 ) and ( p i t N , v Y 0 ) for fastballs.
In other words, their differences are primarily tied to “speed” ( v Y 0 or s t a r t S p ). The coupling effect of ( s p i n R , v Y 0 ) might be closely related to the capacity of control of slider pitches. As for fastballs, the speed ( v Y 0 or s t a r t S p ) is a basic requirement. It is surprising to see that the coupling effects of speed v Y 0 and vertical acceleration a Z on Y = ( p f x X , p f x Z ) are somehow critical in this MSP of fastball pitching dynamics. The candidacy of order-2 major factor ( p i t N , v Y 0 ) indicates that there are significant differences among the 199 fastball pitchers, while this is not the case in sliders.
When Y = ( a X , a Z ) is the response variable in both pitch types, we found three triplet-based collections of major factors in the sliders, but only two pair-based and one singleton-based collections of major factors in the fastballs. The feature s p i n D is apparently an order-1 major factor for Y = ( a X , a Z ) in both pitch types.This clearly points to the fact that the Magnus effect is primarily contained in horizontal and vertical accelerations ( a X , a Z ) .
The effects of heterogeneity among 62 slider pitchers and 199 fastball pitchers are manifested through the same four candidates of order-2 major factors: p i t N coupled with one biomechanical member of feature-set { a Y , v Z 0 , s p i n R , z 0 } . The impacts of such heterogeneity effects via p i t N seem much more severe in fastballs than in sliders.
Further, the presences of ( s p i n R , v X 0 ) , ( s p i n R , v Y 0 ) as candidates for order-2 major factors in sliders, but not in fastballs, imply characteristic differences between these two pitch types. These two pairs clearly reveal the effects of s p i n R by coupling the two speed features v X 0 and v Y 0 , which are commonly shared by all 62 slider pitchers. Therefore, s p i n R is most likely linked to better control and reliable stability of slider pitches. In contrast, s p i n R only plays a minor role in fastballs through one aspect of interacting relations with p i t N .
Here, we reiterate that p i t N appears in candidates of order-2 major factors by coupling wirh primary biomechanical features under Y = ( p f x X , p f x Z ) and Y = ( a X , a Z ) . This evidence strongly indicates that the heterogeneity in these two MSPs must be accommodated before performing any inferences on Y = ( p f x X , p f x Z ) . This is one of the chief characteristics of MSPs.

5.4. Taking Care of Heterogeneity

From the results reported above, it is essential to recognize that p i t N neither ever stands alone as an order-1, nor is a member of identified order-2 major factors. Its effects are always revealed through interacting relations with biomechanical and physical features due to it being a candidates for an order-2 major factor. Such interacting relations severely constrain our understanding of the pitching dynamics of both pitch types. In fact, such interacting relations reflect p i t N ’s diverse associations with a majority of the biomechanical and physical features. As such, these relations are pieces of information on heterogeneity revealed in many feature-specific aspects. What about global information on heterogeneity regarding similarity and dissimilarity among pitchers in both MSPs?
With such global information in mind, it becomes necessary to correspondingly partition the whole collection of pitchers into homogeneous pitcher-groups. By doing so, we want to discover more diverse aspects of pitching dynamics, and at the same time bring out the global distinctions across various homogeneous groups.
For the two MSPs of slider pitchers and fastball pitchers, we propose here to perform hierarchical clustering (HC) on the two collections of pitchers with respect to contingency tables of triplet ( a X , a Z , v Y 0 ) against p i t N . These three features in the triplet involve order-1 and order-2 major factors of Y = ( p f x X , p f x Z ) . The two resulting HC-trees are given in Figure 6. Each HC-tree is marked with various scales of composition of pitcher-groups.
For the sliders, we first consider partitioning the 62 pitchers into five branch-based groups indexed A to E. We designate each branch-based group as one MSP and check whether it embraces homogeneity or not. For such a goal in each branch-based MSP, we conduct feature selection for major factors of the response variable Y = ( a X , a Z ) . We summarize our findings of major factors of these five MSPs in Table 12.
In this table, when checking the criterion [C2: irreplaceable], we make sure that the SCE drop is at least three times that of Condition (a) for any candidate feature-sets. To test for the criterion [C1:confirmable], we use Algorithm 2. The selected collections across these ive MSPs share s p i n D as the chief order-1 major factor. Then s p i n D is coupled with various biomechanical features as order-1 major factors with relatively minor effects.
We see that p i t N appears as one member of the candidates of order-2 major factors of Y = ( a X , a Z ) in MSPs defined by branches B, D and E, but not by branches A and C. That is, the branch A- and C- based MSPs embrace homogeneity among pitcher-systems, while branch B-, D-, and E-based MSPs still embrace heterogeneity to a much less extent because the number of such candidates is just two. It is also noted that throughout these five MSPs, p i t N achieves the top 2 ranked individual CE drops, which are significantly smaller than the CE drops of the order-1 major factor s p i n D . We also observe that the SCE drop of ( s p i n D , p i t N ) is rather small. Therefore, p i t N is ruled out as a candidate for an order-1 major factor.
Another essential visible pattern in Table 12 is that the candidates for order-2 major factors of Y = ( a X , a Z ) become abundant and diverse by involving only biomechanical features in these 5 MSPs. This phenomenon is different from the interacting relations between p i t N and biomechanical features in the original MSP consisting of 62 slider pitchers. This fact reflects P.W. Anderson’s “More is Different” perspective of MSPs; its homogeneous components likely reveal more diverse and detailed group-idiosyncratic characteristics.
For fastballs, we first partition the 199 pitchers into three large groups with respect to the three major branches marked and indexed by A, B, and C, in panel (B) of Figure 6. These three large branches are unlikely to embrace homogeneity. Hence, we further partition the C branch into six sub branches marked and indexed by {C1, C2, …, C6}. These six sub branch-based MSPs seemingly embrace homogeneity, as seen in Table 13. That is, all their major factors of Y = ( a X , a Z ) in the six MSPs do not explicitly contain p i t N . Throughout these six MSPs, p i t N achieves an individual CE drop that is significantly smaller than the CE drop of s p i n D .
The MSP C1 does not support any triplets because of the criterion [C1:confirmable]. Only two collections of order-1 major factors are identified within this MSP. Three collections consist of s p i n D coupling with one of the members of { v Z 0 , a Y , v X 0 } . The fourth collection is { s p i n D , s p i n R } .
The MSP C2 does support one collection of triplets in the format of one order-1 major factor and one order-2 major factor. The other two collections consist of s p i n D coupled with another less effective order-1 major factor such as v Z 0 and a Y . There are 13 candidates for order-2 major factors identified, but not a single pair contains p i t N as a member. Likewise, in MSP C3, there are two collections of triplets: one order-1 major factor and one order-2 major factor. The alternative is a collection of two order-1 major factors: { s p i n D , a Y } . There are also 12 candidates of order-2 major factors identified without p i t N as a member. In a similar fashion, in MSP C5 there are three collections of triplets supported by the data and there are 13 candidates for order-2 that do not include p i t N . In MSP C6, identified collections are similar to those found in MSP C3.
In MSP C4, the first two selected collections of triplets share an unusual format: two order-2 major factors. The members of these two triplets are all biomechanical features without s p i n D . The third collection is { s p i n D , a Y } . There are no other candidates for order-2 major factors.
Overall, we identify abundant candidates for order-2 major factors that do not involve p i t N at all across the six MSPs. In other words, the effect of heterogeneity seems to have disappeared.
Up to this point, our slider and fastball examples strongly suggest that a rigorous way for analyzing any MSPs containing a large number of complex systems must take the following steps. Firstly, we explore and understand how the complexity of the underlying dynamics interacts with heterogeneity. Secondly, this exploration leads us to find a proper clustering and grouping scheme, such that similar members of the original collection of systems are grouped together in order to embrace homogeneity. Thirdly, when such homogeneity is achieved across all subdivided groups, a more diverse and finer scale group-specific complexity can be discovered. This is because a group of similar systems will collectively retain a much larger number of data points than any of its single member systems. This approach for analyzing any large MSPs by going from the heterogeneity of the whole to the diverse homogeneity of its constituents is a way of seeing and understanding “More is Different” in MSPs.

6. Self-Esteem across Gender and Age Human Complex Systems

In the very last section of this paper, we explore a potential merit of MSP study in a research area where psychology, psychiatry, and sociology intersect. The scientific question can be simply stated as: do people of different genders and ages respond to the Rosenberg Self-Esteem Scale differently? Prof. Morris Rosenberg studied dynamic changes of late adolescents’ self-image in his well known 1965 book [10]. Boys and girls of 15 to 18 years of age, the so-called late adolescence, must make drastic self-image changes in response to drastic physiological and psychological developments due to changes in their own bodies, as well as the different societal potentials being discovered and the many career decisions that must be made. In order to study this topic, he designed the popular Rosenberg Self-Esteem Scale with 10 questions for his study subjects.
More than half of a century has passed since the publication of his book. Intuitively speaking, human self-image might be shaped, for example, by the abundant information available on the Internet. In fact, nowadays, all age and gender groups are exposed to and share a huge amount of information in regard to our world’s drastic environmental and technological changes. Do such visible world-wide changes in our living conditions motivate and affect people’s self-image differently or similarly?
Nowadays, the Rosenberg Self-Esteem Scale has become a popular on-line test. Many persons who are far outside the original domain of application of this test have taken the test and their results are recorded. A Rosenberg Self-Esteem Scale data set from Kaggle can be found via the following link: https://www.kaggle.com/yamqwe/rosenberg-selfesteem-scale (accessed on 3 January 2022). When addressing such data, it is reasonable to first ask whether seemingly diverse systems defined across gender and age axes are equal through the perspective of the Rosenberg Self-Esteem Scale, This is one version of an MSP without specifically targeted Re–Co dynamics. Only by arriving at an answer to this MSP question can the second question be formulated accordingly.
Under the MSP setting, the first question can be stated as: does potential heterogeneity exist across spans of the age-axis and gender-axis or not? This is a rather interesting issue from the perspective of psychiatry, psychology, and sociology. We want to shed some light on this issue via the computational major factor selection method developed and illustrated in the previous sections.
This data set consists of the results of 47,974 subjects who took the online Rosenberg Self-Esteem Scale. We deleted subjects younger than 10 or older than 70. Subjects who did not indicate their gender information are collected into the “none was chosen” category of the gender variable. There are 10 questions in the Rosenberg Self-Esteem Test. Each subject can respond on the following scale: 1 = strongly disagree, 2 = disagree, 3 = agree, and 4 = strongly agree ( 0 = no answer).
Q 1 .
I feel that I am a person of worth, at least on an equal plane with others.
Q 2 .
I feel that I have a number of good qualities.
Q 3 .
All in all, I am inclined to feel that I am a failure.
Q 4 .
I am able to do things as well as most other people.
Q 5 .
I feel I do not have much to be proud of.
Q 6 .
I take a positive attitude toward myself.
Q 7 .
On the whole, I am satisfied with myself.
Q 8 .
I wish I could have more respect for myself.
Q 9 .
I certainly feel useless at times.
Q 10 .
At times I think I am no good at all.
Here we propose one fundamental way of resolving the aforementioned issue by looking into all possible response-to-covariate (Re–Co) dynamics and checking the effects of heterogeneity. For expositional simplicity, we use only one instance as an example. Given Q 1 , it is taken as a 1D response feature Y = Q 1 , while the rest of the nine questions Q 2 Q 10 are taken as covariate features. These 10 features are ordinal. We also take A g e and G e n d e r as two additional covariate features. Specifically, A g e takes categorical-membership values among the ordinal age-groups [ 10 , 14 ] , [ 15 , 18 ] , [ 19 , 22 ] , [ 23 , 30 ] , [ 31 , 40 ] , [ 41 , 50 ] , [ 51 , 60 ] , and [ 61 , 70 ] , while V 11 = G e n d e r takes categorical values of 1 = “male”, 2 = “female”, 3 = “other”; 0 = ”none was chosen”. That is, the MSP under study here is specified by the collection of complex systems encoded with the bivariate categories of ( A g e , G e n d e r ) .
The pairwise associations among these 12 categorical features are presented through an MCE-based heatmap and its network in Figure 7. A glimpse of an associative map of these 12 features is seen through the three small communities in the network. When showing heterogeneity effects via our major factor selection, A g e and G e n d e r will be treated as two separate covariate features, as is demonstrated in Table 14 and Table 15.
Selection of major factors in the Self-esteem test example with heterogeneity.
1 
In the one-feature setting of Table 14 and Table 15, Q 2 achieves the smallest CE with the largest CE drop: 0.3364 . Q 6 achieves the second ranked CE with a CE drop of 0.2443 . These two features are potential order-1 major factors. Q 7 , Q 4 and Q 3 are ranked third, fourth, and fifth, with CE drops near 0.2000 . In sharp contrast, A g e and G e n d e r individually have rather low CE drops. These two pieces of evidence indicate the absence of heterogeneity effects. This fact is significantly distinct from the presence of effects of heterogeneity discussed in previous sections for our designed MSP and all MLB-based MSP examples.
2 
In the two-feature setting, we can see that the SCEs are all significantly lower than corresponding CE drops among features from Q 3 to Q 10 . That is, all feature-pairs fail the criterion [C2: unreplacable], including pairs involving A g e and G e n d e r . Therefore, no order-2 major factors can be seen. The primary reason for this phenomenon is that Q 2 is associated with all other covariate features to the degree that there are no pairs satisfying the ecological effects.
We conclude that Q 2 is the solo order-1 major factor of the Q 1 -based Re–Co dynamics and that the effects of heterogeneity of A g e and G e n d e r are absent. In other words, from the perspective of the Q 1 -based Re–Co dynamics, all age-vs-gender defined systems are not heterogeneous. To fully resolve this issue, we need all possible Re–Co dynamics. This task will be undertaken in a separate report.
To further address this issue in depth, we go into the phase of decomposing the collection of complex systems into homogeneous groups; that is, the categories of the bivariate ( A g e , G e n d e r ) are clustered into homogeneous groups, as shown in Figure 8. Then, we apply our major factor selection protocol to further check whether the effects of A g e or G e n d e r exist within any of these homogeneous groups of systems.
Based on Figure 8, we select three obvious cluster-based groups: R 1 = { ( 2 , 2 ) , ( 3 , 2 ) } , R 2 = { ( 4 , 2 ) , ( 3 , 1 ) } , and R 3 = { ( 2 , 1 ) , ( 4 , 1 ) , ( 5 , 1 ) , ( 5 , 2 ) } . The sample size is 16,800 for R 1 , 11,721 for R 2 , and 15,312 for R 3 . The single major factor collection { Q 2 } is selected for R 3 and R 3 , while a collection of two order-1 major factors { Q 2 , A g e } is selected for R 2 , since Q 2 and A g e jointly achieve a SCE drop of ( 0.0046 ) against the CE drop of A g e of ( 0.0024 ) . Thus, the ecological effect is satisfied. In addition, A g e also meets the criterion [C1:confirmable], since its CE of 1.1713 is below the CE distribution of ε (having the same number of categories of A g e ) with a mean of 1.1736 and SD of 1.2193 × 10 4 . Therefore, Q 2 and A g e can both be order-1 major factors. Nevertheless, Q 2 is still the dominant order-1 major factor. Hence, we still conclude with a coherent result: systems within these three homogeneous groups are not heterogeneous from the perspective of Q 1 dynamics.
At the end of this example, we remark that the widely used Rosenberg Self-Esteem Scale was conceptualized by its author as a single-factor scale with scores ranging along a continuum of low self-esteem to high self-esteem [10]. From this perspective, our identification of Q 2 as the single major factor is seemingly coherent. Further, when three versions of rewording on Rosenberg Self-Esteem Scale were devised and a much smaller data set was collected and analyzed [18], the results of factor analysis [19] indeed indicated that the original version fits a two-factor model, while positively and negativey reworded versions fit single-factor models. It is noted that these results based on factor analysis are fundamentally irrelevant to the issue under study here.

7. Conclusions

We demonstrated the versatile merits of a CEDA-based protocol for selecting major factors of various orders as a proper way of studying any MSP. First, it is essential to reveal patter -information of “complexity interacting with heterogeneity” embedded within MSPs. Secondly, the heterogeneity embedded within the whole ensemble of complex systems needs to be broken down into homogeneity embraced by constituent groups of similar complex systems. Finally, similarity-based homogeneity and largeness of the data set will enable each constituent group to manifest detailed multiscale characteristics via major factors of various orders. This is taken as the true implication of P. W. Anderson’s “More is Different” under a structured data setting. This MSP implication of breaking heterogeneity for the sake of unraveling possibly distinct collections of hidden major factors across homogeneous parts may be fundamental and critical in this Big Data era.
With an understanding of such an MSP implication, we further demonstrate its merits in formulating resolutions to a scientific problem related to the Rosenberg Self-Esteem Scale in psychology and sociology. This demonstration helps us recognize the degree of the spread of MSPs in the real world and simultaneously indicates the huge potential of our computational major factor selection techniques.
As scientists, we want to prevent data-driven understandings being twisted and information content being compromised. Nowadays, MSPs can be ubiquitously found in diverse research fields and real-world businesses. Proper discovery of data’s authentic information content and true understanding of it should greatly help our societies. In this paper, we demonstrate that data-driven understandings can be supported by visible and explainable relational patterns found on the simple platform of contingency tables. This is essential and important for data analysis in this Big Data era.
Furthermore, we emphasize that our CEDA computations work for all data types. This is an essential virtue of data analysis, since its categorical nature involves features of any data types. By employing such a categorical nature in data, the contingency table platform becomes natural, as do information theoretical measurements. Consequently, the pattern-formation brought out by conditional entropy and mutual information is natural, so it reveals and facilitates understanding pertaining to MSPs of interest. In other words, this CEDA-based data analysis is an effective way of studying an ensemble of complex systems.
By visualizing how complexity interacts with heterogeneity and how homogeneity unravels many more major factors in the two MLB-related MSPs, we recognize the fact that a group of similar pitchers’ collective pitching dynamics is indeed much more involved in many aspects than the whole collection of pitchers, as well as a single pitcher’s pitching dynamics. This recognition allows us to fundamentally resolve the issue of how to properly compare the pitching dynamics of a group of pitchers.
By resolving these real-world MSPs, we also demonstrated how to successfully implement the two criteria [C1: confirmable] and [C2: irreplaceable] together with reliability checks. By recognizing existential heterogeneity and exploring its complicated effects and then teasing out diverse versions of homogeneity in all constituent groups, our various collections of major factors bring out the common governing mechanisms shared by all involved complex systems and, at the same time, they manifest distinctive mechanisms that characterize each of the identified groups of similar complex systems. The reliability check using Algorithm 2 is critical when subject to the finite sample phenomenon, even in the Big Data era.
Finally, we summarize the chief concept underlying our CEDA-based feature selection for major factors: “Let data’s categorical nature assemble freely and naturally to shed light on complexity, heterogeneity, and homogeneity in MSPs”.

Author Contributions

Conceptualization, T.-L.C. and H.F.; Formal analysis, E.P.C.; Investigation, E.P.C., T.-L.C. and H.F. ; Methodology, E.P.C., T.-L.C. and H.F. ; Project administration, H.F.; Software, E.P.C.; Visualization, E.P.C.; Writing—original draft, H.F.; Writing—review & editing, E.P.C., T.-L.C. and H.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The pitching data are available in PITCHf/x database belonging to Major League Baseball via http://gd2.mlb.com/components/game/mlb (accessed on 3 January 2022). Rosenberg Self-Esteem Scale data set from Kaggle can be found via https://www.kaggle.com/yamqwe/rosenberg-selfesteem-scale (accessed on 3 January 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gell-Mann, M. What is complexity? Complexity 1995, 1, 16–19. [Google Scholar] [CrossRef]
  2. Adami, C. What is Complexity? BioEssays 2002, 24, 1085–1094. [Google Scholar] [CrossRef] [PubMed]
  3. Bar-Yam, Y. General Features of Complex Systems. In Encyclopedia of Life Support Systems; World Wide Web: Geneva, Switzerland, 2002. [Google Scholar]
  4. Wibral, M.; Vicente, R.; Lizier, J. Directed Information Measures in Neuroscience. In Understanding Complex Systems; Springer: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  5. Chiappori, P.-A.; Haddad, L.; Hoddinott, J.; Kanbur, R. Unitary versus Collective Modelsof the Household: Time to Shift the Burden of Proof? In Policy Research Working Paper; World Bank: Washington, DC, USA, 1993; Volume 1217. [Google Scholar]
  6. Tumer, K.; Wolpert, D. Collectives and the Design of Complex Systems; Springer: Berlin/Heidelberg, Germany, 2004. [Google Scholar]
  7. Reif, F. Fundamentals of Statistical and Thermal Physics; Waveland Press: Long Grove, IL, USA, 2009. [Google Scholar]
  8. Briggs, L. Effect of Spin and Speed on the Lateral Deflection (Curve) of a Baseball; and the Magnus Effect for Smooth Spheres. Am. J. Phys. 1959, 27, 589–596. [Google Scholar] [CrossRef]
  9. Anderson, P.W. More is different. Science 1972, 177, 393–396. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Rosenberg, M. Society and the Adolescent Self-Image; Princeton University Press: Princeton, NJ, USA, 1965. [Google Scholar]
  11. Chen, T.-L.; Chou, E.P.; Fushing, H. Categorical nature of major factor selection in complex systems via Information Theoretical measurements. Entropy 2021, 23, 1684. [Google Scholar] [CrossRef] [PubMed]
  12. Fushing, H.; Chou, E.P. Categorical Exploratory Data Analysis: From Multiclass Classification and Response Manifold Analytics perspectives of baseball pitching dynamics. Entropy 2021, 23, 792. [Google Scholar] [CrossRef]
  13. Fushing, H.; Chou, E.P.; Chen, T.-L. Mimicking complexity of structured data matrix’s information content: Categorical Exploratory Data Analysis. Entropy 2021, 23, 594. [Google Scholar] [CrossRef]
  14. Tukey, J.W. Exploratory Data Analysis; Pearson: London, UK, 1977. [Google Scholar]
  15. Fushing, H.; Roy, T. Complexity of Possibly-gapped Histogram and Analysis of Histogram (ANOHT). R. Soc. Open Sci. 2018, 5, 171026. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Fushing, H.; Liu, S.-Y.; Hsieh, Y.-C.; McCowan, B. From patterned response dependency to structured covariate dependency: Categorical-pattern-matching. PLoS ONE 2018, 13, e0198253. [Google Scholar] [CrossRef] [PubMed]
  17. Hlavackova-Schindler, K.; Palus, M.; Vejmelka, M.; Bhattacharya, J. Causality detection based on information-theoretic approaches in time series analysis. Phys. Rep. 2007, 441, 1–46. [Google Scholar] [CrossRef]
  18. Greenberger, E.; Chen, C.-S.; Dmitrieva, J.; Susan, P.; Farruggia, S.P. Item-wording and the dimensionality of the Rosenberg Self-Esteem Scale: Do they matter? Personal. Individ. Differ. 2003, 35, 1241–1254. [Google Scholar] [CrossRef] [Green Version]
  19. Child, D. The Essentials of Factor Analysis, 3rd ed.; Bloomsbury Academic Press: London, UK, 2006. [Google Scholar]
Figure 1. Clustering on 80 IDs’ vectors of means of 11 features.
Figure 1. Clustering on 80 IDs’ vectors of means of 11 features.
Entropy 24 00170 g001
Figure 2. Associative patterns of sliders among 14 features: (A) heatmap based on MCE-matrix; (B) network built with linkages with thickness proportional to one minus pairwise MCE and subject to a threshold 0.2 ( = 1.0 M C E ) .
Figure 2. Associative patterns of sliders among 14 features: (A) heatmap based on MCE-matrix; (B) network built with linkages with thickness proportional to one minus pairwise MCE and subject to a threshold 0.2 ( = 1.0 M C E ) .
Entropy 24 00170 g002
Figure 3. Testing for [C1: confirmable] on Y = ( p f x X , p f x Z ) : (A) ( a X , a Z ) (red-marked) against histogram of ( a X , ξ ) ; (B) ( a X , a Z , p i t N ) and ( a X , a Z , v Y 0 ) (red-marked) against histogram of ( a X , a Z , ξ ) ; (C) ( v Z 0 , z 0 , a Y ) (red-marked) against histogram of ( v Z 0 , z 0 , ξ ) .
Figure 3. Testing for [C1: confirmable] on Y = ( p f x X , p f x Z ) : (A) ( a X , a Z ) (red-marked) against histogram of ( a X , ξ ) ; (B) ( a X , a Z , p i t N ) and ( a X , a Z , v Y 0 ) (red-marked) against histogram of ( a X , a Z , ξ ) ; (C) ( v Z 0 , z 0 , a Y ) (red-marked) against histogram of ( v Z 0 , z 0 , ξ ) .
Entropy 24 00170 g003
Figure 4. Associative patterns of fastballs among 14 features: (A) heatmap based on MCE-matrix; (B) network built with linkages with thickness proportional to one minus pairwise MCE and subject to a threshold 0.2 ( = 1.0 M C E ) .
Figure 4. Associative patterns of fastballs among 14 features: (A) heatmap based on MCE-matrix; (B) network built with linkages with thickness proportional to one minus pairwise MCE and subject to a threshold 0.2 ( = 1.0 M C E ) .
Entropy 24 00170 g004
Figure 5. Testing for [C1: confirmable] on Y = ( p f x X , p f x Z ) : (A) ( a X , a Z , v Y 0 ) and ( a X , a Z , s t a r t S p ) red-marked against histogram of ( a X , a Z , ε ) ; (B) ( a Y , p i t N , s p i n R ) and ( a Y , p i t N , s t a r t S p ) red-marked against histogram of ( a Y , p i t N , ε ) ; on Y = ( a x , a z ) : (C) ( s p i n D , p i t N , a Y ) red-marked against histogram of ( s p i n D , p i t N , ε ) .
Figure 5. Testing for [C1: confirmable] on Y = ( p f x X , p f x Z ) : (A) ( a X , a Z , v Y 0 ) and ( a X , a Z , s t a r t S p ) red-marked against histogram of ( a X , a Z , ε ) ; (B) ( a Y , p i t N , s p i n R ) and ( a Y , p i t N , s t a r t S p ) red-marked against histogram of ( a Y , p i t N , ε ) ; on Y = ( a x , a z ) : (C) ( s p i n D , p i t N , a Y ) red-marked against histogram of ( s p i n D , p i t N , ε ) .
Entropy 24 00170 g005
Figure 6. Hierarchical clustering trees marked with pitcher groups: (A) 62 slider pitchers; (B) 199 fastball pitchers with 3 colored major branches A, B, and C (from left to right). The C-branch is further partitioned into six colored sub-branches: C1 to C6 (left to right).
Figure 6. Hierarchical clustering trees marked with pitcher groups: (A) 62 slider pitchers; (B) 199 fastball pitchers with 3 colored major branches A, B, and C (from left to right). The C-branch is further partitioned into six colored sub-branches: C1 to C6 (left to right).
Entropy 24 00170 g006
Figure 7. Associative patterns among 12 features of Rosenberg Self-Esteem Scale:(A) heatmap based on MCE-matrix; (B) network built with linkages with thickness proportional to one minus pairwise MCE and subject to a threshold 0.2 ( = 1.0 M C E ) .
Figure 7. Associative patterns among 12 features of Rosenberg Self-Esteem Scale:(A) heatmap based on MCE-matrix; (B) network built with linkages with thickness proportional to one minus pairwise MCE and subject to a threshold 0.2 ( = 1.0 M C E ) .
Entropy 24 00170 g007
Figure 8. Decomposing the collection of systems defined by A g e and G e n d e r into homogeneous groups in the Rosenberg Self-Esteem Scale. Bivariate-encoded A g e -vs- G e n d e r systems are arranged along the column-axis, while the bivariate-encoded Q 2 -vs- Q 6 categories are arranged along the row-axis. An HC-tree is superimposed on the column-axis for identifying homogeneous groups of systems.
Figure 8. Decomposing the collection of systems defined by A g e and G e n d e r into homogeneous groups in the Rosenberg Self-Esteem Scale. Bivariate-encoded A g e -vs- G e n d e r systems are arranged along the column-axis, while the bivariate-encoded Q 2 -vs- Q 6 categories are arranged along the row-axis. An HC-tree is superimposed on the column-axis for identifying homogeneous groups of systems.
Entropy 24 00170 g008
Table 1. Top 12 ranked successive CE drops of feature-sets across one-feature to three-feature settings of the designed MSP. CE of Y is equal to 2.4338 .
Table 1. Top 12 ranked successive CE drops of feature-sets across one-feature to three-feature settings of the designed MSP. CE of Y is equal to 2.4338 .
One-FeatureSCE DropTwo-FeatureSCE DropThree-FeatureSCE Drop
V 2 0.3332 V 1 _ V 4 0.1350 V 1 _ V 4 _ V 5 1.0734
V 1 0.0106 V 1 _ V 5 0.1349 V 1 _ V 3 _ V 6 1.0708
V 11 0.0017 V 1 _ V 3 0.1336 V 1 _ V 3 _ V 4 1.0706
V 5 0.0016 V 1 _ V 6 0.1322 V 1 _ V 3 _ V 5 1.0706
V 10 0.0015 V 1 _ V 8 0.1229 V 1 _ V 4 _ V 6 1.0682
V 3 0.0014 V 1 _ V 12 0.1214 V 1 _ V 5 _ V 6 1.0682
V 8 0.0014 V 1 _ V 9 0.1213 V 1 _ V 3 _ V 9 1.0327
V 4 0.0013 V 1 _ V 10 0.1207 V 1 _ V 3 _ V 12 1.0326
V 9 0.0013 V 1 _ V 11 0.1201 V 1 _ V 4 _ V 12 1.0308
V 6 0.0013 V 1 _ V 7 0.1193 V 1 _ V 5 _ V 12 1.0300
V 12 0.0012 V 1 _ V 2 0.1035 V 1 _ V 3 _ V 10 1.0300
V 7 0.0011 V 10 _ V 11 0.0194 V 1 _ V 5 _ V 9 1.0289
Table 2. Designed MSP: comparisons of collections of major factors based on the whole data set with heterogeneity and homogeneous groups.
Table 2. Designed MSP: comparisons of collections of major factors based on the whole data set with heterogeneity and homogeneous groups.
Whole or GroupOrder-1 MFOrder-3 MFAlternative MF
Whole-data V 2 none confirmednone confirmed
G 1 V 2 ( V 4 , V 5 , V 6 ) none confirmed
G 2 V 2 ( V 3 , V 5 , V 6 ) none confirmed
G 3 V 2 ( V 3 , V 4 , V 6 ) none confirmed
G 4 V 2 ( V 3 , V 4 , V 5 ) none confirmed
G 5 V 2 none confirmednone confirmed
Table 3. Sliders: 12 top ranked CEs of five settings of feature-sets for response variable Y = ( p f x X , p f x Z ) . CE of Y is 4.6890.
Table 3. Sliders: 12 top ranked CEs of five settings of feature-sets for response variable Y = ( p f x X , p f x Z ) . CE of Y is 4.6890.
One-FeatureCETwo-FeatureCEThree-FeatureCEFour-FeatureCEFive-FeatureCE
aX2.7631aZ_aX0.9264aZ_aX_pitN0.6617aY_aZ_aX_pitN0.3694aY_aZ_vZ0_aX_pitN0.1070
aZ2.8121aX_spinD1.8411aZ_vY0_aX0.6922aZ_vZ0_aX_pitN0.3726aY_aZ_vZ0_z0_aX0.1313
spinD3.0063aZ_spinD1.8678aZ_aX_startSp0.6946aZ_vY0_aX_pitN0.4377aY_aZ_vY0_vZ0_aX0.1400
pitN3.5526aZ_pitN1.9508aZ_aX_spinR0.8216aZ_aX_startSp_pitN0.4390aY_aZ_vZ0_aX_startSp0.1430
x04.3582aX_pitN2.1953aZ_z0_aX0.8238aZ_aX_spinR_pitN0.4537aY_aZ_vY0_z0_aX0.1439
vX04.3719spinD_pitN2.4467aY_aZ_aX0.8251aZ_vY0_vZ0_aX0.4596aY_aZ_z0_aX_startSp0.1448
startSp4.4369aZ_vX02.4929aZ_vZ0_aX0.8342aZ_vY0_z0_aX0.4603aZ_vY0_vZ0_z0_aX0.1479
vY04.4392aZ_x02.4961aZ_vX0_aX0.8388aZ_z0_aX_startSp0.4618aZ_vZ0_z0_aX_startSp0.1485
vZ04.5399aZ_vY02.5263aZ_x0_aX0.8505aZ_vZ0_aX_startSp0.4627aY_aZ_vZ0_aX_spinR0.1538
spinR4.5482aZ_startSp2.5290aZ_aX_spinD0.8845aY_aZ_vY0_aX0.4635aY_aZ_z0_aX_spinR0.1556
z04.6012aX_startSp2.5462aZ_spinD_pitN1.3791aZ_vX0_aX_pitN0.4640aY_aZ_vY0_aX_pitN0.1575
aY4.6261vY0_aX2.5463aY_aZ_pitN1.4649aY_aZ_aX_startSp0.4663aY_aZ_aX_startSp_pitN0.1594
Table 4. Sliders: 12 top ranked CE drops of four settings of feature-sets for the response variable Y = ( p f x X , p f x Z ) .
Table 4. Sliders: 12 top ranked CE drops of four settings of feature-sets for the response variable Y = ( p f x X , p f x Z ) .
One-FeatureSCE DropTwo-FeatureSCE DropThree-FeatureSCE DropFour-FeatureSCE Drop
aX1.9261aZ_aX1.8367aY_vZ0_pitN1.1533aY_vY0_vZ0_z01.6601
aZ1.8771aZ_spinD0.9443aY_vZ0_z01.0929aY_vZ0_z0_spinR1.6535
spinD1.6829aX_spinD0.9220vZ0_z0_spinR1.0343aY_vZ0_z0_startSp1.6521
pitN1.1366aZ_pitN0.8614aY_z0_spinR1.0241aY_vX0_vZ0_z01.6374
x00.3310aX_pitN0.5679aY_vZ0_spinR1.0032aY_vX0_vZ0_spinR1.6105
vX00.3173spinD_pitN0.5595vY0_vZ0_z00.9929aY_vY0_vZ0_spinR1.5936
startSp0.2523vZ0_pitN0.4544vZ0_z0_startSp0.9863aY_vZ0_spinR_startSp1.5859
vY00.2500aY_pitN0.4456aY_vY0_z00.9850aY_vY0_z0_spinR1.5651
vZ00.1493spinR_startSp0.3788aY_vX0_z00.9798aY_vX0_vY0_z01.5645
spinR0.1410vY0_spinR0.3772aY_z0_startSp0.9784aY_vX0_z0_spinR1.5644
z00.0880vX0_startSp0.3728aY_vY0_vZ00.9674vY0_vZ0_z0_spinR1.5629
aY0.0631vX0_vY00.3722aY_vZ0_startSp0.9589aY_vX0_vY0_vZ01.5616
Table 5. Sliders: 10 top ranked CEs of five settings of feature-sets for the response variable Y = ( a X , a Z ) . CE of Y is 4.7213.
Table 5. Sliders: 10 top ranked CEs of five settings of feature-sets for the response variable Y = ( a X , a Z ) . CE of Y is 4.7213.
One-FeatureCETwo-FeatureCEThree-FeatureCEFour-FeatureCEFive-FeatureCE
spinD3.0301spinD_pitN2.4891aY_vZ0_pitN1.9577aY_vZ0_spinR_pitN0.9624aY_vY0_vZ0_spinR_pitN0.3837
pitN3.6086vX0_spinD2.8168vZ0_spinD_pitN1.9577aY_vZ0_spinD_pitN0.9732aY_vZ0_spinR_startSp_pitN0.3873
x04.3981x0_spinD2.8689aY_spinD_pitN1.9657aY_vY0_vZ0_pitN0.9867aY_vX0_vZ0_spinR_pitN0.3922
vX04.4110vZ0_spinD2.8779vX0_spinD_pitN2.1228aY_vZ0_startSp_pitN0.9961aY_vX0_vZ0_z0_spinR0.4034
startSp4.4923spinD_startSp2.8848spinD_startSp_pitN2.1355aY_vX0_vZ0_pitN1.0258aY_vY0_vZ0_z0_spinR0.4053
vY04.4950vY0_spinD2.8854vY0_spinD_pitN2.1387aY_vZ0_z0_pitN1.0853aY_vZ0_spinR_spinD_pitN0.4074
vZ04.5698spinR_spinD2.9170spinR_spinD_pitN2.1541aY_vX0_spinD_pitN1.2414aY_vZ0_z0_spinR_startSp0.4098
spinR4.5811z0_spinD2.9294z0_spinD_pitN2.1892aY_spinR_spinD_pitN1.2517aY_vX0_vY0_vZ0_pitN0.4125
z04.6370aY_spinD2.9413aY_spinR_pitN2.2981vZ0_spinR_spinD_pitN1.2582aY_vX0_vY0_vZ0_z00.4153
aY4.6526vZ0_pitN3.1350vZ0_spinR_pitN2.3014vY0_vZ0_spinD_pitN1.2639aY_vX0_vZ0_startSpeed_pitN0.4176
Table 6. Sliders: 10 top ranked CE drops of four settings of feature-sets for the response variable Y = ( a X , a Z ) .
Table 6. Sliders: 10 top ranked CE drops of four settings of feature-sets for the response variable Y = ( a X , a Z ) .
One-FeatureSCE DropTwo-FeatureSCE DropThree-FeatureSCE DropFour-FeatureSCE Drop
spinD1.6912spinD_pitN0.5410aY_vZ0_pitN1.1774aY_vY0_vZ0_z01.6804
pitN1.1127vZ0_pitN0.4736aY_vZ0_z01.1105aY_vZ0_z0_startSp1.6727
x00.3233aY_pitN0.4644vZ0_z0_spinR1.0418aY_vZ0_z0_spinR1.6705
vX00.3104spinR_startSp0.3855aY_z0_spinR1.0417aY_vX0_vZ0_z01.6459
startSp0.2290vY0_spinR0.3832aY_vZ0_spinR1.0220aY_vX0_vZ0_spinR1.6207
vY00.2263vX0_startSp0.3641vY0_vZ0_z01.0173aY_vY0_vZ0_spinR1.6133
vZ00.1515vX0_vY00.3627aY_vY0_z01.0129aY_vZ0_spinR_startSp1.6066
spinR0.1402x0_startSp0.3625vZ0_z0_startSp1.0103aY_vY0_z0_spinR1.5845
z00.0844vY0_x00.3597aY_z0_startSp1.0040vY0_vZ0_z0_spinR1.5809
aY0.0687vX0_spinR0.3530aY_vX0_z00.9978aY_vX0_z0_spinR1.5770
Table 7. Fastballs: 12 top ranked CE of five settings of feature-sets for the response variable Y = ( p f x X , p f x Z ) . The CE of Y is 4.7167.
Table 7. Fastballs: 12 top ranked CE of five settings of feature-sets for the response variable Y = ( p f x X , p f x Z ) . The CE of Y is 4.7167.
One-FeatureCETwo-FeatureCEThree-FeatureCEFour-FeatureCEFive-FeatureCE
spinD2.9688aZ_aX1.7424aZ_vY0_aX1.0888aZ_vY0_aX_pitN0.7295aY_aZ_vZ0_spinR_pitN0.2540
aX2.9935aZ_spinD1.9326aZ_aX_startSp1.1046aZ_vZ0_aX_pitN0.7318aY_aZ_vZ0_aX_pitN0.2638
pitN3.5048aZ_pitN2.0718aZ_aX_pitN1.1695aZ_aX_startSp_pitN0.7407aY_aZ_vY0_vZ0_pitN0.2763
aZ3.5102aX_spinD2.4713aZ_spinD_pitN1.2602aY_aZ_vZ0_pitN0.7540aY_aZ_vZ0_startSp_pitN0.2806
vX04.0981spinD_pitN2.5521aZ_vY0_spinD1.2951aY_aZ_aX_pitN0.7687aZ_vY0_vZ0_aX_pitN0.2944
x04.1266aX_pitN2.5772aZ_spinD_startSp1.3106aZ_vZ0_spinD_pitN0.8125aZ_vZ0_aX_startSp_pitN0.2984
aY4.5772vX0_spinD2.8220aZ_vZ0_pitN1.5956aZ_vY0_spinD_pitN0.8190aY_aZ_vZ0_spinD_pitN0.3062
vZ04.5960aY_spinD2.8287aZ_vY0_pitN1.5965aZ_aX_spinR_pitN0.8211aZ_vZ0_aX_spinR_pitN0.3080
z04.6173aZ_vY02.8443aZ_startSp_pitN1.6070aZ_spinD_startSp_pitN0.8293aY_aZ_vY0_aX_pitN0.3249
spinR4.6676aY_aX2.8474aY_aZ_pitN1.6196aY_aZ_spinD_pitN0.8453aY_aZ_aX_startSp_pitN0.3306
startSp4.6728vZ0_spinD2.8520aZ_aX_spinD1.6315aZ_vZ0_spinR_pitN0.8567aY_aZ_vX0_vZ0_pitN0.3307
vY04.6740aZ_startSp2.8558aY_aZ_aX1.6560aZ_vY0_vZ0_pitN0.8660aZ_vY0_vZ0_spinR_pitN0.3348
Table 8. Fastballs: 12 top ranked CE drops of four settings of feature-sets for the response variable Y = ( p f x X , p f x Z ) .
Table 8. Fastballs: 12 top ranked CE drops of four settings of feature-sets for the response variable Y = ( p f x X , p f x Z ) .
One-FeatureSCE DropTwo-FeatureSCE DropThree-FeatureSCE DropFour-FeatureSCE Drop
spinD1.7480aZ_pitN1.4329aY_vZ0_pitN1.0297vY0_vZ0_z0_spinR1.4772
aX1.7232aZ_aX1.2511vZ0_spinR_pitN0.8179aY_vZ0_z0_spinR1.4731
pitN1.2120aZ_spinD1.0362aY_spinR_pitN0.7552vZ0_z0_spinR_startSp1.4700
aZ1.2065aZ_vY00.6659vY0_vZ0_pitN0.7349aY_vY0_vZ0_spinR1.3818
vX00.6186aZ_startS0.6544vZ0_startSp_pitN0.7316aY_vZ0_spinR_startSp1.3622
x00.5901aZ_vX00.6232aZ_spinD_pitN0.6724aY_vY0_z0_spinR1.3555
aY0.1395aZ_x00.5920aY_vY0_pitN0.6717aY_vY0_vZ0_z01.3527
vZ00.1207aX_spinD0.4975aY_startSp_pitN0.6612aY_z0_spinR_startSp1.3390
z00.0995aY_pitN0.4280aZ_vY0_aX0.6536aY_vZ0_z0_startSp1.3321
spinR0.0491spinD_pitN0.4167aZ_vX0_vY00.6467vX0_vY0_vZ0_spinR1.1691
startSp0.0439aX_pitN0.4163aZ_vX0_startSp0.6465vY0_vZ0_x0_spinR1.1612
vY00.0428vZ0_pitN0.3883aZ_aX_startSp0.6378vX0_vZ0_spinR_startSp1.1605
Table 9. Fastballs: 10 top ranked CE drops of four settings of feature-sets for the response variable Y = ( a X , a Z ) . CE of Y is 4.7000.
Table 9. Fastballs: 10 top ranked CE drops of four settings of feature-sets for the response variable Y = ( a X , a Z ) . CE of Y is 4.7000.
One-FeatureCETwo-FeatureCEThree-FeatureCEFour-FeatureCEFive-FeatureCE
spinD2.9447spinD_pitN2.4991aY_spinD_pitN1.9962aY_vZ0_spinR_pitN0.9875aY_vY0_vZ0_spinR_pitN0.3918
pitN3.4559aY_spinD2.7193vZ0_spinD_pitN2.0159aY_vZ0_spinD_pitN1.0371aY_vZ0_spinR_spinD_pitN0.3965
vX04.0743vX0_spinD2.7908aY_vZ0_pitN2.0286aY_vY0_vZ0_pitN1.0750aY_vZ0_spinR_startSp_pitN0.3989
x04.1036vZ0_spinD2.8114spinR_spinD_pitN2.1612aY_vZ0_startSp_pitN1.0889aY_vZ0_z0_spinR_pitN0.4428
aY4.4743x0_spinD2.8391spinD_startSp_pitN2.1878aY_vX0_vZ0_pitN1.1993aY_vX0_vZ0_spinR_pitN0.4434
vZ04.5634spinD_startSp2.8422vY0_spinD_pitN2.1966aY_vZ0_z0_pitN1.1993aY_vY0_vZ0_spinD_pitN0.4435
startSp4.5861vY0_spinD2.8526vX0_spinD_pitN2.2366vZ0_spinR_spinD_pitN1.2426aY_vZ0_spinD_startSp_pitN0.4504
vY04.5964z0_spinD2.8873z0_spinD_pitN2.2529aY_spinR_spinD_pitN1.2773aY_vX0_vY0_vZ0_pitN0.4966
z04.6030spinR_spinD2.8900vZ0_spinR_pitN2.2775vY0_vZ0_spinD_pitN1.3018aY_vY0_vZ0_z0_pitN0.5008
spinR4.6380aY_pitN3.0293aY_spinR_pitN2.2950vZ0_spinD_startSp_pitN1.3043aY_vX0_vZ0_startSp_pitN0.5064
Table 10. Fastballs: 10 top ranked CE drops of four settings of feature-sets for the response variable Y = ( a X , a Z ) .
Table 10. Fastballs: 10 top ranked CE drops of four settings of feature-sets for the response variable Y = ( a X , a Z ) .
One-FeatureSCE DropTwo-FeatureSCE DropThree-FeatureSCE-DropFour-FeatureSCE Drop
spinD1.7553spinD_pitN0.4457aY_vZ0_pitN1.0006vY0_vZ0_z0_spinR1.4383
pitN1.2441aY_pitN0.4266vZ0_spinR_pitN0.8026vZ0_z0_spinR_startSp1.4257
vX00.6257vZ0_pitN0.3757aY_spinR_pitN0.7343aY_vZ0_z0_spinR1.4129
x00.5964spinR_pitN0.2689vY0_vZ0_pitN0.7276aY_vY0_vZ0_spinR1.3119
aY0.2257aY_x00.2566vZ0_startSp_pitN0.7262aY_vY0_z0_spinR1.3024
vZ00.1366aY_vX00.2550aY_vY0_pitN0.6490aY_vZ0_spinR_startSp1.2933
startSp0.1139startSp_pitN0.2485aY_startSp_pitN0.6422aY_z0_spinR_startSp1.2879
vY00.1036vY0_pitN0.2380vZ0_z0_pitN0.6047aY_vY0_vZ0_z01.2832
z00.0970aY_spinD0.2255spinR_startSp_pitN0.5933aY_vZ0_z0_startSp1.2641
spinR0.0620z0_pitN0.2069vX0_vZ0_pitN0.5911vX0_vY0_vZ0_spinR1.1406
Table 11. Comparison of major factors of sliders and fastballs with respect to Y = ( p f x X , p f x Z ) and Y = ( a X , a Z ) .
Table 11. Comparison of major factors of sliders and fastballs with respect to Y = ( p f x X , p f x Z ) and Y = ( a X , a Z ) .
Pitch Type Y Order-1 MFOrder-2 MFAlternative MFsAlternative MFs
slider ( p f x X , p f x Z ) a Z ( z 0 , v Y 0 ) { a X , ( z 0 , v Z 0 ) } { a X , ( a Y , v Z 0 ) }
slider ( a X , a Z ) { s p i n D } ( z 0 , v Z 0 ) { s p i n D , ( a Y , v Z 0 ) } { s p i n D , ( a Y , z 0 ) }
fastball ( p f x X , p f x Z ) a X ( a Z , v Y 0 ) { a Z , ( v Y 0 , a Y ) } { a Z , ( v Y 0 , z 0 ) }
fastball ( a X , a Z ) s p i n D , a Y none confirmed! { s p i n D , v Z 0 } { s p i n D }
Table 12. Table of major factors of Y = ( a X , a Z ) in five groups of slider pitchers (A–E).
Table 12. Table of major factors of Y = ( a X , a Z ) in five groups of slider pitchers (A–E).
Pitch GroupsCE of Y Order-1 MFOrder-2 MFAlternative MFsAlternative MFs
A3.3574 s p i n D , a Y None confirmed!NoneNone
B4.2448 s p i n D , v X 0 None confirmed! { s p i n D , x 0 } { s p i n D , v Z 0 }
C4.1125 s p i n D , z 0 None confirmed! { s p i n D , a Y } { s p i n D , s p i n R }
D4.0750 s p i n D , v Z 0 None confirmed! { s p i n D , z 0 } { s p i n D , a Y }
E4.6577 s p i n D , v Z 0 None confirmed! { s p i n D , a Y } { s p i n D }
Table 13. Table of major factors of Y = ( a X , a Z ) in six subgroups of the C group of pitchers.
Table 13. Table of major factors of Y = ( a X , a Z ) in six subgroups of the C group of pitchers.
Pitch GroupsCE of Y Order-1 MFOrder-2 MFAlternative MFsAlternative MFs
C13.6074 s p i n D , v Z 0 none confirmed { s p i n D , a Y } { s p i n D , v X 0 }
C23.8715 s p i n D ( a Y , v Z 0 ) { s p i n D , v Z 0 } { s p i n D , a Y }
C34.1491 s p i n D ( a Y , v Z 0 ) { s p i n D , ( a Y , s p i n R ) } { s p i n D , a Y }
C43.5763None confirmed! ( v Z 0 , s p i n R ) , ( a Y , v Z 0 ) { ( a Y , v Z 0 ) , ( v Z 0 , z 0 ) } { s p i n D , a Y }
C53.9358 s p i n D ( a Y , z 0 ) { s p i n D , ( a Y , s p i n R ) } { s p i n D , ( v Z 0 , s p i n R ) }
C63.9408 s p i n D ( v Z 0 , s p i n R ) { s p i n D , ( v Z 0 , z 0 ) } { s p i n D , v Z 0 }
Table 14. Top 12 ranked CEs of feature-sets across one-feature to three-feature settings on the Rosenberg Self-Esteem Test. CE of Y ( = Q 1 ) is 1.2204.
Table 14. Top 12 ranked CEs of feature-sets across one-feature to three-feature settings on the Rosenberg Self-Esteem Test. CE of Y ( = Q 1 ) is 1.2204.
One-FeatureCETwo-FeatureCEThree-FeatureCE
Q 2 0.8840 Q 2 _ Q 6 0.8039 Q 2 _ Q 4 _ Q 6 0.7696
Q 6 0.9761 Q 2 _ Q 3 0.8131 Q 2 _ Q 3 _ Q 6 0.7709
Q 7 1.0057 Q 2 _ Q 7 0.8136 Q 2 _ Q 6 _ Q 10 0.7766
Q 4 1.0209 Q 2 _ Q 10 0.8216 Q 2 _ Q 3 _ Q 7 0.7782
Q 3 1.0211 Q 2 _ Q 4 0.8304 Q 2 _ Q 3 _ Q 4 0.7793
Q 10 1.0349 Q 2 _ Q 5 0.8334 Q 2 _ Q 4 _ Q 7 0.7795
Q 5 1.0409 Q 2 _ Q 9 0.8354 Q 2 _ Q 5 _ Q 6 0.7795
Q 9 1.0727 Q 2 _ Q 8 0.8429 Q 2 _ Q 7 _ Q 10 0.7831
V 8 1.1141 Q 2 _ A g e 0.8806 Q 2 _ Q 6 _ Q 7 0.7831
A g e 1.2002 Q 2 _ G e n d e r 0.8821 Q 2 _ Q 6 _ Q 9 0.7832
G e n d e r 1.2133 Q 4 _ Q 6 0.8908 Q 2 _ Q 4 _ Q 10 0.7849
Table 15. Top 12 ranked successive CE drops of feature-sets across one-feature to three-feature settings of the Rosenberg Self-Esteem Test.
Table 15. Top 12 ranked successive CE drops of feature-sets across one-feature to three-feature settings of the Rosenberg Self-Esteem Test.
One-FeatureSCE DropTwo-FeatureSCE DropThree-FeatureSCE Drop
Q 2 0.3364 Q 3 _ Q 4 0.1022 Q 3 _ Q 4 _ Q 7 0.0447
Q 6 0.2443 Q 4 _ Q 10 0.0965 Q 4 _ Q 7 _ Q 10 0.0439
Q 7 0.2147 Q 4 _ Q 7 0.0937 Q 4 _ Q 5 _ Q 10 0.0408
Q 4 0.1995 Q 4 _ Q 5 0.0899 Q 3 _ Q 4 _ Q 6 0.0394
Q 3 0.1993 Q 4 _ Q 6 0.0852 Q 3 _ Q 4 _ Q 10 0.0390
Q 10 0.1855 Q 2 _ Q 6 0.0800 Q 4 _ Q 5 _ Q 7 0.0389
Q 5 0.1795 Q 2 _ Q 3 0.0708 Q 4 _ Q 6 _ Q 10 0.0369
Q 9 0.1477 Q 4 _ Q 9 0.0707 Q 2 _ Q 4 _ Q 10 0.0367
V 8 0.1063 Q 2 _ Q 7 0.0704 Q 4 _ Q 5 _ Q 6 0.0361
A g e 0.0202 Q 3 _ Q 7 0.0703 Q 2 _ Q 4 _ Q 5 0.0354
G e n d e r 0.0071 Q 5 _ Q 10 0.0662 Q 2 _ Q 3 _ Q 7 0.0349
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chou, E.P.; Chen, T.-L.; Fushing, H. Unraveling Hidden Major Factors by Breaking Heterogeneity into Homogeneous Parts within Many-System Problems. Entropy 2022, 24, 170. https://doi.org/10.3390/e24020170

AMA Style

Chou EP, Chen T-L, Fushing H. Unraveling Hidden Major Factors by Breaking Heterogeneity into Homogeneous Parts within Many-System Problems. Entropy. 2022; 24(2):170. https://doi.org/10.3390/e24020170

Chicago/Turabian Style

Chou, Elizabeth P., Ting-Li Chen, and Hsieh Fushing. 2022. "Unraveling Hidden Major Factors by Breaking Heterogeneity into Homogeneous Parts within Many-System Problems" Entropy 24, no. 2: 170. https://doi.org/10.3390/e24020170

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop