Next Article in Journal
Research on the Relationship between Resistivity and Resistance between Two Points on RCS Test Model
Next Article in Special Issue
Model-Driven Engineering Techniques and Tools for Machine Learning-Enabled IoT Applications: A Scoping Review
Previous Article in Journal
Starch-g-Acrylic Acid/Magnetic Nanochitin Self-Healing Ferrogels as Flexible Soft Strain Sensors
Previous Article in Special Issue
Technological Transformation of Telco Operators towards Seamless IoT Edge-Cloud Continuum
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inertial Sensor-Based Sport Activity Advisory System Using Machine Learning Algorithms

1
Institute of Mechanical Engineering, University of Zielona Góra, 65-417 Zielona Gora, Poland
2
Institute for Machine Tools and Production Processes, Chemnitz University of Technology, 09126 Chemnitz, Germany
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(3), 1137; https://doi.org/10.3390/s23031137
Submission received: 15 December 2022 / Revised: 13 January 2023 / Accepted: 17 January 2023 / Published: 19 January 2023
(This article belongs to the Special Issue Feature Papers in the Internet of Things Section 2022)

Abstract

:
The aim of this study was to develop a physical activity advisory system supporting the correct implementation of sport exercises using inertial sensors and machine learning algorithms. Specifically, three mobile sensors (tags), six stationary anchors and a system-controlling server (gateway) were employed for 15 scenarios of the series of subsequent activities, namely squats, pull-ups and dips. The proposed solution consists of two modules: an activity recognition module (ARM) and a repetition-counting module (RCM). The former is responsible for extracting the series of subsequent activities (so-called scenario), and the latter determines the number of repetitions of a given activity in a single series. Data used in this study contained 488 three defined sport activity occurrences. Data processing was conducted to enhance performance, including an overlapping and non-overlapping window, raw and normalized data, a convolutional neural network (CNN) with an additional post-processing block (PPB) and repetition counting. The developed system achieved satisfactory accuracy: CNN + PPB: non-overlapping window and raw data, 0.88; non-overlapping window and normalized data, 0.78; overlapping window and raw data, 0.92; overlapping window and normalized data, 0.87. For repetition counting, the achieved accuracies were 0.93 and 0.97 within an error of ±1 and ±2 repetitions, respectively. The archived results indicate that the proposed system could be a helpful tool to support the correct implementation of sport exercises and could be successfully implemented in further work in the form of web application detecting the user’s sport activity.

1. Introduction

The effects of constantly changing lifestyles are leading to new challenges and potentials in daily life. Thus, the use of digital technologies is increasingly gaining importance in the fields of physiotherapy, fitness and health care in monitoring human activities in free training without additional supervision. The focus here on the research side is the usage of progressive sensor technologies based on multisensors in combination with machine learning (ML) tools for accurate sport activity recognition, as well as to improve computing power in order to build useful and practical tools [1].
For sport activity recording, the following wearables can be distinguished: chromatic smart glasses (for activity tracking for fitness purposes), smart watches (mostly integrated with headphones) and smart bands (for tracking the fitness of the user) [2]. Naturally, the accuracy of the measurements is a main issue; therefore, such devices can be complemented by new sensors. Inertial sensors have been widely applied in the study of human activity recognition (HAR), as well as in the context of sport activities. The authors of [3] analyzed the usage of inertial sensors in sport. It was found that further development is needed to support the so-called standardization of data collection and analysis procedures for physical activities [4] and to find a balance between the accuracy and practicability of the proposed solution. Moreover, although detecting the user’s activity and environment is beneficial, it is not yet sufficient to provide personalized advice for a specific type of sport.
Therefore, the goal of this research is to use inertia-based sensing analysis to provide sport activity recognition based on the example sport exercises of squats, pull-ups and dips and build a tool supporting training without a personal trainer through system feedback. It is known that applying artificial intelligence (AI) enables a system to support the automation of activities such as identification of correctly/incorrectly performed sport activities [5]. According to the statement that AI methods can help in training monitoring and training schedule optimization [6], in previous work [7], a critical analysis of the literature on this topic was performed from the point of view of convolutional neural network (CNNs) and of CNNs combined with classifiers for HAR applications in sports. In this work, we provide an overview of current solutions that support some feedback features for the correct performance of sports exercises with inertial sensors and CNNs (Table 1).
Table 1 presents examples of physical activity advisory systems found in the currently literature in order to enrich the discussion with a rapid overview of the main outcome of the presented systems with respect to the analyzed state of the art. In particular, Table 1 reports information on the related works about (a) the approach to advising on the realization of physical activity, (b) the adoption of wearables for activity recording, (c) the applied AI methods, and (d) the exploitation of the type of sport and the achieved accuracy. To the best of our knowledge and as already highlighted in the state-of-the-art analysis, no existing studies have developed a system supporting the correct implementation of sport exercises, namely squats, pull-ups and dips, using mobile sensors (tags) and stationary anchors in combination with a system-controlling server (gateway) and for data processing, i.e., a convolutional neural network (CNN) classifier with an additional post-processing block (PPB) and repetition counting. The authors of [31] suggested that research in the area of applications of AI in HAR in the context of the sport activities should focus on heterogeneous sensor fusion, combining expert knowledge with a combination of different AI methods to improve the accuracy of activity recognition systems. In previous research on sports activity detection using CNNs, only homogeneous signals with a certain type of activity were considered. According to the suggestion of the authors of [31] and to distinguish this work from previous studies, we applied CNN classifiers for the analysis of ongoing registered signals and for the classification of heterogeneous activity signals. In summary, the key contributions of this paper are as follows:
  • Employment of three mobile sensors (tags), six stationary anchors and a system-controlling server (gateway) for 15 scenarios of the series of subsequent activities, namely squats, pull-ups and dips, as wearables for activity recording;
  • Building a model supporting the identification of correctly performed activities of sport exercises, namely squats, pull-ups and dips, based on the data of 488 sport activities using a convolutional neural network (CNN) classifier with an additional post-processing block (PPB) and repetition counting;
  • Design of an activity recognition module (ARM) and repetition-counting module (RCM) as the integral parts of the proposed advisory system;
  • Achievement of satisfactory accuracy of the proposed system: CNN + PPB: non-overlapping window and raw data, 0.88; non-overlapping window and normalized data, 0.78; overlapping window and raw data, 0.92; overlapping window and normalized data, 0.87. For repetition counting, the achieved accuracies were 0.93 and 0.97 within an error of ±1 and ±2 repetitions, respectively, indicating that the proposed system is a helpful tool to support the correct implementation of sport exercises.
The remainder of this paper is organized as follows. In Section 2, we present an overview of our proposed advisory system for sport activities, namely squats, pull-ups and dips, as well as the process of data acquisition and analysis. In Section 3, we explores the experimental results and present the activity recognition module (ARM) and repetition-counting module (RCM) as the integral parts of the proposed advisory system. In Section 4, we discuss the achieved results, as well as directions for further works.

2. Materials and Methods

2.1. Overview of the System

The task of the advisory system considered in this paper is to supervise the correctness of the implementation of the training plan previously selected by the exercising person. The subjects perform the defined exercises (namely three activities: dips, pull-ups, and squats and breaks) in the developed research unit (Figure 1). The subjects wear three inertial measurement units (IMUs) on their chest, hand and foot. Therefore, the dataset includes three activities: dips and pull-ups with squats and breaks in between, with data collected in real time during exercises performed by eight participants. The recorded quantities include the axial size acceleration, rotation rate, magnetic field strength, the absolute position of the IMUs in the room, Euler angles, quaternions and pressure. A total of 20 measured quantities per IMU were received with a recording rate of 30 Hz. The maximum recording rate of the system is 100 Hz per tag, but performance problems occurred during data acquisition. The authors of [32] also showed that satisfactory classification of human activities can be achieved with sampling rates as low as 15 Hz. As in further work, the fusion of camera and sensor data is investigated, a 30 Hz recording rate was considered suitable. To develop a system supporting the identification of correctly performed activities of sport exercises, namely squats, pull-ups and dips, a validation module (VM) has to be designed. The proposed module consists of two main subsystems: an activity recognition module (ARM), which is responsible for extracting the series of subsequent sport activities, and a repetition-counting module (RCM), which determines the number of repetitions of a given activity in single series. Such an approach allows for the assessment of the activity compliance of the exercising person according to the previously established protocol (i.e., compliance with the scenario and the assumed number of repetitions). In our solution, the data are collected in real time, but in the presented solution, the posture of the exerciser is not detected. The purpose of the present research is to build a smart phone application acting as personal trainer, so this aspect was not taken into account in the advisory system presented in this paper.
Considering the originality of this work, an overview of the proposed advisory system supporting the correct implementation of sport exercises, namely squats, pull-ups and dips, is presented in Figure 1.
It is known that the methodology of modelling of the advisory system should be integrated with the use of scenario methods [33]. Therefore, our developed advisory system (Figure 1) includes: (1) data collection based on three mobile sensors (tags), six stationary anchors and a system-controlling server (gateway) during live sports exercise sessions (i.e., squats, pull-ups and dips); (2) 15 scenarios of the series of subsequent activities, namely squats, pull-ups and dips; (3) sport activity recognition using a convolutional neural network (CNN) with an additional post-processing block (PPB); (4) determination of the number of repetitions of a given activity; and (5) support for the correct implementation of sport exercises with the addition of new data.

2.2. Data Acquisition

A measurement system consisting of three mobile sensors (tags), six stationary anchors and a system-controlling server (gateway) developed by Pozyx was used to collect the data. The anchors are connected to the gateway via a patch cable, and the tags communicate via an ultra-wide band with the anchors distributed throughout the room. In addition to transmitting the inertial measurement variables, the system also determines the absolute position of the tags based on trilateration. This feature is particularly useful for monitoring several spatially distributed training stations. The data recorded by the tags are passed through the anchor lines to the gateway, which makes them available in the local network via MQTT. The recording and preprocessing of the MQTT data stream was achieved with MATLAB R2020b. Preprocessing consists of linear interpolation of the data when the transmission rate falls below 30 Hz and temporal synchronization of the three recorded sensor data streams. After preprocessing, the recorded data are available in an N × M matrix, where N corresponds to the number of samples and M represents the number of measured variables. During real-time data collection, the three exercises were performed by the subjects in several sequences, and one or more exercises could be performed within a sequence. The number of repetitions and the length of the pauses between individual sports exercises were carried out at the individual discretion of the test subjects. The subjects could also decide whether to perform all three exercises or only a selection thereof.

2.3. Dataset Analysis

The exercises were carried out in accordance with the protocols specifying the scenarios (i.e., sequences of activities) and the length of series of each activity (i.e., number of repetitions). It should be noted that the protocols were developed based on the activities performed sport by test persons. The exercises were carried out according to the sporting abilities of the test persons, and the order in which the exercises were completed was also decided by the test persons. The participants realized 78 protocols according to 15 scenarios with the number of activity repetitions ranging between 2 and 15. The used scenarios and the number of protocols are summarized in Table 2.
An example of raw acceleration signals from the sensor placed on chest of a participant registered while performing a protocol realized in accordance with the scenario comprising the pull-ups, five dips and five squats is presented in Figure 2.
The raw signals collected during realizations of 78 protocols were split into parts containing one type of activity (break, dip, pull-up, squat). In further considerations, each of the parts obtained in this way is referred to as a single occurrence of the activity. The distribution of collected data is presented in Table 3 and Table 4 and Figure 3 (in all cases, time is expressed as the number of samples). The number of activity occurrences and the statistics of time duration in all scenarios within the protocols are shown in Table 3. The minimum and maximum number of repetitions and statistics of time duration of one activity repetition are presented in Table 4. In Figure 3, box-and-whisker plots show the distribution of total time and time per repetition based on the obtained dataset. Such analysis allows for determination of ranges of duration (expressed as the number of samples) for the occurrence of each activity, as well as the duration of one repetition. The parameters calculated in this way were used to tune the ARM and RCM modules of the advisory system described in detail in Section 3.
As shown in the Table 3 and Table 4, dataset is unbalanced because the distribution of activities is unequal, and the number of series of squats is approximately equal to the number of remaining activities (i.e., dips and pull-ups). A similar relationship can be observed in the duration of series (i.e., the number of registered samples); however, in this case, the differences are not so clear. The length of series (i.e., the number of repetitions) is similar for all activities and varies between 1 and 15. It takes the least amount of time to complete one squat, and the longest time is needed to complete one pull-up.

3. Results

As mentioned in the previous section, the proposed validation module (VM) that is a part of the designed advisory system (Figure 1) consists of two main subsystems. The first (ARM) is based on a CNN classifier and is responsible for the extraction of the scenario (the series of subsequent activities) performed by the exercising person. The second subsystem (RCM) uses information from the ARM and raw signals from sensors to determine the number of repetitions of a given activity. The scheme of the designed module is shown in Figure 4, and a detailed description of both subsystems is presented in the following subsections.
In order to design the ARM unit, a low-computational-cost CNN classifier based on our previous work is proposed. Owing to the nature of the task, in this work, the classifier was modified to process a live signal; moreover, it was supplemented with an additional post-processing block (PPB), enabling correct identification of the performed scenario. The second unit of the system (RCM) uses the known technique of finding peaks; however, the proposed approach to preprocessing of raw signals and tuning the parameters of the RCM block based on statistics of registered protocols is an original concept of the authors.

3.1. Activity Recognition Module

In order to design the ARM (Figure 4), an approach involving the use of a convolutional neural network (CNN) is proposed. The concept of this module is based on the results of our previous research on human activity recognition (HAR). In our previous work [7], a solution utilizing the technique known as image recognition [1,3] was proposed. The main goal of this work was to find a classifier structure demanding the lowest possible computational effort for the forward pass able to solve the HAR problem. A series of experiments involving a CNN with different hyperparameters allowed us to choose a classifier with one convolutional layer, five filters and a 7 × 1 kernel size. In [34], further research was conducted on decreasing computational effort by identification of significant sensor signals leading to input signal reduction. The studies focused on two issues: selecting the signals of greatest importance and sensor location. To evaluate the significance of the signals, an approach based on a comparison of the energy of signals transformed by a convolutional layer was used. In order to select sensor locations, the effectiveness of CNN classifiers for different sensor configurations was analyzed. As follows from the obtained results, the HAR problem can be satisfactorily solved using two acceleration sensors placed on the chest and the hand, which allowed for a reduction in the original set of 60 inputs to a set of only 6 signals. The results achieved in cited works constituted the basis for the studies presented in this paper. Finally, the classifier used in the experiments presented below is composed of one convolutional layer, five filters, a 7 × 1 kernel size, a ReLU activation function, a downsampling layer and a max pooling-type, flattened and dense output layer. The classifier input consists of 50 samples obtained from 6 acceleration signals from sensors placed on the chest and hand. The general structure of the employed CNN classifier is presented in Figure 5, and its parameters are listed in Table 5.
An important problem that had to be taken into account in the studies presented in this paper was the need to use a classifier for analysis of a live stream of data. In [7,34], the focus was on activity detection, and the input data of the CNN classifier contained only homogeneous signals covering one type of activity in both the training and testing phases. In contrast, in the presented work, the application of a CNN classifier for analysis of ongoing registered signals was considered, so the CNN had to also properly classify heterogenous activity signals. For this reason, two approaches for training the classifier were implemented and compared. In both cases, in order to prepare input data, the sliding-window method was used. In this technique, signals from sensors were sliced in the time domain to form 2D windows including 50 successive samples from 6 acceleration signals. In the first case, a non-overlapping window with homogenous signals presented in [34] was used; in the second case, a successive window overlap was used that can contain heterogenous activity signals. Additionally, in both cases, the classifier was trained and tested using raw and normalized signals.
The number of input data points (windows) used in training and testing phases is presented in Table 6. In the case of the overlapping window approach, the overlap of half the window size (25 samples) was used, and the number of windows including homogenous and heterogenous samples is presented separately. Based on assumption, each exercise scenario includes breaks between two activities, and according to Table 3, the duration the shortest break is equal to 63 samples, so heterogenous windows contain only single activities and breaks. For this reason, such data are treated as corresponding activities.
All experiments presented in this work were performed using algorithms implemented in Python 3.7.12. To create CNN classifiers with the structure shown in Figure 5, TensorFlow and Keras libraries in (version 2.7.0) were utilized. The code was run in the Google Colab environment with GPU support. In all experiments, datasets were split in a ratio of 70:15:15 between training, validation and test sets, respectively. The validation set with the early stopping approach was used to avoid classifier overtraining. In the case of the dataset created using the overlapping window method, the desired responses of the classifier were determined based on the percentage of samples corresponding to the activities registered within the window (i.e., for heterogenous data outputs less than 1). As a result of the training process based on the dataset created using the overlapping and non-overlapping window methods and raw and normalized data, four classifiers were obtained: C1 (non-overlapping window, raw data), C2 (non-overlapping window, normalized data), C3 (overlapping window, raw data) and C4 (overlapping window, normalized data). The accuracies of training and testing phases for these classifiers are presented in Table 7.
In all cases, the obtained results are satisfactory; however, the classifiers based on datasets created using a non-overlapping window achieve better accuracy in both the training and testing phases. Additionally, the accuracies obtained for normalized data are approximately equal (to the second decimal place) in contrast to raw data, which achieve significantly worse testing accuracy. The results presented in Table 7 allow for assessment of the quality of obtained classifiers, but they do not allow for the assessment of the effectiveness of classifiers operating as ARM, which should process the live stream of data.
In real applications, the samples registered by the sensors arrive to ARM in real time and must be immediately processed by the classifier. Such a scenario can be interpreted as a sliding of the overlapping window in the time domain with a step equal to one sample. An exemplary result of live signal processing by classifier C4 to recognize the scenario in the protocol shown in Figure 2 is presented in Figure 6. The subsequent plots show output signals from the classifier corresponding to breaks, dips, pull-ups and squats. The gray line shows the raw output signal (probability of activity occurring), the orange line represents the output signal after conversion to binary classes and the green line describes the scenario used in the realized protocol. A simple conversion of the output signals to binary classes is insufficient because there are short periods of time (few subsequent samples) during which the activities are not correctly recognized and, as a consequence, converted to output signals containing peaks and gaps.
In order to eliminate the disadvantages mentioned above, the ARM was expanded with an additional post-processing block (PPB) for CNN output signals. As a result of experimentation, the two-stage algorithm was chosen. In the first step, each raw output signal of the CNN was smoothed with a moving average filter (MAF) according to dependency (1).
y ˜ i = 1 w j = 0 w 1 y i + j ,
where y i and y ˜ i are the raw and filtered values, respectively, of the CNN output signal corresponding to the i -th sample, and w is the size of the window used by MAF. The output signal filtered in this way was converted to binary classes, and such a piecewise constant signal was processed in order to remove constant segments of insufficient length. For this reason, in the presented work, the eliminating peaks and gaps (EPG) algorithm was proposed according to dependency (2):
Y i = v i f       j     i P j   1 v o t h e r w i s e ,
where Y i is the output signal processed by EPG, v is equal to 1 or 0 for the EPG used to remove peaks or gaps, P = P 1 , , P j , , P n is a set of intervals with a constant signal (1 for removing peaks and 0 for removing gaps) with the lengths exceeding or threshold number of samples) for removing peaks and gaps, respectively.
In the presented work, parameters of filters described by dependencies (1) and (2) were assumed based on statistical measures of registered protocols shown in Table 3 as follows: w = 50 , δ 1 = 90 , δ 0 = 60 . To prepare a set of intervals ( P ) a function find_peaks from scipy.signal of Python module was used. The result of the CNN output signal shown in Figure 6 transformed by the PPB filters described above is presented in Figure 7.
As shown in Figure 7, after transformation by PPB using the MAF filter and binarization (olive line), most of the peaks and gaps in the output signal of the classifier shown in Figure 6 are removed. There are only a few peaks and gaps in the dip activity signal recognized incorrectly as breaks. After the EPG filter is added, the scenario is recognized correctly (orange line); however, slight shifts in time with respect to the protocol (green line) can be observed.
Finally, the accuracies of the considered classifiers and complete ARM unit (CNN + PPB) in activity recognition (Table 8) and scenario recognition (Table 9) were determined. The application of a post-processing block significantly improved the accuracy of both activity and scenario recognition. It is worth noting that the use of PPB significantly improved scenario recognition, which is especially important in the case of a system supporting the correct implementation of sport exercises. Based on the presented results, the ARM with a C3 classifier (overlapping window, raw data) was chosen as the component of the designed system.

3.2. Repetition-Counting Module

The second module of the designed system (Figure 4) has to calculate number of repetitions for each activity using information about activities provided by the ARM unit and raw signals registered by sensors. For this reason, in this study, the approach based on finding peaks in the original signals was used. Owing to the nature of the acquired signals (see Figure 2), their direct use for repetition counting led to overestimated results as a consequence of detecting false peaks. In order to eliminate this problem, preprocessing of original signals was performed.
The preprocessing of raw signals should remove the noise resulting from sensor characteristics and accidental/unnecessary movements of the exercising person. To solve this problem, the application of a low-pass filter to attenuate signals with frequencies higher than the cutoff frequency was proposed. In this study, the Butterworth filter was chosen. The significant feature of this filter in eliminating false peaks is a maximally flat frequency response in the passband and near-zero response in the stopband. In this application, cutoff frequencies result from the time of performing single exercises. They were determined individually for each activity based on statistics presented in Table 4 and are listed in Table 10. The acceleration signals from chest and hand sensors registered during the performance of dips according to the protocol presented in Figure 2 before and after filtering are shown in Figure 8.
In the discussed system, each acceleration sensor provides three signals; nevertheless, as shown in Figure 2 and Figure 8, some signals do not contain information relevant to repetition counting (i.e., x-axis signal from the chest). To eliminate irrelevant signals, a series of experiments were carried out, and the chosen signals are presented in Table 10. Using data from sensors preprocessed and selected as described above, the number of repetitions of each activity was determined by counting the number of peaks of the signals. The parameters defining the minimum distances between the peaks required for correct performance of this operation were determined based on the minimum time needed to complete one repetition (Table 4) and are presented in last column of Table 10. Finally, the number of repetitions of each activity was determined as the average number of peaks across all previously chosen relevant signals.
To assess the effectiveness of the proposed system to supervise the correctness of the implementation of the training plan, a summary of obtained results is presented in Table 11 and Table 12. Table 11 contains the number of protocols, repetitions and miscounts grouped by activity. A low total number of miscounts was achieved, calculated as the total number of miscounts relative to the total number of repetitions (6%; 68/1105). The accuracies obtained in counting the repetition for each activity are shown in Table 12. Three types of accuracy were calculated: exact (ratio of correctly detected repetitions to the total number of repetitions), with errors of ±1 and ±2 repetitions. Accuracies within ±1 and ±2 equal 0.93 and 0.97, respectively, which are satisfactory, especially because they were determined based on scenario recognition performed by the ARM unit.

4. Discussion

In this work, an advisory system was developed to evaluate the correctness of the execution of sports activities (squats, pull-ups and dips). Each exercise was carried out according to the sporting abilities of the test persons, and the order in which the exercises were completed was also decided by the test persons. Based on the results, 15 scenarios of the series of subsequent activities were developed. Using three mobile sensors and six stationary anchors in combination with a system-controlling server (gateway), the data containing 488 sport activity occurrences were acquired. The proposed solution consists of two modules: an activity recognition module (ARM) and a repetition-counting module (RCM). The ARM, as a combination of a classifier and a post-processing block, showed very good performance in the recognition of three sports exercises with pauses in between. Squats were best-detected, which can be attributed to the predominant amount of data from squat intervals. Depending on the activity, classifiers C1 to C4 achieved different results. This should be investigated again with a more comprehensive dataset. The PPB block is an important component for the live recognition of sports exercises, especially when feedback functionalities have to be implemented.
Our results indicate that the developed advisory system can be treated as suitable tool to support the correct implementation of sport exercises. The developed system (CNN + PPB) achieved satisfactory accuracy:
  • Non-overlapping window and raw data: 0.88;
  • Non-overlapping window and normalized data: 0.78;
  • Overlapping window and raw data: 0.92;
  • Overlapping window and normalized data: 0.87.
  • Repetition counting: 0.93 within an error of ±1 and 0.97 within an error of ±2.
The research results are comparable to another solution, and a higher recognition accuracy (0.97) was achieved when using repetition counting (see Table 1). The originality of this work is the combination of a convolutional neural network (CNN) with an additional post-processing block (PPB) and repetition counting. In addition, CNNs were used for the analysis of heterogeneous activity signals. The main limitation of the proposed system is that the protocols specifying the scenarios (i.e., sequences of activities) were developed based on the on the sport activities performed by test persons. This process of creating sport activity scenarios should be fully prepared by a sports trainer to guarantee the correctness of the exercises performed. In further work, the scenarios specifying the correctness of each sport activity will be formulated and added. The workflow and limitations described in this paper can also serve as a basis for further work in order to implement our advisory system in the form of a web application. This application will be a system acting as a personal trainer supervising the correctness of training in accordance with a previously prepared protocol (order, number and pace of exercises) using only sensors available in commonly used mobile devices (e.g., smart phones or smart watches).

5. Conclusions

In summary, the proposed advisory system that processes the data of 488 sports activities (squats, pull-ups and dips) collected from mobile sensors and stationary anchors provided through a system-controlling server (gateway) using a convolutional neural network (CNN) with an additional post-processing block (PPB) followed by repetition counting is a promising solution for the generation of a variety of conditions, providing the basis for a statement about the correctness of the performed sports exercise.
In a future work, sport activity scenarios prepared by a sports trainer will be applied, and experiments based on the new data will be conducted. Thus, the scenarios, new dataset and further research experiments on various neural networks to achieve better accuracy and to reduce the computational efforts would be especially beneficial for the development of a system in the form of, e.g., a web application available for users to supervise the performance of exercises without the need for the physical presence of a trainer.

Author Contributions

Conceptualization, J.P.-M. and I.P.; formal analysis, M.D.; funding acquisition, J.P.-M. and M.D.; investigation, J.P.-M., I.P., P.K., G.P., M.R. and H.S.; methodology, J.P.-M., I.P., P.K., G.P., M.R. and H.S.; project administration, M.D.; resources, P.K., H.S. and M.D.; software, I.P., P.K. and G.P.; supervision, J.P.-M., H.S. and M.D.; validation, I.P., G.P., M.R. and H.S.; writing—original draft, J.P.-M., I.P., P.K., G.P., M.R. and H.S.; writing—review and editing, J.P.-M., I.P., P.K., G.P., M.R., H.S. and M.D. All authors have read and agreed to the published version of the manuscript.

Funding

Parts of this publication were funded by the Federal German Ministry for Economic Affairs and Energy. Sensors 23 01137 i001

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of Chemnitz University of Technology (protocol code #101556284 by 14 July 2022.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Y.; Chen, R.; Niu, X.; Zhuang, Y.; Gao, Z.; Hu, X.; El-Sheimy, N. Inertial Sensing Meets Machine Learning: Opportunity or Challenge? IEEE Trans. Intell. Transp. Syst. 2022, 23, 9995–10011. [Google Scholar] [CrossRef]
  2. Nahavandi, D.; Alizadehsani, R.; Khosravi, A.; Acharya, U.R. Application of artificial intelligence in wearable devices: Opportunities and challenges. Comput. Methods Programs Biomed. 2022, 213, 106541. [Google Scholar] [CrossRef]
  3. Camomilla, V.; Bergamini, E.; Fantozzi, S.; Vannozzi, G. Trends supporting the in-field use of wearable inertial sensors for sport performance evaluation: A systematic review. Sensors 2018, 18, 873. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Cummins, C.; Orr, R. Analysis of physical collisions in elite national rugby league match play. Int. J. Sports Physiol. Perform. 2015, 10, 732–739. [Google Scholar] [CrossRef]
  5. Baker, S.B.; Xiang, W.; Atkinson, I. Internet of things for smart healthcare: Technologies, challenges, and opportunities. IEEE Access 2017, 5, 26521–26544. [Google Scholar] [CrossRef]
  6. Vidal, M.C.; Murphy, S.M. Quantitative measure of fitness in tri-trophic interactions and its influence on diet breadth of insect herbivores. Ecology 2018, 99, 2681–2691. [Google Scholar] [CrossRef] [PubMed]
  7. Pajak, G.; Krutz, P.; Patalas-Maliszewska, J.; Rehm, M.; Pajak, I.; Dix, M. An approach to sport activities recognition based on an inertial sensor and deep learning. Sens. Actuators A Phys. 2022, 345, 113773. [Google Scholar] [CrossRef]
  8. Hussain, A.; Zafar, K.; Baig, A.R.; Almakki, R.; AlSuwaidan, L.; Khan, S. Sensor-Based Gym Physical Exercise Recognition: Data Acquisition and Experiments. Sensors 2022, 22, 2489. [Google Scholar] [CrossRef]
  9. Bian, S.; Rey, V.F.; Hevesi, P.; Lukowicz, P. Passive Capacitive based Approach for Full Body Gym Workout Recognition and Counting. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications, Kyoto, Japan, 11–15 March 2019; pp. 1–10. [Google Scholar] [CrossRef]
  10. Fu, B.; Kirchbuchner, F.; Kuijper, A.; Braun, A.; Vaithyalingam Gangatharan, D. Fitness Activity Recognition on Smartphones Using Doppler Measurements. Informatics 2018, 5, 24. [Google Scholar] [CrossRef]
  11. Balkhi, P.; Moallem, M. A Multipurpose Wearable Sensor-Based System for Weight Training. Automation 2022, 3, 132–152. [Google Scholar] [CrossRef]
  12. Soro, A.; Brunner, G.; Tanner, S.; Wattenhofer, R. Recognition and Repetition Counting for Complex Physical Exercises with Deep Learning. Sensors 2019, 19, 714. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Burns, D.; Leung, N.; Hardisty, M.; Whyne, C.M.; Henry, P.; McLachlin, S. Shoulder physiotherapy exercise recognition: Machine learning the inertial signals from a smartwatch. Physiol. Meas. 2018, 39, 075007. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Dib, W.; Ghanem, K.; Ababou, A.; Eskofier, B.M. Human Activity Recognition Based on the Fading Characteristics of the On-Body Channel. IEEE Sens. J. 2022, 22, 8094–8103. [Google Scholar] [CrossRef]
  15. Rojanavasu, P.; Jitpattanakul, A.; Mekruksavanich, S. Comparative Analysis of LSTM-based Deep Learning Models for HAR using Smartphone Sensor. In Proceedings of the Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunication Engineering, Nanjing, China, 3–5 December 2021; pp. 269–272. [Google Scholar] [CrossRef]
  16. Golestani, N.; Moghaddam, M. Magnetic Induction-based Human Activity Recognition (MI-HAR). In Proceedings of the IEEE International Symposium on Antennas and Propagation and USNC-URSI Radio Science Meeting, Atlanta, GA, USA, 7–12 July 2019; pp. 17–18. [Google Scholar] [CrossRef]
  17. Ghazali, N.F.; Sharar, N.; Rahmad, N.A.; Sufri, N.A.J.; As’ari, M.A.; Latif, H.F.M. Common Sport Activity Recognition using Inertial Sensor. In Proceedings of the IEEE 14th International Colloquium on Signal Processing & its Applications, Batu Feringghi, Malaysia, 9–10 March 2018; pp. 67–71. [Google Scholar]
  18. Steels, T.; Herbruggen, B.V.; Fontaine, J.; Pessemier, T.D.; Plets, D.; Poorter, E.D. Badminton Activity Recognition Using Accelerometer Data. Sensors 2020, 20, 4685. [Google Scholar] [CrossRef] [PubMed]
  19. Kautz, T.; Groh, B.H.; Hannink, J.; Jensen, U.; Strubberg, H.; Eskofier, B.M. Activity recognition in beach volleyball using a Deep Convolutional Neural Network. Data Min. Knowl. Discov. 2017, 31, 1678–1705. [Google Scholar] [CrossRef]
  20. Kiazolu, G.D.; Aslam, S.; Ullah, M.Z.; Han, M.; Weamie, S.J.Y.; Miller, R.H.B. Location-Independent Human Activity Recognition Using WiFi Signal. In Signal and Information Processing, Networking and Computers; Sun, J., Wang, Y., Huo, M., Xu, L., Eds.; Lecture Notes in Electrical, Engineering; Springer: Singapore, 2023; p. 917. [Google Scholar] [CrossRef]
  21. Ayvaz, U.; Elmoughni, H.; Atalay, A.; Atalay, Ö.; Ince, G. Real-Time Human Activity Recognition Using Textile-Based Sensors. In Body Area Networks. Smart IoT and Big Data for Intelligent Health. BODYNETS 2020; Alam, M.M., Hämäläinen, M., Mucchi, L., Niazi, I.K., Le Moullec, Y., Eds.; Lecture Notes of the Institute for Computer, Sciences; Springer: Cham, Switzerland, 2020; p. 330. [Google Scholar] [CrossRef]
  22. Wang, H.; Li, L.; Chen, H.; Li, Y.; Qiu, S.; Gravina, R. Motion Recognition for Smart Sports Based on Wearable Inertial Sensors. In Body Area Networks: Smart IoT and Big Data for Intelligent Health Management; Mucchi, L., Hämäläinen, M., Jayousi, S., Morosi, S., Eds.; Lecture Notes of the Institute for Computer, Sciences; Springer: Cham, Switzerland, 2019; Volume 297. [Google Scholar] [CrossRef]
  23. Biagetti, G.; Crippa, P.; Falaschetti, L.; Orcioni, S.; Turchetti, C. Human Activity Recognition Using Accelerometer and Photoplethysmographic Signals. In Intelligent Decision Technologies; Czarnowski, I., Howlett, R., Jain, L., Eds.; Springer: Cham, Switzerland, 2017; Volume 73. [Google Scholar] [CrossRef]
  24. Mehrang, S.; Pietila, J.; Tolonen, J.; Helander, E.; Jimison, H.; Pavel, M.; Korhonen, I. Human Activity Recognition Using A Single Optical Heart Rate Monitoring Wristband Equipped with Triaxial Accelerometer. In EMBEC & NBC; Eskola, H., Väisänen, O., Viik, J., Hyttinen, J., Eds.; Springer: Singapore, 2017; Volume 65. [Google Scholar] [CrossRef]
  25. Shahar, N.; Ghazali, N.F.; As’ari, M.A.; Tan, T.S.; Ibrahim, M.F. Investigation of Different Time-Series Segmented Windows from Inertial Sensor for Field Hockey Activity Recognition. In Enhancing Health and Sports Performance by Design; Springer: Singapore, 2019. [Google Scholar] [CrossRef]
  26. Polo-Rodriguez, A.; Montoro-Lendinez, A.; Espinilla, M.; Medina-Quero, J. Classifying Sport-Related Human Activity from Thermal Vision Sensors Using CNN and LSTM. In Image Analysis and Processing; Mazzeo, P.L., Frontoni, E., Sclaroff, S., Distante, C., Eds.; Springer: Cham, Switzerland, 2022; Volume 13373. [Google Scholar] [CrossRef]
  27. Chakraborty, A.; Mukherjee, N. A deep-CNN based low-cost, multi-modal sensing system for efficient walking activity identification. Multimed. Tools Appl. 2022, in press. [Google Scholar] [CrossRef]
  28. Jang, Y.; Kim, S.; Kim, K.; Lee, D. Deep learning-based classification with improved time resolution for physical activities of children. PeerJ 2018, 6, e5764. [Google Scholar] [CrossRef] [Green Version]
  29. Mekruksavanich, S.; Jitpattanakul, A.; Youplao, P.; Yupapin, P. Enhanced Hand-Oriented Activity Recognition Based on Smartwatch Sensor Data Using LSTMs. Symmetry 2020, 12, 1570. [Google Scholar] [CrossRef]
  30. Javed, A.R.; Sarwar, M.U.; Khan, S.; Iwendi, C.; Mittal, M.; Kumar, N. Analyzing the Effectiveness and Contribution of Each Axis of Tri-Axial Accelerometer Sensor for Accurate Activity Recognition. Sensors 2020, 20, 2216. [Google Scholar] [CrossRef] [Green Version]
  31. Nweke, H.F.; Teh, Y.W.; Al-garadi, M.A.; Alo, U.R. Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges. Expert Syst. Appl. 2018, 105, 233–261. [Google Scholar] [CrossRef]
  32. Khan, A.; Hammerla, N.; Mellor, S.; Ploetz, T. Optimising sampling rates for accelerometer-based human activity recognition. Pattern Recognit. Lett. 2016, 73, 33–40. [Google Scholar] [CrossRef]
  33. Antosz, K.; Pasko, L.; Gola, A. The Use of Intelligent Systems to Support the Decision-Making Process in Lean Maintenance Management. IFAC-Pap. 2019, 52, 148–153. [Google Scholar] [CrossRef]
  34. Pajak, I.; Krutz, P.; Patalas-Maliszewska, J.; Rehm, M.; Pajak, G.; Schlegel, H.; Dix, M. Sports activity recognition with UWB and inertial sensors using deep learning approach. In Proceedings of the IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar] [CrossRef]
Figure 1. Schematic representation of the advisory system supporting the correct implementation of sport exercises, namely squats, pull-ups and dips.
Figure 1. Schematic representation of the advisory system supporting the correct implementation of sport exercises, namely squats, pull-ups and dips.
Sensors 23 01137 g001
Figure 2. Raw acceleration signals from a sensor placed on the chest of a participant performing three pull-ups, five dips and five squats.
Figure 2. Raw acceleration signals from a sensor placed on the chest of a participant performing three pull-ups, five dips and five squats.
Sensors 23 01137 g002
Figure 3. (a) Distribution of total time; (b) distribution of time per repetition of a given activity.
Figure 3. (a) Distribution of total time; (b) distribution of time per repetition of a given activity.
Sensors 23 01137 g003
Figure 4. The scheme of the validation module.
Figure 4. The scheme of the validation module.
Sensors 23 01137 g004
Figure 5. The general structure of the CNN classifier.
Figure 5. The general structure of the CNN classifier.
Sensors 23 01137 g005
Figure 6. An exemplary result of live signal processing.
Figure 6. An exemplary result of live signal processing.
Sensors 23 01137 g006
Figure 7. CNN output signal transformed by PPB filters.
Figure 7. CNN output signal transformed by PPB filters.
Sensors 23 01137 g007
Figure 8. Acceleration signals from chest and hand sensors before and after filtering.
Figure 8. Acceleration signals from chest and hand sensors before and after filtering.
Sensors 23 01137 g008
Table 1. A review of currently available solutions supporting the performance of sport exercises.
Table 1. A review of currently available solutions supporting the performance of sport exercises.
Physical Activity Advisory SystemApplied Wearables for Activity RecordingType of SportApplied AIAchieved
Accuracy
Ref.
Gym physical exercise recognition systemSingle chest-mounted triaxial accelerometerSix muscle groups, gym exerciseLong short-term memory (LSTM) neural networksNon-overlapping dataset: 0.57–0.81; overlapping dataset: 0.74–0.91[8]
System for weight trainingAn inertial measurement unit (IMU), an accelerometer and three force sensorsWeightliftingNearest neighbor (KNN), decision tree (DT), random forest (RF) and repetition countingKNN: 0.97; DT: 0.99;
SVM: 0.98;
RF: 100;
repetition counting: 0.96
[9]
Recognizing gym workoutsA micro-watt-level power consumption human body capacitance based sensorSeven gym workoutsRepetition counting0.91[10]
Fitness activity recognitionMicrophone integrated into a smartphoneBody weight exercises, bicycles, toe touches and squatsSupport vector machines (SVMs) and convolutional neural networks (CNNs)0.88 for bicycles, 0.97 for toe touches and 0.9 for squats[11]
Exercises in CrossFit recognitionSmart watches in communication with each other over Bluetooth low-energy (BLE)10 CrossFit exercisesClassifiers of a series of 2D convolutions and two fully connected layers, as well as
repetition counting
0.99;
repetition counting: 0.91
[12]
Shoulder physiotherapy exercise recognitionSix-axis inertial sensorSeven shoulder exercisesK-nearest neighbor (k-NN), random forest (RF), support vector machine (SVM) and convolutional recurrent neural network (CRNN)Classification algorithms: 0.94; CRNN algorithm: 0.99[13]
Static and dynamic activity recognitionTwo radio devices (waist- and ankle-worn) based on measurements of variations in the received signal strengthStatic (standing, sitting and lying) and dynamic activities (walking, running and dancing)K-nearest neighbor (k-NN), support vector machine (SVM) and a combination thereof0.99[14]
Daily activity recognitionRecorded with smart watch and smart phoneSix non-hand-oriented activitiesConvolutional neural networks (CNNs), long short-term memory (LSTM) and combinations thereofHybrid LSTM: 0.99[15]
Magnetic-induction-based human activity recognitionWireless system based on magnetic inductionWalking and knockingDeep recurrent neural networks (DRNNs)0.88
(synthetically generated motion dataset)
[16]
Sport activity recognitionIMU attached to the chestWalking, jogging, sprinting and jumpingDecision trees (DT), discriminant analysis, support vector machine (SVM) and k-nearest neighbor (k-NN)Cubic SVM: 0.91[17]
Badminton activity recognitionAccelerometer/gyroscope sensors attached to the wrist, upper arm and racket gripSeven different strokes and two movement typesConvolutional neural networks (CNNs) and Deep CNNsCNN: 0.99 (with accelerometer and gyroscope data)[18]
Beach volleyball activity recognitionThree-axis acceleration sensor worn on the dominant handTen action classes in beach volleyballDeep convolutional neural networks (DCNNs)DCNN: 0.83[19]
Motion activity recognition based on Wi-Fi signal analysisTwo laptops equipped with three omnidirectional antennae and analysis of channel-state informationNine motion activities (e.g., walking, paper toss and hand clap)Decision tree (DT), convolutional neural networks (CNNs) and long short-term memory (LSTM)DT: 0.94[20]
Activity recognition in real timeTextile-based capacitive sensors implemented as knee bracesWalking, standing, running and squattingSupport vector machine (SVM), decision tree (DT), k-nearest neighbor (k-NN) and random forest (RF)RF: 0.83[21]
Table tennis activity recognitionSix-axis accelerometer/gyroscope attached to the armFive typical table tennis strokesSupport vector machine (SVM) and k-nearest neighbor (k-NN)SVM: 0.96[22]
Activity recognition with additional photoplethysmographic signalsAccelerometer and PPG sensor dataWalking, running, cycling (with high and low resistance)Bayesian classifier0.78[23]
Activity recognitionHeart-rate-monitoring wrist band equipped with a triaxial accelerometerHome-specific activities (sitting, standing, household activities and stationary cycling)Support vector machine (SVM) and random forest (RF)SVM: 0.85;
RF: 0.89
[24]
Field hockey activity recognitionTwo IMUs worn on the chest and waistSix field hockey activitiesCubic support vector machine (SVM)0.91[25]
Fitness activity recognitionThermal vision sensorPush-ups, sit-ups, jumping jacks, squats and planksConvolutional neural networks (CNNs) for feature extraction and long short-term memory (LSTM) for final classification0.98[26]
Walking activity recognitionA heterogeneous sensor system: leg-worn IMU and finger-tip-based pulse sensorsWalking activity and leg-swing activitiesDeep convolutional neural network (DCNN)0.97[27]
Recognition of the physical activities of childrenThree-axis accelerometer modules around the waistSlow/fast walking, slow/fast running, walking up/down stairs, jumping rope, standing up and sitting downConvolutional neural network (CNN)0.81[28]
Hand-oriented activity recognitionTriaxial accelerometer data and triaxial gyroscopeJogging and walking (public benchmark dataset called WISDM)Convolutional neural network (CNN) and long short-term memory (LSTM)0.96[29]
Daily life activity recognitionTwo-axis smart phone accelerometer sensorJogging and walking (public benchmark dataset called WISDM)Multilayer perceptron (MLP) classifier0.93[30]
Table 2. The used scenarios and the number of protocols.
Table 2. The used scenarios and the number of protocols.
ScenarioNumber of Protocols
dips1
pull-ups2
squats7
dips, squats3
pull-ups, squats1
squats, dips1
squats, squats5
dips, pull-ups, squats5
dips, squats, pull-ups4
pull-ups, dips, squats28
pull-ups, squats, dips1
squats, dips, pull-ups4
squats, pull-ups, dips2
squats, squats, squats13
pull-ups, dips, squats, squats1
Table 3. The number of activity occurrences and the statistics of duration.
Table 3. The number of activity occurrences and the statistics of duration.
ActivityNumber of
Occurrences
Duration in Number of Samples
MinAvgMax
breaks283633821844
dips50135297768
pull-ups4891249765
squats107157474956
Total488633801844
Table 4. The number of repetitions and the statistics of the duration of one activity repetition.
Table 4. The number of repetitions and the statistics of the duration of one activity repetition.
ActivityRepetitionsDuration in Number of Samples
MinMaxMinAvgMax
dips2124879171
pull-ups1145590220
squats2153368130
Total1153376220
Table 5. Parameters of the CNN classifier.
Table 5. Parameters of the CNN classifier.
LayerLayer TypeOutput ShapeNumber of
Parameters
1.Convolution (5 filters, 7 × 1 kernel, ReLU)(44, 6, 5)5 × (7 + 1)
2.Max pooling(22, 6, 5)0
3.Flatten(660)0
4.Dense (softmax)(4)4 × (660 + 1)
Total parameters2684
Table 6. The number of input data points used in the training and testing phases for both approaches.
Table 6. The number of input data points used in the training and testing phases for both approaches.
ActivityNon-Overlapping WindowOverlapping Window
Homogenous DataHeterogenous Data
breaks20243812-
dips273493192
pull-ups215384189
squats9561810423
Total34687303
Table 7. The accuracies of training and testing phases for obtained classifiers.
Table 7. The accuracies of training and testing phases for obtained classifiers.
PhaseNon-Overlapping WindowOverlapping Window
Raw Data
Classifier C1
Normalized Data
Classifier C2
Raw Data
Classifier C3
Normalized Data
Classifier C4
training0.980.940.950.92
testing0.930.940.900.92
Table 8. The accuracies for single and all activities.
Table 8. The accuracies for single and all activities.
ActivityCNN ClassifiersARM (CNN + PPB)
C1C2C3C4C1C2C3C4
dips0.740.630.790.720.910.670.970.8
pull-ups0.770.870.780.90.910.930.910.95
squats0.90.930.860.920.950.970.940.97
Total0.850.860.830.880.940.910.940.93
Table 9. The accuracies for scenario recognition.
Table 9. The accuracies for scenario recognition.
CNN ClassifiersARM (CNN + PPB)
C1C2C3C4C1C2C3C4
0.000.130.000.210.880.780.920.87
Table 10. RCM parameters.
Table 10. RCM parameters.
ActivitySignalsButterworth
Filter
Find Peaks
ChestHand
X AxisY AxisZ AxisX AxisY AxisZ AxisOrder W n * Distance
dips20.0540
pull-ups20.0450
squats20.0630
* W n = f c / 0.5 f s —normalized cutoff frequency, f c —cutoff frequency, f s —sampling frequency, ☑ given signal is used in RCM module, □ given signal is not used in RCM module.
Table 11. Repetition count results.
Table 11. Repetition count results.
ActivityNo. of ProtocolsTotal
Repetition Count
Total
Miscount
dips5120220
pull-ups4914419
squats7575929
Total175110568
Table 12. Accuracies of repetition counting.
Table 12. Accuracies of repetition counting.
ActivityExactwithin ±1within ±2
dips0.780.900.94
pull-ups0.670.960.98
squats0.770.920.97
Total0.750.930.97
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Patalas-Maliszewska, J.; Pajak, I.; Krutz, P.; Pajak, G.; Rehm, M.; Schlegel, H.; Dix, M. Inertial Sensor-Based Sport Activity Advisory System Using Machine Learning Algorithms. Sensors 2023, 23, 1137. https://doi.org/10.3390/s23031137

AMA Style

Patalas-Maliszewska J, Pajak I, Krutz P, Pajak G, Rehm M, Schlegel H, Dix M. Inertial Sensor-Based Sport Activity Advisory System Using Machine Learning Algorithms. Sensors. 2023; 23(3):1137. https://doi.org/10.3390/s23031137

Chicago/Turabian Style

Patalas-Maliszewska, Justyna, Iwona Pajak, Pascal Krutz, Grzegorz Pajak, Matthias Rehm, Holger Schlegel, and Martin Dix. 2023. "Inertial Sensor-Based Sport Activity Advisory System Using Machine Learning Algorithms" Sensors 23, no. 3: 1137. https://doi.org/10.3390/s23031137

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop