Next Article in Journal
Wearable Feet Pressure Sensor for Human Gait and Falling Diagnosis
Next Article in Special Issue
Pulmonary COVID-19: Learning Spatiotemporal Features Combining CNN and LSTM Networks for Lung Ultrasound Video Classification
Previous Article in Journal
Convolutional Neural Networks for Challenges in Automated Nuclide Identification
Previous Article in Special Issue
Forecasting of Typhoon-Induced Wind-Wave by Using Convolutional Deep Learning on Fused Data of Remote Sensing and Ground Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic, Qualitative Scoring of the Clock Drawing Test (CDT) Based on U-Net, CNN and Mobile Sensor Data

Department of Electronic Engineering, Hallym University, Chuncheon 24252, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(15), 5239; https://doi.org/10.3390/s21155239
Submission received: 22 June 2021 / Revised: 26 July 2021 / Accepted: 29 July 2021 / Published: 3 August 2021
(This article belongs to the Special Issue Image Sensing and Processing with Convolutional Neural Networks)

Abstract

:
The Clock Drawing Test (CDT) is a rapid, inexpensive, and popular screening tool for cognitive functions. In spite of its qualitative capabilities in diagnosis of neurological diseases, the assessment of the CDT has depended on quantitative methods as well as manual paper based methods. Furthermore, due to the impact of the advancement of mobile smart devices imbedding several sensors and deep learning algorithms, the necessity of a standardized, qualitative, and automatic scoring system for CDT has been increased. This study presents a mobile phone application, mCDT, for the CDT and suggests a novel, automatic and qualitative scoring method using mobile sensor data and deep learning algorithms: CNN, a convolutional network, U-Net, a convolutional network for biomedical image segmentation, and the MNIST (Modified National Institute of Standards and Technology) database. To obtain DeepC, a trained model for segmenting a contour image from a hand drawn clock image, U-Net was trained with 159 CDT hand-drawn images at 128 × 128 resolution, obtained via mCDT. To construct DeepH, a trained model for segmenting the hands in a clock image, U-Net was trained with the same 159 CDT 128 × 128 resolution images. For obtaining DeepN, a trained model for classifying the digit images from a hand drawn clock image, CNN was trained with the MNIST database. Using DeepC, DeepH and DeepN with the sensor data, parameters of contour (0–3 points), numbers (0–4 points), hands (0–5 points), and the center (0–1 points) were scored for a total of 13 points. From 219 subjects, performance testing was completed with images and sensor data obtained via mCDT. For an objective performance analysis, all the images were scored and crosschecked by two clinical experts in CDT scaling. Performance test analysis derived a sensitivity, specificity, accuracy and precision for the contour parameter of 89.33, 92.68, 89.95 and 98.15%, for the hands parameter of 80.21, 95.93, 89.04 and 93.90%, for the numbers parameter of 83.87, 95.31, 87.21 and 97.74%, and for the center parameter of 98.42, 86.21, 96.80 and 97.91%, respectively. From these results, the mCDT application and its scoring system provide utility in differentiating dementia disease subtypes, being valuable in clinical practice and for studies in the field.

1. Introduction

As a sub-test of the Mini-Mental State Examination (MMSE), the Clock Drawing test (CDT) along with the Pentagon Drawing test (PDT) and the Rey–Osterrieth Complex Figure test (ROCF) have been widely used in neuropsychology and related areas for neurological and cognitive exams for dementias such as Alzheimer’s and others [1,2,3,4,5,6,7]. For the CDT, a qualitative neurological drawing test commonly used as a screening instrument for cognitive capabilities, a subject is asked to draw a clock showing a specific time. Placement of the numbers around the clock contour requires visual–spatial, numerical sequencing and planning abilities [8]. Drawing hands indicating a specific time requires long-term attention, memory, auditory processing, motor programming, and frustration tolerance [8]. Hence, in CDT, the test subject uses various cortical areas at the same time for the task being performed; these include use of the frontal, parietal, and temporal lobes [9,10]. As such, various cognitive skills such as selective, sustained attention, visuospatial skills, verbal working memory, visual memory and reconstruction, on-demand motor execution (praxis), auditory comprehension, numerical knowledge, and executive function can be tested [9,11,12]. It was Shulman and collaborators who employed the CDT as a screening tool for older patients for cognitive disorders and, since then, multiple studies have affirmed the utility of the CDT for screening and diagnosis of cognitive impairment for suspected pathologies such as Huntington’s disease, schizophrenia, unilateral neglect, delirium, multiple sclerosis, etc. [13]. Given the utility of CDT, studies on generating a scoring system for CDT are valuable.
To gauge visuo-constructional disorders in moderate and severe dementia, quantitative scoring methods had been developed for the CDT, and these scoring methods have indicated that CDT may be practical in allowing to distinguish the various clinical features of cognitive deficits [14,15,16,17]. However, to improve such methods, neuropsychological approaches using information processing and its qualitative aspects have become necessary, with the aims being to assess the executive functions for the task and analyzing errors in the execution of the drawing. Qualitative approaches analyze various error types such as graphing problems, conceptual deficits, stimulus-bound responses, spatial/planning deficits, and perseveration, and these have helped describe various dementia profiles [18,19]. Furthermore, the assessment of the CDT had been conventionally conducted manually by a medical expert based on identifying abnormalities in the drawings including poor number positioning, omission of numbers, incorrect sequencing, missing clock hands and the presence of irrelevant writing, which is labor intensive, complex and also prone to human subjective errors [20]. Thus, needs of an automatic scoring system have been increased. As such, qualitative and automatic scoring systems are helpful for differential diagnoses [21]. There are several ways of interpreting CDT quantitatively and/or qualitatively [22,23,24,25,26,27,28]. For several of these methods, the scoring involves assessment of drawing parameters, including the size of the clock, the closure of the contour, the circularity of the contour, the existence of the two hands, the proportion of the minute and the hour hands, the positioning of the hands, the correct target numbers of the hands according to the time-setting instruction, the presence of all the digit numbers, the correct position of the digit numbers, the order of drawing of the digit numbers, and the existence of the center in the hand-drawn images for CDT. However, the scoring of CDT is dependent on a clinician’s subjectivity, and so it is prone to be biased and have human errors. Furthermore, it is not very practical in analyzing big data, such as for personal lifelogs. With the CDT results being qualitative and not numeric, it is also difficult to evaluate objectively. Therefore, the necessity of a standardized, qualitative, and automatic scoring system for CDT has been increased. An optoelectronic protocol was suggested to qualitatively induce a parameter related to the movement kinematics in the CDT execution by acquiring the graphic gesture from video recording using six cameras, where a trial duration index was evaluated as a temporal parameter classifying between groups of subjects with Parkinson’s Disease and with both Parkinson’s Disease and dementia [29]. However, the optoelectronic protocol needs a lot of equipment and does not show parametric analysis of the specific behavior patterns during CDT. Due to sensors and sensing systems of modern mobile phones, accurate and rapid measurements of the user’s behavior patterns are available and can be easily implemented without using a lot of hardware equipment. A comparative study was presented for an automatic digit and hand sketch recognition, where several conventional machine learning classification algorithms were considered such as a decision tree, k-nearest neighbors, and multilayer perceptron [30]. Another machine learning based classification was developed for the CDT using a digital pen, where the scoring system used was the original Rouleau scoring system, and a comparative study was executed on the digital pen stroke analysis [31]. A deep learning approach was recently applied for automatic dementia screening and scoring on CDT, where the screening was for distinguishing between sick and healthy groups without scaling on the levels of dementia, and the scoring was for a performed dementia test not for an assessment of CDT [8]. In other words, the deep learning approach scored the CDT image in six levels according to the severity and status of dementia by using the image classification neural networks, such as VGG16 [32], ResNet-152 [33], and DenseNet-121 [34]. Therefore, a qualitative and automatic scoring system for CDT in mobile environments is still strongly demanded.
For this study, a mobile-phone app version for CDT, namely mCDT, was developed that includes a novel, original, automatic, and qualitative scoring system. The scoring methodology employs U-Net, a convolutional network for biomedical image segmentation, CNN, a convolutional network for digit classification, and MNIST database, the Modified National Institute of Standards and Technology database. Smart phone mobile sensor data were used to develop mCDT and its scoring system. Hand drawn clock images numbered 159 and were of 128 × 128 resolution; they were obtained via mCDT to train the U-Net to generate the trained model, DeepC, tasked with segmenting the contour of a drawn clock image. Similarly, the U-Net was trained with the same 159 CDT 128 × 128 resolution images to obtain DeepH, also a trained model, designed for segmenting the hands of a drawn clock image. The MNIST database was used to train CNN to obtain the trained model, DeepN, employed for classifying the digit images from a hand drawn clock image. The accuracies of greater than 75% and those being saturated were obtained with the epochs iterated for DeepC and DeepH. Similarly, accuracies of greater than 98% that were saturated were achieved with epochs iterated for DeepN. The mobile sensor data were the x and y coordinates, timestamps, and touch events for all the samples with a 20 ms sampling period extracted from the mobile touch sensor. Four parameters including contour (0–3 points), numbers (0–4 points), hands (0–5 points), and the center (0–1 points) were estimated by using DeepC, DeepH and DeepN along with the sensor data, resulting in scaling with a total of 13 points.
As a result, this study comes up with a significantly effective and accurate CDT scoring system available in a mobile device, which not only achieves the state-of-the-art accuracy on the assessment and the scaling of each CDT parametric criterion, but is also applicable to neurological disease diagnosis as well as temporal difference assessment of cognitive functioning in a daily lifelog. Especially, it is noticeable from the performance results for this system to show relatively more excellent specificity and precision for the PD test group than the young volunteer group. These results suggest that our mCDT application and the scoring system are valuable in differentiating dementia disease subtypes and also useful for clinical practice and field studies.
The rest of this paper is organized as follows. Section 2 describes the number of subjects enrolled in this study and the information related to the obtained approval of the ethics committee for gathering data from the subjects and the protocols suggested in this study. Section 2 also describes the implementation of mCDT, a mobile-phone app of CDT, the training models DeepC, DeepH and DeepN generated in this study, and the CDT scoring methods using the training models and the sensor data collected from mCDT for each of the parameters, such as the contour, numbers, hands, and center. Section 3 describes the results of the performance test of mCDT on a case-by-case basis for each of the four parameters. Section 4 presents a discussion of the results, the limitations, and the future works. Finally, Section 5 summarizes the overall results of this study.

2. Materials and Methods

2.1. Subjects

Right-handed young volunteers (238 total, 147 males and 89 females, aged 23.98 ± 2.83 years) were enrolled and took part in developing mCDT, the deep learning based mobile application in this study. Parkinson’s disease (PD) patients visiting the university hospital were also part of the study, submitting clock drawing images (140 total, 76 males and 64 females, aged 75.09 ± 8.57 years). The above-mentioned 238 volunteers were used to create the pre-training models of DeepC and DeepH. Image data from 159 volunteers (112 males and 45 females, aged 24.78 ± 1.63 years) and their clock drawing were used for the U-Net algorithm. The remaining 79 volunteers (35 males and 44 females, aged 22.81 ± 0.79 years) and those from the 140 PD patients with their clock drawing image data were subjected to testing the application’s scoring method. Two neurologists assisted in scoring of the CDT as the experts in CDT scaling, and cross checked all the images for an objective performance analysis, as well as gathering the data of the CDT test from the 140 PD patients. The Institutional Review Board of the Hallym University Sacred Heart Hospital, as an independent ethics committee (IEC), ethical review board (ERB), or research ethics board (REB), approved the gathering data and the protocols used for this study (IRB number: 2019-03-001). Table 1 provides the age, gender, and binary CDT score summary of the 238 volunteers and 140 PD patients.

2.2. Implementation of the Deep Learning Based Mobile Clock Drawing Test, mCDT

The Android Studio development environment was used to develop the deep learning based mobile application mCDT for the clock drawing test. A user of mCDT draws the face of a clock with all the numbers present and sets the hands to a specific time such as 10 after 11. Here, the clock face contour could be pre-drawn by mCDT as an option chosen by the user and the specific time is randomly selected by mCDT. Then, mCDT scores the drawn image qualitatively; this scoring is based on mobile sensor data of the drawing image and pre-trained models, DeepC, DeepH and DeepN created in this study. Fast and precise segmenting of the clock face contour and the hands in the images were accomplished by DeepC and DeepH, respectively, using U-Net, a convolutional network architecture. In turn, DeepN classifies the numbers using CNN and the MNIST database. The mobile sensor data of x and y coordinates in pixels, timestamps in seconds, and touch events for each samples of the drawing image are made with a 50 Hz sampling frequency. Three types of the touch events, ‘up’, ‘down’, and ‘move’ were considered in mCDT. The touch event ‘down’ occurs when the user starts to touch on the screen; ‘up’ when the user ends it; and ‘move’ when the user moves the finger or the pen across the screen. Figure 1a provides the flow chart of the processes by mCDT. Figure 1b–d provide the screen shots of the registration window, the CDT window, and the result window of mCDT, respectively. As shown in Figure 1a,b, an informed consent prompt appears at the launch of mCDT, followed by a registration window for entering the subject’s information; these include age, name, gender, education level and handedness of the subject plus optional parameters including an email address. After pressing the start button in the registration window, the CDT window appears as shown in Figure 1c, and the user is instructed to draw numbers and hands on a clock face contour; the contour of the clock face is asked to be drawn by the user or pre-drawn by mCDT as an option chosen by the user. In the drawing, the clock hands have to set to a specific time given randomly by mCDT. The sensor data are saved as the subject draws a contour, numbers and hands of a clock face on the touch screen of the CDT window. The sensor data along with the drawn image are then provided in the results window as shown in Figure 1d. The results could then be forwarded to the email address submitted at the registration window.

2.3. Pre-Trained Models, DeepC, DeepH and DeepN Based on the U-Net and the CNN

Novel pre-trained models DeepC, DeepH and DeepN were developed for the segmentation and classification of the contour, the hands, and the numbers, respectively, of the clock face from a drawn clock image. DeepC and DeepH were created based on the U-Net convolutional network architecture and DeepN, based on the CNN in keras [6]. The U-Net and CNN network architecture implemented in this study are illustrated in Figure 2 and Figure 3, respectively. The U-Net network architecture consists of a contracting path, an expansive path, and a final layer as shown in Figure 2. The contracting path consists of repeated applications of two 3 × 3 convolutions and a 2 × 2 max pooling operation with stride 2 for down-sampling. At each repetition, the number of feature channels is doubled. The expansive path consists of two 3 × 3 convolutions and a 2 × 2 convolution (“up-convolution”) for up-sampling to recover the size of the segmentation map. At the final layer, a 1 × 1 convolution was used to map each 16-component feature vector to the desired number of classes. In total, the network has 23 convolutional layers. The training data for both DeepC and DeepH contain 477 images of 128 × 128 resolution, which were augmented using a module called ImageDataGenerator in keras.preprocessing.image and resized from the original 159 images of 2400 × 1200 resolution. The augmentation was carried out by randomly translating horizontally or vertically using the parameter value 0.2 for both width_shifting_range and height_shifting_range. DeepC and DeepH were both trained for 100 epochs with an accuracy of about 77.47% and 79.56%, respectively. The loss function used for the training was basically a binary cross entropy. The CNN network architecture consists of two convolution layers (C1 and C3), two pooling layers (D2 and D4), and two fully connected layers (F5 and F6), as shown in Figure 3. The first convolution layer C1 filters the 28 × 28 input number image with 32 kernels of size 5 × 5, while the second convolution layer C3 filters the down-sampled 12 × 12 × 32 feature maps with 64 kernels of size 5 × 5 × 32. A unit stride is used in both the convolution layers, and a ReLU nonlinear function is used at the output of each of them. Down-sampling occurs at layer D2 and D4 by applying 2 × 2 non-overlapping max pooling. Finally, the two fully-connected layers, F5 and F6, have 1024 and 10 neurons, respectively. The MNIST handwritten digit database (about 60,000 images) and the digit images from the 477 CDT images were used to train the CNN architecture used here to obtain the trained model, DeepN.

2.4. Scoring Method of mCDT

The novel, automatic and qualitative scoring method of mCDT was developed, based on the sensor data and the pre-trained models, DeepC, DeepH and DeepN. Four parameters were included in the scoring method: contour (0–3 points), numbers (0–4 points), hands (0–5 points), and the center (0–1 point). Some of the scoring criteria were adopted from a previous study by Paolo Caffarra et al. [22]. A total score corresponding to the sum of individual scores of each parameter ranged from 0 to 13. When a subject executes the CDT more than once, the best copy is then scored. A detailed list of the scoring criteria of each parameter used in this study is presented in Table 2 and the overall flowchart of the scoring method is shown in Figure 4.
Two forms of data that include the sensor data and the clock drawing image are generated for output once mCDT has been completed, as shown in Figure 4a. The clock drawing image, I C , is intended to be of a clock face with numbers, hands and a contour. From the original 2400 × 1200 pixel drawing image at the CDT window, the clock drawing image I C is resized to 128 × 128 pixels. Time stamps t [ n ] in sec, x- and y- coordinates, x [ n ] and y [ n ] in pixels, and touch-events e [ n ] of the sensor data for the 128 × 128 drawing image have a sampling rate of 50   Hz with n being the index of a sample point. Each touch-event e [ n ] has a value such as −1, 0, 1; the assigned value of −1 designates the event ‘down’ for the screen being touched, 1 is for the event ‘up’ with the screen touch not continuing, and 0 is for the event ‘move’ with moving and touching on the screen continuing.
The sensor data x [ n ] and y [ n ] , c i n c f belonging to the contour in the clock drawing image I C are obtained using the touch-events e [ n ] , c i n c f , where c i and c f are the start and the end indices of the contour, respectively, which can be estimated by the touch-event down-shifting from the event ‘down’ into the event ‘move’ and the touch-event up-shifting from the event ‘move’ into the event ‘up’, respectively. Besides, the touch-events e [ n ] , c i < n < c f between the touch-event down and up shiftings have to be successively stayed in the event ‘move’ for the longest time if such a period occurs more than once. In other words, the longest continuous sequence of 0 s in the touch-event e [ n ] , starting with the digit −1 and ending with the digit 1, identifies itself as belonging to the contour in the clock drawing image I C .
The contour image I f c is segmented from the clock drawing image I C using DeepC, the pre-trained models. Next, percentages p f c of the segmented image I f c , matching to the corresponding portion of the clock drawing image I C , is estimated by Equation (1), where n ( I C , c i n c f I f c ) is the number of pixel coordinates that I f c and the contour image I C , c i n c f have in common; and n ( I C , c i n c f ) is the total number of pixel coordinates in the sensor data belonging to the contour image.
p f c = n ( I C , c i n c f I f c ) n ( I C , c i n c f )
A modified clock drawing image I C is generated as shown in Figure 4b by redrawing the remaining part of the clock drawing image I C after excluding the sensor data x [ n ] , y [ n ] , t [ n ] , c i n c f belonging to the contour, and then by binarizing for the background to be black and for the number digits to be drawn white.
The hand image I f h is separately segmented from the clock drawing image I C using DeepH, the pre-trained models. Next, percentages p f h k , k = 1 , 2 of the segmented images I f h , matching to the corresponding portion of the clock drawing image I C are estimated by Equation (2), where n ( I C , h i k n h f k I f h ) is the number of pixel coordinates that I f h and one of the hands images I C , h i k n h f k , k = 1 , 2 have in common; and n ( I C , h i k n h f k ) , k = 1 , 2 is the total number of pixel coordinates in the sensor data belonging to one of the hands images.
p f h k = n ( I C , h i k n h f k I f h ) n ( I C , h i k n h f k )
The sensor data x [ n ] and y [ n ] , h i k n h f k , k = 1 , 2 belonging to one of the hand and minute hands images in the modified clock drawing image I C , are obtained using the touch events e [ n ] , h i k n h f k , k = 1 , 2 , where h i k and h f k are the start and the end indices of the hand, respectively, which can be estimated by a touch-event down shifting from the event ‘down’ into the event ‘move’ and by the touch-event up shifting from the event ‘move’ into the event ‘up’, respectively. Here, the time stamps between t [ h i k ] and t [ h f k ] have to be overlapped with those between t i [ j ] and t f [ j ] , j = 1 , 2 , , N . If there is only one or no hands, so such a touch-event down or up shifting is not identified in the modified clock drawing image I C , then the corresponding time stamps t [ h i k ] and t [ h f k ] are treated to have missing values, NAs.
The minimum and maximum x- and y- coordinates in pixels, x min c and x max c , and y min c and y max c , of the sensor data x [ n ] and y [ n ] , c i n c f are estimated as the boundary of the contour and formulated by Equations (3)–(6), respectively.
x min c = min c i n c f x [ n ]
x max c = max c i n c f x [ n ]
y min c = min c i n c f y [ n ]
y max c = max c i n c f y [ n ]
The x- and y- coordinates in pixels, x m i d c and y m i d c of the center point of the contour are defined by the bisecting point of the minimum and maximum x- and y- coordinates in pixels, x min c and x max c , and y min c and y max c , that are formulated by Equations (7) and (8), respectively:
x m i d c = ( x min c + x max c ) / 2
y m i d c = ( y min c + y max c ) / 2
Pre-estimated point positions P k = ( x d [ k ] , y d [ k ] ) , k = 1 , 2 , , 12 of the clock number digits from 0 to 12 are evaluated as shown in Figure 4c by using the boundaries x min c , x max c , y min c and y max c of the contour along with the center point x m i d c and y m i d c , where x d [ k ] and y d [ k ] are x- and y- coordinates in pixels of the k t h digit number from 0 to 12. Table 3 summarizes the corresponding formula of each of the pre-estimated positions P k , k = 1 , 2 , , 12 .
Next, each of the number images I C j , j = 1 , 2 , , N corresponding to a digit number is cropped out from the modified clock drawing image I C using the function findContours() of OpenCV2, where N is the total number of the digit images cropped out and j is the index sorted by the time stamps in ascending order. Here, the function findContours() can be used for finding the suburb contours of white objects from a black background [35].
The model DeepN classifies each of the number images I C j , j = 1 , 2 , , N into one of 10 integer values ranging from 0 to 9, inclusive, and saves the identified integer in D [ j ] . At the same time, spatial data, L u x [ j ] L u y [ j ] , L d x [ j ] , and L d y [ j ] , j = 1 , 2 , , N , of the suburb contours and the corresponding time stamps t i [ j ] and t f [ j ] , j = 1 , 2 , , N are generated, where L u x [ j ] and L u y [ j ] are the upper left x- and y- coordinates in pixel of the j t h suburb contour, respectively; L d x [ j ] and L d y [ j ] are the lower right x- and y- coordinates in pixels of the j t h suburb contour, respectively; t i [ j ] t [ n ] and t f [ j ] t [ n ] are the corresponding initial and final time stamps of the sensor data belonging to the number image in the j t h suburb contour, respectively, and the index j is sorted by the time stamp t i [ j ] .

2.4.1. Scoring on Criteria of Contour Parameter

Scoring on the circular shape, the closure (opening), and the size properness of the contour are evaluated by the percentage p f c of the segmented image I f c matching to the corresponding portion of the clock drawing image I C , the maximum contour closure distance d max c , and the ratio A c / W c of the contour A c to the CDT window sizes W c as shown in Figure 4. The circular shape is identified if the value of the percentage p f c is larger than a given threshold θ c 1 . Scoring on the closure(opening) of the contour is evaluated by the first p i c = ( x [ c i ] , y [ c i ] ) and the last p f c = ( x [ c f ] , y [ c f ] ) sample points along with the contour sample points p s h k = ( x [ c s h k ] , y [ c s h k ] ) at down or up shiftings in the touch events, that are shifting from ‘up’ to ‘move’ or from ‘move’ to ‘up’ at the time stamp t [ c s h k ] , c i < c s h k < c f . The closure of the contour is identified if it is greater than a given threshold θ c 2 , the maximum value d max c of ( p i c , p f c ) , the distance between the first and the last contour sample points, and ( p s h k , p s h k + 1 ) the distances between the k t h and the ( k + 1 ) t h sample points shifting down or up in the touch events, where the index k is sorted by the time stamp of the sample points. The appropriateness of the contour size is evaluated by the ratio of the size of the contour to that of the CDT window. The contour size A c in pixels is calculated by the expression A c = ( x max c x min c ) ( y max c y min c ) using x min c , x max c , y min c and y max c . The appropriate of the contour size is identified if the ratio A c / W c is larger than a given threshold θ c 3 , where W c is the size in pixels of the CDT window.

2.4.2. Scoring on Criteria of Numbers Parameter

Figure 5 shows the flowcharts suggested in this study for scoring presence of all the numbers and no additional numbers, correctness of the order of the numbers, correctness of the positions of the numbers, and positioning of the numbers within the contour.
The presence of all the numbers and no additional numbers is evaluated by using the classified output D [ j ] , j = 1 , 2 , , N of DeepN for the cropped number images I C j , j = 1 , 2 , , N and identified if the total number N is equal to 15, all the values in D [ j ] , j = 1 , 2 , , N are in the range from 0 to 9, and the count number of digits 1 and 2 are 5 and 2, respectively.
The correctness of the order of the numbers is evaluated by using the classified outputs D [ j ] , j = 1 , 2 , , N of DeepN for the cropped number images I C j , j = 1 , 2 , , N . First of all, the number sequence S N = [ D [ 1 ] D [ 2 ] D [ N ] ] is obtained and compared to reference number sequences considering general human habits. Here, three types of ordering in drawing numbers were considered as reference number sequences, such as drawing numbers starting from digits 1 through 12 in ascending order, starting from digit 12, and then digits 1 through 11 in ascending order, or starting from digits 12, 3, 6, and 9 and then inserting digits 1 through 11 in ascending order. Therefore, the reference number sequences considered were S 1 = [ 1 , 2 , , 9 , 1 , 0 , 1 , 1 , 1 , 2 ] , S 2 = [ 1 , 2 , 1 , 2 , , 9 , 1 , 0 , 1 , 1 ] , and S 3 = [ 1 , 2 , 3 , 6 , 9 , 1 , 2 , 4 , 5 , 7 , 8 , 1 , 0 , 1 , 1 ] . The correctness of the order of the numbers is identified if the maximum value R s e q = max i = 1 , 2 , 3 S e q u e n c e M a t c h e r ( S N , S i ) of percentages of matched sequences between the number sequence S N and each of the reference number sequences S 1 , S 2 and S 3 is greater than a given threshold θ n , where the function SequenceMatcher() is in the python library difflib [36].
The correctness of the position of the numbers is evaluated by the classified outputs D [ j ] , L u x [ j ] L u y [ j ] , L d x [ j ] , and L d y [ j ] , t i [ j ] and t f [ j ] , j = 1 , 2 , , N of DeepN for the cropped number images I C j , j = 1 , 2 , , N as well as the predefined positions P ( x d [ k ] , y d [ k ] ) , k = 1 , 2 , , 12 of the number digits from 0 to 12. The center position P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , N of each of the cropped number images I C j , j = 1 , 2 , , N is estimated by the bisecting point of the upper left point P ( L u x [ j ] , L u y [ j ] ) and the lower right point P ( L d x [ j ] , L d y [ j ] ) , j = 1 , 2 , , N , estimated by Equations (9) and (10).
x m [ j ] = ( L u x [ j ] + L d x [ j ] ) / 2 ,   j = 1 , 2 , , N
y m [ j ] = ( L u y [ j ] + L d y [ j ] ) / 2 ,   j = 1 , 2 , , N
The digit number k in D [ j ] is identified and the distance d c n [ j ] is estimated between the center position P ( x m [ j ] , y m [ j ] ) of the j t h cropped number images I C j and the predefined position P ( x d [ k ] , y d [ k ] ) corresponding to the identified digit number k in D [ j ] . Then, the correctness of the position of the numbers is identified if the percentage of the distances d c n [ j ] , j = 1 , 2 , , N ; within a given limit, d c is greater than a given value θ d c .
The positioning of the numbers within the contour is evaluated by using the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , N of each of the cropped number images I C j , j = 1 , 2 , , N and the contour sensor data, x [ n ] and y [ n ] , c i n c f . A circle F c L is fitted to the contour sensor data, x [ n ] and y [ n ] , c i n c f using the least squares circle fitting algorithm [37], where the center point P ( x [ n c ] , y [ n c ] ) , n c { n | c i n c f } and the radius R c L of the fitted circle F c L are obtained. Similarly, a circle F c N is fitted to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , N using also the least squares circle fitting algorithm, where the center point P ( x [ n c N ] , y [ n c N ] ) and the radius R c N of the fitted circle F c N are obtained. Then, the positioning of the numbers is identified if the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L .

2.4.3. Scoring on Criteria of Hands Parameter

Figure 6 shows the flowchart suggested in this study for scoring the presence of two or one hand, correctness of the proportion of the hands, correctness of the hour target number, and correctness of the minute target number.
Presence of two or one hand is evaluated by the percentage p f h of the segmented image I f h matching to the corresponding portion of the clock drawing image I C . The presence of two hands is identified if the value of the percentage p f h is larger than a given threshold θ h 1 . The presence of one hand is identified if the value of the percentage p f h is larger than a given threshold θ h 2 . Here, the value of the given threshold θ h 2 is smaller than that of the given threshold θ h 1 since DeepH is trained with images of the clock face with two hands so that the criteria of the two hands is included in the criteria of the one hand.
Indication of the correct proportion of hands is evaluated by using the hands sensor data x [ n ] , y [ n ] , t [ n ] , and e [ n ] h i n h f between the time stamps t [ h i ] and t [ h f ] . Here, indication of the presence of two hands is a prerequisite of the indication of the correct proportion of hands. The hands sensor data x [ n ] , y [ n ] , t [ n ] , and e [ n ] h i n h f is divided into two sets H 1 = { x [ n ] , y [ n ] , t [ n ] , e [ n ] | h i n < h m } and H 2 = { x [ n ] , y [ n ] , t [ n ] , e [ n ] | h m < n h f } using a line clustering algorithm [38]. Here, the time stamp t [ h m ] h i < h m < h f is the time point of the intermission between the two sets H 1 and H 2 . Then, the length h 1 of one hand is estimated as the maximum distance between P ( x [ n ] , y min 1 ) , h i n < h m and P ( x [ n ] , y max 1 ) , h i n < h m using the hand sensor data in the set H 1 , where y min 1 and y max 1 are the minimum and maximum of y coordinates in the set H 1 . Similarly, the length h 2 of another hand is estimated as the maximum distance between P ( x [ n ] , y min 2 ) , h m < n h f and P ( x [ n ] , y max 2 ) , h m < n h f using the hand sensor data in the set H 2 , where y min 2 and y max 2 are the minimum and maximum of y coordinates in the set H 2 . Finally, the indication of the correct proportion of hands is identified if the length difference Δ h = | h 2 h 1 | between the lengths of one and another hand is larger than a given number θ p r .
Indication of the hour target number is evaluated by using the hands sensor data x [ n ] , y [ n ] , t [ n ] , and e [ n ] h i n h f between the time stamps t [ h i ] and t [ h f ] . Two different cases are considered here; one is that the presence of two hands is identified, and the other is that the presence of only one hand is identified. For the first case, the hour hand sensor data S h with larger data size of the two sets H 1 and H 2 is fitted into a line and then the fitted line is extrapolated within a range [ y h , min , y h , max ] of y pixel coordinates, where y h , min is the minimum value of y coordinates y [ n ] , h i n < h f in the hands sensor data and y h , max , the maximum of y coordinates y [ n ] , c i n < c f in the contour sensor data. For the second case, the whole hands sensor data is fitted into a line and then the fitted line is extrapolated within a range [ y h , min , y h , max ] of y pixel coordinates. Next, the closest point P h ( x [ n h ] , y [ n h ] ) , c i n < c f of the contour sensor data to the extrapolated line is evaluated. Finally, the indication of the hour target number is identified if the point P h ( x [ n h ] , y [ n h ] ) is within a given range from a predefined pixel point P ( x d [ h t ] , y d [ h t ] ) of the given hour target digit h t .
Indication of minute target number is similarly evaluated by using the hands sensor data x [ n ] , y [ n ] , t [ n ] , and e [ n ] h i n h f between the time stamps t [ h i ] and t [ h f ] . Two different cases are considered here; one is that the presence of two hands is identified, and the other is that the presence of only one hand is identified. For the first case, the minute hand sensor data S m with smaller data size of the two sets S 1 and S 2 is fitted into a line and then the fitted line is extrapolated within a range [ y h , min , y h , max ] of y pixel coordinates, where y h , min is the minimum value of y coordinates y [ n ] , h i n < h f in the hands sensor data, and y h , max is the maximum of y coordinates y [ n ] , c i n < c f in the contour sensor data. For the second case, the whole hands sensor data is fitted into a line and then the fitted line is extrapolated within a range [ y h , min , y h , max ] of y pixel coordinates. Next, the closest point P ( x [ n m ] , y [ n m ] ) , c i n < c f of the contour sensor data to the extrapolated line is evaluated. Finally, the indication of a minute target number is identified if the point P ( x [ n m ] , y [ n m ] ) is within a given range from a predefined pixel point P ( x d [ m t ] , y d [ m t ] ) of the given minute target digit m t . Here, the point P ( x [ n m ] , y [ n m ] ) has to be the same as the point P ( x [ n h ] , y [ n h ] ) if the indication of the hour target number is already identified.

2.4.4. Scoring Criteria of Center Parameter

Presence or inference of the center point of the clock face in the drawing image I C is identified, as shown in Figure 4 if the presence of two or one hand is identified. Also, presence or inference of the center is identified if there is a data point within a given range from the center point P ( x m i d c , y m i d c ) .

2.4.5. Assignment of Scores

Table 4 lists the conditions for assigning scores for each parameter in mCDT, where the heuristic values of all the thresholds used in this study were summarized at the footnote.
The score of the contour parameter is via the percentage p f c , the maximum contour closure distance d max c , and the ratio of the contour to the CDT window sizes A c / W c . The score of the circular contour is a 1 if the percentage p f c is greater than a given threshold θ c 1 ; the score of the closure contour is a 1 if the maximum contour closure distance d max c is greater than a given threshold θ c 2 ; the score of the appropriate sized contour is a 1 if the ratio of the contour to the CDT window sizes A c / W c is greater than a given threshold θ c 3 .
The score of the numbers parameter is determined by the contour sensor data, x [ n ] and y [ n ] , c i n c f , the outputs, D [ j ] , L u x [ j ] L u y [ j ] , L d x [ j ] , L d y [ j ] , t i [ j ] and t f [ j ] , j = 1 , 2 , , N of DeepN, the reference number sequences S i , i = 1 , 2 , 3 , and the predefined number positions P ( x d [ k ] , y d [ k ] ) , k = 1 , 2 , , 12 of the number digits from 0 to 12. There are four criteria on the numbers parameter, such as the presence of all the numbers and no additional numbers, the correctness of the order of the numbers, the correctness of the position of the numbers, and the positioning of the numbers within the contour. The score of the presence of all the numbers and no additional numbers is a 1 if the total number N is equal to 15, all the values in D [ j ] , j = 1 , 2 , , N are in the range from 0 to 9, and the count number of digits 1 and 2 are 5 and 2, respectively. The score of the correctness of the order of the numbers is a 1 if the maximum value R s e q of percentages of matched sequence between the number sequence S N and each of the reference number sequences S 1 , S 2 and S 3 is greater than a given threshold θ n . The score of the correctness of the position of the numbers is a 1 if the percentage of the distances d c n [ j ] , j = 1 , 2 , , N within a given limit d c is greater than a given value θ d c , where the distance d c n [ j ] is estimated between the j t h center position P ( ( L u x [ j ] + L d x [ j ] ) / 2 , ( L u y [ j ] + L d y [ j ] ) / 2 ) and the predefined position P ( x d [ k ] , y d [ k ] ) corresponding to the identified digit number k in D [ j ] . The score of the positioning of the numbers within the contour is a 1 if the radius R c N is smaller than the radius R c L , where the radius R c N is of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 15 and the radius R c L is of the fitted circle F c L to the contour sensor data, x [ n ] and y [ n ] , c i n c f . The score of criteria on the numbers parameter is the sum of the scores of the presence of all the numbers and no additional numbers, the correctness of the order of the numbers, the correctness of the position of the numbers, and the positioning of the numbers within the contour.
The score of the hands parameter is determined by the percentage p f h , the linearly clustered sets, H 1 and H 2 of the hands sensor data x [ n ] and y [ n ] , h i n h f , the contour sensor data x [ n ] and y [ n ] , c i n c f , the point P ( x [ n h ] , y [ n h ] ) in the range [ y min i , y max h ] , i = 1 , 2 where y min i is the minimum of y coordinates y [ n ] in H i , i = 1 , 2 and y h , max is the maximum of y coordinates y [ n ] , c i n < c f . There are four criteria on hands parameter, such as the presence of two or one hand, the indication of the correct proportion of hands, the indication of the hour target number, the indication of the minute target number, and the positioning of the numbers within the contour. The score of the presence of two hands or one hand is a 2 if the value of the percentage p f h is larger than a given threshold, θ h 1 ; a 1 if the value of the percentage p f h is larger than a given threshold θ h 2 . The score of the indication of the correct proportion of hands is a 1 if the value of the percentage p f h is larger than a given threshold θ h 1 and the size difference Δ h of the fitted and extrapolated lines in data between H 1 and H 2 is larger than a given number θ p r , where H 1 and H 2 are the two sets divided by a line cluster algorithm on the hands sensor data, x [ n ] and y [ n ] , h i n h f . The score of the indication of the hour target number is a 1 if the point P ( x [ n h ] , y [ n h ] ) is within a given range of a predefined pixel point of the given hour target, where the point P ( x [ n h ] , y [ n h ] ) is obtained by estimating the closest one of the extrapolated hands sensor data spatially to the contour sensor data within the range [ y h , min , y h , max ] . The score of the indication of the minute target number is a 1 if the point P ( x [ n m ] , y [ n m ] ) is within a given range of predefined pixel points of the given minute target, where the point P ( x [ n m ] , y [ n m ] ) is obtained by estimating the closest of the extrapolated hands sensor data spatially to the contour sensor data within the range [ y h , min , y h , max ] . Here, the point P ( x [ n m ] , y [ n m ] ) has to be the same as the point P ( x [ n h ] , y [ n h ] ) if the indication of the hour target number is already identified. The score of the criteria on the hands parameter is the sum of the scores of the presence of two hands or one hand, the indication of the correct proportion of hands, the indication of the hour target number, the indication of the minute target number, and the positioning of the numbers within the contour.
The score of the criteria of the center parameter is determined by the center point P ( x m i d c , y m i d c ) of the contour sensor data, x [ n ] and y [ n ] , c i n c f . The score of the presence or the inference of the center of the clock face is a 1 if the presence of two or one hands is identified, or if there is a data point within a given range from the center point P ( x m i d c , y m i d c ) .

3. Results

3.1. Scoring on Criteria of Contour Parameter

Figure 7 depicts separate examples of original drawings (first column) each with the segmented image (second column) perceived by DeepC for a detected contour, the overlapping image (third column) of the original and the segmented images and the corresponding parameter values including the total score estimated.
In Figure 7a where the original image is of a closed circular contour sized appropriately, the segmented image has the estimated percentage p f c of 95.74%, the maximum contour closure distance has the evaluated pixel value d max c of 9.67 pixels, and the ratio of the contour to the CDT window size has the estimated value A c / W c of 0.339. Both the closure and circular of the contour were evaluated to be one, as p f c was greater than the 75.00 score for θ c 1 , a threshold heuristically set, and d max c was less than the 50.00 score for θ c 2 , a threshold heuristically set. The size of the contour was also evaluated to be one, as A c / W c was greater than the 0.1 score for θ c 3 , a threshold heuristically set. Therefore, the total score of the contour parameter was evaluated to be three.
Figure 7b has an original drawing image of a circular contour sized appropriately, but not wholly closed. The segmented image has the estimated percentage p f c of 89.66%, the maximum contour closure distance has the evaluated pixel value d max c of 55.56 pixels, and the ratio of the contour to the CDT window sizes has the estimated value A c / W c of 0.337. The closure of the contour was evaluated to be zero, as d max c was greater than the 50.00 score for θ c 2 ; however, both the circular and the size were gauged to be one, as the estimated percentage p f c was greater than the 75.00 score for θ c 1 and A c / W c was greater than the 0.1 score for θ c 3 . Therefore, the total score of the contour parameter was evaluated to be two.
Figure 7c shows an example of an original drawing of an appropriately sized, but neither closed nor circular, contour. The segmented image has the estimated percentage p f c of 52.31%, the maximum contour closure distance has the evaluated pixel value d max c of 51.56 pixels, and the ratio of the contour to the CDT window sizes has the estimated value A c / W c of 0.237. Both the closure and circular of the contour were evaluated to be zero, as p f c was not greater than the 75.00 score for θ c 1 and d max c was greater than the 50.00 score for θ c 2 ; however, the size was gauged to be one, as A c / W c was greater than the 0.1 score for θ c 3 . Therefore, the total score of the contour parameter was evaluated to be one.
Finally, the original drawing image of Figure 7d depicts an example of a closed circular contour, but not sized appropriately. The segmented image has the estimated percentage p f c of 97.44%, the maximum contour closure distance has the evaluated pixel value d max c of 32.01 pixels, and the ratio of the contour to the CDT window sizes has the estimated value A c / W c of 0.061. Both the closure and circular of the contour were evaluated to be one, as p f c was greater than the 75.00 score for θ c 1 and d max c was less than the 50.00 score for θ c 2 ; however, the size was gauged to be zero, as A c / W c was less than the 0.1 score for θ c 3 . The total score of the contour parameter was evaluated to be.

3.2. Scoring on Criteria of Number Parameter

In eight representative examples, Figure 8 demonstrates how numbers in original images are detected and classified in their orders and positions, showing the binarized image (left), the original image with the cropped areas for the numbers (middle), and the corresponding parameter values (right) of each example.
Figure 8a displays the case in which all the numbers without any additional numbers are present in the correct orders and the proper positions within the contour. The classified output D [ j ] , j = 1 , 2 , , N of DeepN, the pretrained model, was identified to have no missing values, since the total umber N was equal to 15, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were five and two, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be one. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , was evaluated to be 100.00%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 100.00 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was greater than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] in pixels was specifically 46.0, 66.9, 27.5, 32.0, 33.0, 16.6, 119.0, 123.0, 38.0, 51.1, 40.7 and 22.6. Therefore, the correctness of the position of the numbers was evaluated to be one. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 454.7 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 351.6 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be four.
Figure 8b displays the case in which all the numbers without any additional numbers are present in the correct orders within the contour but not in proper positions. The classified output D [ j ] , j = 1 , 2 , , N of DeepN, the pretrained model, was identified to have no missing values, since the total number N was equal to 15, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were five and two, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be one. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , was evaluated to be 86.66%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 86.66 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] in pixels for each of the digit numbers k in D [ j ] was specifically 243.0, 367.8, 366.93, 594.4, 518.2, 398.0, 491.1, 365.9, 418.2, 404.0, 628.2 and 143.6. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 447.6 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 338.1 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be three.
Figure 8c displays the case in which all the numbers without any additional numbers are present in the correct orders, but not in the proper positions nor within the contour. The classified output D [ j ] , j = 1 , 2 , , N of DeepN, the pretrained model, was identified to have no missing values, since the total number N was equal to 15, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were five and two, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be one. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , was evaluated to be 93.33%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 93.33 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] was specifically 147.7, 170.7, 214.2, 364.0, 369.4, 139.1, 114.3, 226.4, 157.0, 242.5, 224.0 and 127.6. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 538.5 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 694.2 pixels. Now, the positioning of the numbers within the contour was evaluated to be zero, since the radius R c N of the fitted circle F c N is larger than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be two.
Figure 8d displays the case in which all the numbers without any additional numbers are present in the correct orders within the contour, but not in the proper positions. The classified output D [ j ] , j = 1 , 2 , , N of DeepN, the pretrained model, was identified to have no missing values, since the total number N was equal to 15, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were five and two, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be one. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , was evaluated to be 100.00%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 100.00 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] was specifically 618.19, 1106.19, 1408.42, 931.44, 378.00, 185.95, 630.61, 1079.10, 1381.07, 1314.39, 720.95 and 58.69. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 554.8 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 330.8 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be three.
Figure 8e displays the case in which some numbers are missing and the presented numbers are not in proper positions, but mostly in correct order within the contour. The classified output D [ j ] , j = 1 , 2 , , N of DeepN, the pretrained model, was identified to have some missing values, since the total number N was equal to 13, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were four and two, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be zero. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , was evaluated to be 85.71%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 85.71 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] was specifically 497.8, 462.5, 350.7, 399.3, 415.1, 254.5, 208.5, 1037.8, 836.2, 792.1, 743.6 and 952.1. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 319.8 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 193.2 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be two.
Figure 8f displays the case in which there are additional numbers not belonging to a clock but the numbers are in correct orders within the contour. The classified output D [ j ] , j = 1 , 2 , , N of DeepN the pretrained model was identified to have some additional numbers, since the total number N was equal to 42, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) was thirteen and nine, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be zero. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , were evaluated to be 83.57%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 83.57 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] was specifically 376.5, 483.4, 432.1, 636.2, 743.4, 856.4, 947.7, 1056.9, 1171.4, 1113.0, 826.2 and 837.7. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 694.2 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 446.1 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N was smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be two.
Figure 8g displays the case in which some numbers are missing and the presented numbers are not in proper positions, but mostly in correct order within the contour. The classified output D [ j ] , j = 1 , 2 , , N of DeepN, the pretrained model, was identified to have some additional numbers, since the total number N was equal to 11, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were two and one, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be zero. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences S i i = 1 , 2 , 3 , were evaluated to be 66.66%. Therefore, the correctness of the order of the numbers was evaluated to be one, since the value 66.66 of the maximum ratio R s e q was greater than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] was specifically 87.7, 68.9, 56.1, 78.0, 163.0, 190.1, 232.8, 265.3, 894.3, 860.6, 802.5 and 990.8. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 242.4 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 12 of each of the cropped number images I C j , j = 1 , 2 , , N was obtained to be 149.9 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be two.
Figure 8h displays the case in which many numbers are missing and the presented numbers are not in proper positions and correct order, but within the contour. The classified output D [ j ] , j = 1 , 2 , , N of DeepN the pretrained model was identified to have some missing numbers, since the total number N is equal to five, and the classified output D [ j ] , j = 1 , 2 , , N was in the range from zero to nine, that is 0 D [ j ] 9 for all j , and the numbers n ( D [ j ] = 1 ) and n ( D [ j ] = 2 ) were one and two, respectively. Therefore, the presence of all the numbers and no additional numbers was evaluated to be zero. The maximum ratio R s e q in percentage between the number sequence S N and the reference sequences, S i i = 1 , 2 , 3 , were evaluated to be 42.10%. Therefore, the correctness of the order of the numbers was evaluated to be zero since the value 42.10 of the maximum ratio R s e q was less than 65.00, the heuristically given value of θ n . The ratio n( d c n [ j ] d c )/n( d c n [ j ] ) of the distances d c n [ j ] , j = 1 , 2 , , 12 within 100.00 pixels, a given limit d c , was less than 0.65, a heuristically given value θ d c , where the evaluated distance d c n [ j ] for each of the digit numbers k in D [ j ] was specifically 545.5, 920.4, 1167.1, 1739.7, 1644.2, 1464.5, 1338.6, 1095.1, 715.6, 630.7, 734.1 and 153.0. Therefore, the correctness of the position of the numbers was evaluated to be zero. The radius R c L of the fitted circle F c L to the contour sensor data was obtained to be 601.1 pixels. The radius R c N of the fitted circle F c N to the center point P ( x m [ j ] , y m [ j ] ) , j = 1 , 2 , , 15 of each of the cropped number images I C j , j = 1 , 2 , , 15 was obtained to be 446.3 pixels. Now, the positioning of the numbers within the contour was evaluated to be one, since the radius R c N of the fitted circle F c N is smaller than the radius R c L of the fitted circle F c L . Finally, the total score of the numbers parameter was evaluated to be one.

3.3. Scoring on Criteria of Hand Parameter

The analytical ability of the pre-trained model DeepH is demonstrated in Figure 9, with eight separate examples on how two hands in the original images are evaluated in the presence, the correctness of the proportions, and the correctness of the target numbers, showing the segmented image of hands (left), the original image with the cropped areas of the target numbers, and the extrapolated lines of hands (middle), and the corresponding parameter values (right) of each examples.
Figure 9a displays the case in which two hands are present with the proper proportions and target numbers. In this case, the percentage p f h was evaluated to be 100.00% greater than 65% the score for θ h 1 . Therefore, the presence of two hands was scored to be two. The length difference Δ h was evaluated to be 89.4 pixels greater than 30.0 pixels the score for θ p r . Therefore, the correct proportion of the two hands was also scored to be one. Both of the distances a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) and a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) were estimated to be 123.9 and 86.6 pixels less than 200.0 pixels, the score for ε , respectively. Therefore, both the correctness of hand and minute target numbers were evaluated to be one. Finally, the total score of the hands parameter was evaluated to be five.
Figure 9b displays another example of the case in which two hands are present with the proper proportions and target numbers, where one of the target numbers is not in the proper position. In this case, the percentage p f h was evaluated to be 73.37% greater than 65% the score for θ h 1 . Therefore, the presence of two hands was scored to be two. The length difference Δ h was evaluated to be 219.3 pixels greater than 30.0 pixels, the score for θ p r . Therefore, the correct proportion of the two hands was also scored to be one. Both of the distances a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) and a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) were estimated to be 57.7 and 44.5 pixels less than 200.0 pixels, the score for ε , respectively. Therefore, both the correctness of hand and minute target numbers were evaluated to be one. Finally, the total score of the hands parameter was evaluated to be five.
Figure 9c displays the case in which two hands are present with the proper proportions, but one of them is not indicating the target number. In this case, the percentage p f h was evaluated to be 65.35% greater than 65%, the score for θ h 1 . Therefore, the presence of two hands was scored to be two. The length difference Δ h was evaluated to be 101.7 pixels less than 30.0 pixels, the score for θ p r . Therefore, the correct proportion of the two hands was also scored to be one. The distance a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) was estimated to be 110.7 pixels less than 200.0 pixels, the score for ε , but the distance a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) was estimated to be 292.0 pixels greater than 200.0 pixels, the score for ε . Therefore, the correctness of the hand target number was evaluated to be one, but the correctness of the minute target number was also evaluated to be one. Finally, the total score of the hands parameter was evaluated to be four.
Figure 9d displays the case in which two hands are present with the proper target numbers but not the proper proportions. In this case, the percentage p f h was evaluated to be 89.8% greater than 65.0%, the score for θ h 1 . Therefore, the presence of two hands was scored to be two. The length difference Δ h was evaluated to be 63.9 pixels less than 30.0 pixels, the score for θ p r . Therefore, the correct proportion of the two hands was also scored to be zero. Both of the distances a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) and a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) were estimated to be 60.5 and 109.5 pixels less than 200.0 pixels, the score for ε , respectively. Therefore, both the correctness of the hand and minute target numbers were evaluated to be one. Finally, the total score of the hands parameter was evaluated to be four.
Figure 9e displays the case in which two hands are present with the proper proportions but not the proper target numbers. In this case, the percentage p f h was evaluated to be 91.1% greater than 65.0%, the score for θ h 1 . Therefore, the presence of two hands was scored to be two. The length difference Δ h was evaluated to be 37.80 pixels greater than 30.0 pixels, the score for θ p r . Therefore, the correct proportion of the two hands was also scored to be one. Both of the distances a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) and a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) were estimated to be 610.1 and 540.1 pixels greater than 200.0 pixels, the score for ε , respectively. Therefore, both the correctness of hand and minute target numbers were evaluated to be zero. Finally, the total score of the hands parameter was evaluated to be three.
Figure 9f displays the case in which only one hand is present with the proper target number. In this case, the percentage p f h was evaluated to be 64.4% less than 65.0%, the score for θ h 1 , and greater than 50.0%, the score for θ h 2 . Therefore, the presence of two hands was scored to be one. The distance a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) was estimated to be 133.6 pixels less than 200.0 pixels, the score for ε but the distance a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) was estimated to be 1888.6 pixels greater than 200.0 pixels, the score for ε . Therefore, the correctness of the hand target number was evaluated to be one, and the correctness of the minute target number was also evaluated to be one. Finally, the total score of the hands parameter was evaluated to be two.
Figure 9g displays the case in which only one hand is present with neither the proper proportions nor the target numbers. In this case, the percentage p f h was evaluated to be 63.7% less than 65.0%, the score for θ h 1 and greater than 50.0%, the score for θ h 2 . Therefore, the presence of two hands was scored to be one. Both of the distances a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) and a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) were estimated to be 229.63 greater than 200.0 pixels, the score for ε . Therefore, the correctness of hand target number was evaluated to be zero. Finally, the total score of the hands parameter was evaluated to be one.
Figure 9h displays the case in which no hands are present. Therefore, the total score of the hands parameter was evaluated to be zero.

3.4. Scoring on Criteria of Center Parameter

For presence or inference of a center in the hand drawn images, Figure 10 represents four cases of three having center points detected or inferred and the other having no center point. In each of the cases, the original clock drawing image example (left) and the corresponding parameter values (right) are presented.
Figure 10a displays the case in which two hands are present so the center is inferred. In this case, the percentage p f h was evaluated to be 100.00% greater than 65.0%, the score for θ h 1 . Therefore, the presence of a center point was evaluated to be one.
Figure 10b displays the case in which only one hand is present so the center is inferred. In this case, the percentage p f h was evaluated to be 64.4% less than 65.0%, the score for θ h 1 , and greater than 50.0%, the score for θ h 2 . This case also was evaluated to be one for the presence of a center point.
Figure 10c displays the case in which no hands are present, but a data point exists near to the center of the contour. Here, the number of the sensor data points P ( x [ n ] , y [ n ] ) with distance a b s ( P ( x [ n ] , y [ n ] ) P ( x m i d c , y m i d c ) ) , less than 75.0 pixels, the given heuristic value of ε c , from the predefined center point P ( x m i d c , y m i d c ) , which was evaluated to be 94. The presence of a center point was inferred and therefore evaluated to be one.
Figure 10d shows the case in which no center point is present or inferred. In this case, there are no hands and the number of the sensor data points P ( x [ n ] , y [ n ] ) with distance a b s ( P ( x [ n ] , y [ n ] ) P ( x m i d c , y m i d c ) ) less than 75.0 pixels, the given heuristic value of ε c , from the predefined center point P ( x m i d c , y m i d c ) , which was evaluated to be zero. Therefore, the presence or the inference of a center was evaluated to be zero.

3.5. Performance Test Result

A total of 219 drawing images were used to test the performance of the scoring method by mCDT. Table 5 summarizes the frequency of the ground truth for the 219 images with the score in each of the parameters. For the parameter contour, the frequencies were 217, 178, and 215, with an error in estimation of 6, 13, and 1, respectively for the criteria of the circular contour, closure contour, and appropriately sized contour. For the parameter numbers, the frequencies were 153, 181, 88, and 202 with an error in estimation of 11, 5, 2, and 2, respectively, for the criteria of all the numbers present without additional numbers, numbers in corrected order, numbers in correct positions, and numbers within the contour. For the parameter hands, the frequencies were 171, 181, 170, 153, and 149 with an error in estimation of 13, 6, 13, 1, and 6, respectively, for the criteria of presence of two hands, presence of one hand, correct proportion of two hands, hour target number indication and minute target number indication. For the parameter center, the frequency was 190 with an error in estimation of three for the criteria of presence or inference of a center.
Table 6 and Table 7 list the distribution of the estimated scores and the performance of each scoring parameter, respectively in total as well as in the two separate groups, young volunteers and PD patients. As shown in Table 7, for the parameter contour, sensitivity, specificity, accuracy and precision, values were 89.33%, 92.68%, 89.95% and 98.15%; for numbers, they were 80.21%, 95.93%, 89.04% and 93.90%; for hands, they were 83.87%, 95.31%, 87.21% and 97.74%; and for center, they were 98.42%, 86.21%, 96.80% and 97.91%, respectively.

4. Discussion

A conventional CDT based on a paper and pencil test for examining the active and dynamic mechanisms of the cognitive function is inadequate. With the conventional CDT, multiple studies have indicated that a number of brain regions are recruited for the tasks required; these include the temporal lobes, frontal and parietal lobes in addition to the cerebellum, thalamus, premotor area and inferior temporal sulcus, the bilateral parietal lobe, and the sensorimotor cortex [39,40]. What is not clearly known are which portions of the cognitive function are required for recruiting these areas, as with the conventional CDT such an association and a quantitation would be difficult to accomplish. Our study sought to address this requirement from the CDT, and by introducing mCDT as a mobile phone application with a qualitative, automatic scoring system of CDT, this may have been realized. As elaborated previously, the mCDT scoring system was constructed using CNN, a convolutional network for digit classification, U-Net, a convolutional network for biomedical image segmentation, and the MNIST database, the Modified National Institute of Standards and Technology database. The sensor data is also collected by mCDT. From the performance test results, the scoring algorithm in mCDT is efficient and accurate when compared with those of the traditional CDT. In addition, mCDT is able to evaluate the relevant components of the cognitive function. The subjects in our study carried out the drawings with a smart pen on a smartphone screen when required to reproduce figures in a setting similar to the conventional CDT using a pen and paper. This method also allows for increased accuracy in gauging the drawing process and also minimizing any noise in an assay for activated brain function. The smartphone could also provide the motor-related markers of speed and pausing as the test is being carried out; in a conventional CDT pencil and paper test, such motor ability function cognitive tools may not be easily implemented. In summary, our study introduces the developed mCDT as a tool for increasing the accuracy required for the cognitive function evaluation in CDT. As described in the performance test results, mCDT showed fairly good statistical indicators, especially excellent values in specificity and precision. Furthermore, the values of specificity and precision for PD patient groups were better than those for the young volunteer group, which suggested that mCDT does classify the two groups very well and consistently so that it is applicable as a diagnostic tool in neurological disease group and also as a correlation tool between the scores of each criteria and the regional functions of the degenerated brain. Of course, the ability as a correlation tool needs to be investigated in future work, some preliminary studies of which is ongoing, with several clinical groups in collaboration with primary care physicians and neurology subspecialists. Furthermore, since the presented CDT scoring method here uses sensor data collected from a smart mobile device and deep learning based algorithm in the CDT image segmentation and processing, other stroke behavior patterns due to neurological disease symptoms such as motor, memory, and cognitive disorders could be additionally extracted using stroke speed variation and touch event sequence patterns that could be estimated from the sensor data; even the CDT scoring is limited to four parameters with thirteen criteria.

5. Conclusions

In this study, a mobile phone application mCDT for the CDT was implemented, and also an automatic and qualitative scoring method in thirteen criteria was developed using mobile sensor data and deep learning algorithms, U-Net and CNN. A young healthy volunteer (n = 238, 147 males and 89 females, aged 23.98 ± 2.83 years) and a PD patient (n = 140, 76 males and 64 females, aged 75.09 ± 8.57 years) group were recruited and participated in the training models DeepC, DeepH and DeepN, and in validating the performance of the CDT scoring algorithm. Most of the resulting overall statistical indicators, sensitivity, specificity, accuracy, and precision greater than 85%, were acquired at the performance validation of the 79 young healthy volunteers and 140 PD patients. Two exceptions were recognized at the sensitivities of the number and the hands parameters. Especially, the specificities of the contour and hand parameter of the young volunteer group were shown to be far too low (60.00% and 66.67%, respectively), which was because the number of true negatives and false positives were a lot smaller, as well as they were in relatively similar proportions. Furthermore, the specificities and the precisions of the PD patients group were better than those of the young volunteer group, which suggests that the mCDT along with the scoring method is available to be used as a tool of classifying neurological disease groups and also as a tool of scaling the disease symptoms related to degenerated regions of the brain. Further clinical studies should be established in differentiating neurological disease subtypes, being valuable in clinical practice and for studies in the field.

Author Contributions

Conceptualization, U.L.; data curation, I.P. and U.L.; formal analysis, U.L.; funding acquisition, U.L.; investigation, U.L.; methodology, U.L.; project administration, U.L.; resources, U.L.; software, I.P. and U.L.; supervision, U.L.; validation, U.L.; visualization, I.P. and U.L.; writing—original draft, U.L.; writing—review and editing, U.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research (2020R1F1A1048281) and Hallym University Research Fund (HRF-202003-021).

Institutional Review Board Statement

The Institutional Review Board of the Hallym University Sacred Heart Hospital approved the gathering data and the protocols used for this study (IRB number: 2019-03-001).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy issue.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Biundo, R.; Weis, L.; Bostantjopoulou, S.; Stefanova, E.; Falup-Pecurariu, C.; Kramberger, M.G.; Geurtsen, G.J.; Antonini, A.; Weintraub, D.; Aarsland, D. MMSE and MoCA in Parkinson’s disease and dementia with Lewy bodies: A multicenter 1-year follow-up study. J. Neural. Transm. 2016, 123, 431–438. [Google Scholar] [CrossRef] [Green Version]
  2. Mittal, C.; Gorthi, S.P.; Rohatgi, S. Early Cognitive Impairment: Role of Clock Drawing Test. Med. J. Armd. Forces India 2010, 66, 25–28. [Google Scholar] [CrossRef] [Green Version]
  3. Aprahamian, I.; Martinelli, J.E.; Neri, A.L.; Yassuda, M.S. The Clock Drawing Test: A review of its accuracy in screening for dementia. Dement. Neuropsychol. 2009, 3, 74–81. [Google Scholar] [CrossRef]
  4. Youn, Y.C.; Pyun, J.-M.; Ryu, N.; Baek, M.J.; Jang, J.-W.; Park, Y.H.; Ahn, S.-W.; Shin, H.-W.; Park, K.-Y.; Kim, S.Y. Use of the Clock Drawing Test and the Rey–Osterrieth Complex Figure Test-copy with convolutional neural networks to predict cognitive impairment. Alzheimer’s Res. Ther. 2021, 13, 1–7. [Google Scholar] [CrossRef]
  5. Straus, S.H. Use of the automatic clock drawing test to rapidly screen for cognitive impairment in older adults, drivers, and the physically challenged. J. Am. Geriatr. Soc. 2007, 55, 310–311. [Google Scholar] [CrossRef] [PubMed]
  6. Chen, S.; Stromer, D.; Alabdalrahim, H.A.; Schwab, S.; Weih, M.; Maier, A. Automatic dementia screening and scoring by applying deep learning on clock-drawing tests. Sci. Rep. 2020, 10, 1–11. [Google Scholar] [CrossRef]
  7. Park, I.; Kim, Y.J.; Kim, Y.J.; Lee, U. Automatic, Qualitative Scoring of the Interlocking Pentagon Drawing Test (PDT) Based on U-Net and Mobile Sensor Data. Sensors 2020, 20, 1283. [Google Scholar] [CrossRef] [Green Version]
  8. Mann, D.L. Heart Failure: A Companion to Braunwald’s Heart Disease; Elsevier: Amsterdam, The Netherlands, 2011. [Google Scholar] [CrossRef]
  9. Spenciere, B.; Alves, H.; Charchat-Fichman, H. Scoring systems for the Clock Drawing Test: A historical review. Dement. Neuropsychol. 2017, 11, 6–14. [Google Scholar] [CrossRef] [Green Version]
  10. Eknoyan, D.; Hurley, R.A.; Taber, K.H. The Clock Drawing Task: Common Errors and Functional Neuroanatomy. J. Neuropsychiatry Clin. Neurosci. 2012, 24, 260–265. [Google Scholar] [CrossRef]
  11. Talwar, N.A.; Churchill, N.W.; Hird, M.A.; Pshonyak, I.; Tam, F.; Fischer, C.E.; Graham, S.J.; Schweizer, T.A. The Neural Correlates of the Clock-Drawing Test in Healthy Aging. Front. Hum. Neurosci. 2019, 13, 25. [Google Scholar] [CrossRef]
  12. Yuan, J.; Libon, D.J.; Karjadi, C.; Ang, A.F.; Devine, S.; Auerbach, S.H.; Au, R.; Lin, H. Association Between the Digital Clock Drawing Test and Neuropsychological Test Performance: Large Community-Based Prospective Cohort (Framingham Heart Study). J. Med. Internet Res. 2021, 23, e27407. [Google Scholar] [CrossRef] [PubMed]
  13. Shulman, K. Clock-drawing: Is it the ideal cognitive screening test? Int. J. Geriatr. Psychiatry 2000, 15, 548–561. [Google Scholar] [CrossRef]
  14. Sunderland, T.; Hill, J.L.; Mellow, A.M.; Lawlor, B.A.; Gundersheimer, J.; Newhouse, P.A.; Grafman, J.H. Clock drawing in Alzheimer’s disease. Nov. Meas. Dement. Sev. J. Am. Geriatr Soc. 1989, 37, 725–729. [Google Scholar] [CrossRef] [PubMed]
  15. Nasreddine, Z.S.; Phillips, N.A.; Bédirian, V.; Charbonneau, S.; Whitehead, V.; Collin, I. The Montreal Cognitive Assessment, MoCA: A brief screening tool for mild cognitive impairment. J. Am. Geriatr Soc 2005, 53, 695–699. [Google Scholar] [CrossRef]
  16. Souillard-Mandar, W.; Davis, R.; Rudin, C.; Au, R.; Libon, D.J.; Swenson, R.; Price, C.C.; Lamar, M.; Penney, D.L. Learning Classification Models of Cognitive Conditions from Subtle Behaviors in the Digital Clock Drawing Test. Mach. Learn. 2016, 102, 393–441. [Google Scholar] [CrossRef]
  17. Nirjon, S.; Emi, I.A.; Mondol, A.S.; Salekin, A.; Stankovic, J.A. MOBI-COG: A Mobile Application for Instant Screening of Dementia Using the Mini-Cog Test. In Proceedings of the Wireless Health 2014 on National Institutes of Health, Bethesda, MD, USA, 1–7 October 2014. [Google Scholar] [CrossRef]
  18. Fabricio, A.T.; Aprahamian, I.; Yassuda, M.S. Qualitative analysis of the Clock Drawing Test by educational level and cognitive profile. Arq. Neuropsiquiatr. 2014, 72, 289–295. [Google Scholar] [CrossRef] [Green Version]
  19. Borson, S.; Scanlan, J.; Brush, M.; Vitaliano, P.; Dokmak, A. The mini-cog: A cognitive “vital signs” measure for dementia screening in multi-lingual elderly. Int. J. Geriatr. Psychiatry 2000, 15, 1021–1027. [Google Scholar] [CrossRef]
  20. Harbi, Z.; Hicks, Y.; Setchi, R. Clock Drawing Test Interpretation System. Procedia Comput. Sci. 2017, 112, 1641–1650. [Google Scholar] [CrossRef]
  21. Kim, H.; Cho, Y.S.; Do, E.I. Computational clock drawing analysis for cognitive impairment screening. In Proceedings of the Fifth International Conference on Tangible, Embedded, and Embodied Interaction; Gross, M.D., Ed.; ACM: New York, NY, USA, 2011; pp. 297–300. [Google Scholar]
  22. Caffarraa, P.; Gardinia, S.; Diecib, F.; Copellib, S.; Maseta, L.; Concaria, L.; Farinac, E.; Grossid, E. The qualitative scoring MMSE pentagon test (QSPT): A new method for differentiating dementia with Lewy Body from Alzheimer’s Disease. Behav. Neurol. 2013, 27, 213–220. [Google Scholar] [CrossRef]
  23. Davis, R.; Libon, D.; Au, R.; Pitman, D.; Penney, D. Think: Inferring cognitive status from subtle behaviors. IEEE Int. Conf. Robot. Autom. 2014, 2898–2905. Available online: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4825804/ (accessed on 2 August 2021). [CrossRef] [Green Version]
  24. Manos, P.J.; Wu, R. The Ten Point Clock Test: A Quick Screen and Grading Method for Cognitive Impairment in Medical and Surgical Patients. Int. J. Psychiatry Med. 1994, 24, 229–244. [Google Scholar] [CrossRef] [PubMed]
  25. Royall, D.R.; Cordes, J.A.; Polk, M. CLOX: An executive clock drawing task. J. Neurol. Neurosurg. Psychiatry 1998, 64, 588–594. [Google Scholar] [CrossRef] [PubMed]
  26. Rouleau, I.; Salmon, D.P.; Butters, N.; Kennedy, C.; McGuire, K. Quantitative and qualitative analyses of clock drawings in Alzheimer’s and Huntington’s disease. Brain Cogn. 1992, 18, 70–87. [Google Scholar] [CrossRef]
  27. Muayqil, T.A.; Tarakji, A.R.; Khattab, A.M.; Balbaid, N.T.; Al-Dawalibi, A.M.; AlQarni, S.A.; Hazazi, R.A.; Alanazy, M.H. Comparison of Performance on the Clock Drawing Test Using Three Different Scales in Dialysis Patients. Behav. Neurol. 2020, 2020, 1–7. [Google Scholar] [CrossRef] [PubMed]
  28. Shao, K.; Dong, F.; Guo, S.; Wang, W.; Zhao, Z.; Yang, Y.; Wang, P.; Wang, J. Clock-drawing test: Normative data of three quantitative scoring methods for Chinese-speaking adults in Shijiazhuang City and clinical utility in patients with acute ischemic stroke. Brain Behav. 2020, 10, e01806. [Google Scholar] [CrossRef]
  29. De Pandis, M.F.; Galli, M.; Vimercati, S.; Cimolin, V.; De Angelis, M.V.; Albertini, G. A New Approach for the Quantitative Evaluation of the Clock Drawing Test: Preliminary Results on Subjects with Parkinson’s Disease. Neurol. Res. Int. 2010, 2010, 1–6. [Google Scholar] [CrossRef] [Green Version]
  30. Guha, A.; Kim, H.; Do, E.Y. Automated Clock Drawing Test through Machine Learning and Geometric Analysis. In Proceedings of the 16th International Conference on Distributed Multimedia Systems, DMS 2010, Hyatt Lodge at McDonald’s Campus, Oak Brook, IL, USA, 14–16 October 2010. [Google Scholar]
  31. William, S.; Randall, D.; Cynthia, R.; Rhoda, A.; Dana, L.P. Interpretable Machine Learning Models for the Digital Clock Drawing Test. In Proceedings of the 2016 ICML Workshop on Human Interpretability in Machine Learning (WHI 2016), New York, NY, USA, 23 June 2016. [Google Scholar]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Huang, G.; Liu, Z.; Weinberger, K.Q. Densely Connected Convolutional Networks. 2016. Available online: http://arxiv.org/abs/1608.06993 (accessed on 2 August 2021).
  34. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. mini-mental state: A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef]
  35. Available online: https://docs.opencv.org/3.4/df/d0d/tutorial_find_contours.html (accessed on 2 August 2021).
  36. Available online: https://www.kite.com/python/docs/difflib.SequenceMatcher (accessed on 2 August 2021).
  37. Available online: https://numpy.org/doc/stable/reference/generated/numpy.polyfit.html (accessed on 2 August 2021).
  38. Available online: https://pythonprogramming.net/how-to-program-best-fit-line-machine-learning-tutorial/ (accessed on 2 August 2021).
  39. Freedman, M.; Leach, L.; Kaplan, E.; Winocur, G.; Shulman, K.I.; Delis, D. Clock Drawing: A Neuropsychological Analysis; Oxford University Press, Inc.: Oxford, UK, 1994. [Google Scholar]
  40. Mendes-Santos, L.C.; Mograbi, D.; Spenciere, B.; Charchat-Fichman, H. Specific algorithm method of scoring the Clock Drawing Test applied in cognitively normal elderly. Dement. Neuropsychol. 2015, 9, 128–135. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) Flow diagram of mCDT operation; screen shots of (b) registration window; (c) CDT window; and (d) results window of mCDT.
Figure 1. (a) Flow diagram of mCDT operation; screen shots of (b) registration window; (c) CDT window; and (d) results window of mCDT.
Sensors 21 05239 g001
Figure 2. U-Net network architecture for this study. The U-Net network architecture is composed of a contracting path, an expansive path, and a final layer. The contracting path is made up of repetitive applications of two 3 × 3 convolutions and a 2 × 2 max pooling operation with stride 2 for down-sampling. The expansive path exists of two 3 × 3 convolutions and a 2 × 2 convolution for up-sampling. The final layer employs a 1 × 1 convolution for mapping each 16-component feature vector for the required number of classes.
Figure 2. U-Net network architecture for this study. The U-Net network architecture is composed of a contracting path, an expansive path, and a final layer. The contracting path is made up of repetitive applications of two 3 × 3 convolutions and a 2 × 2 max pooling operation with stride 2 for down-sampling. The expansive path exists of two 3 × 3 convolutions and a 2 × 2 convolution for up-sampling. The final layer employs a 1 × 1 convolution for mapping each 16-component feature vector for the required number of classes.
Sensors 21 05239 g002
Figure 3. CNN network architecture for this study. It is made up of two convolution layers (C1 and C3), two pooling layers (D2 and D4), and two fully connected layers (F5 and F6). C1, the first convolution layer, filters the 28 × 28 input number image with 32 kernels of 5 × 5 size; C3, the second convolution layer, filters the down-sampled 12 × 12 × 32 feature maps with 64 kernels of 5 × 5 × 32 size. Both of the convolution layers use a unit stride, and at the output of each, a ReLU nonlinear function is used. At layers D2 and D4, down-sampling is performed with 2 × 2 non-overlapping max pooling. For the two final fully-connected layers, F5 and F6, they respectively have 1024 and 10 neurons.
Figure 3. CNN network architecture for this study. It is made up of two convolution layers (C1 and C3), two pooling layers (D2 and D4), and two fully connected layers (F5 and F6). C1, the first convolution layer, filters the 28 × 28 input number image with 32 kernels of 5 × 5 size; C3, the second convolution layer, filters the down-sampled 12 × 12 × 32 feature maps with 64 kernels of 5 × 5 × 32 size. Both of the convolution layers use a unit stride, and at the output of each, a ReLU nonlinear function is used. At layers D2 and D4, down-sampling is performed with 2 × 2 non-overlapping max pooling. For the two final fully-connected layers, F5 and F6, they respectively have 1024 and 10 neurons.
Sensors 21 05239 g003
Figure 4. Overall flowchart of the scoring method, (a) for parameter contour, (b) for parameters hands and center and (c) for parameter numbers.
Figure 4. Overall flowchart of the scoring method, (a) for parameter contour, (b) for parameters hands and center and (c) for parameter numbers.
Sensors 21 05239 g004
Figure 5. Flowcharts for scoring presence of all the numbers and no additional numbers; for scoring correctness of the order of the numbers; for scoring correctness of the positions of the numbers; and for scoring positioning of the numbers within the contour.
Figure 5. Flowcharts for scoring presence of all the numbers and no additional numbers; for scoring correctness of the order of the numbers; for scoring correctness of the positions of the numbers; and for scoring positioning of the numbers within the contour.
Sensors 21 05239 g005
Figure 6. Flowcharts for scoring presence of one or two hands, correctness of the proportion of the hands, correctness of the hour target number, and correctness of the minute target number.
Figure 6. Flowcharts for scoring presence of one or two hands, correctness of the proportion of the hands, correctness of the hour target number, and correctness of the minute target number.
Sensors 21 05239 g006
Figure 7. Four cases of original drawings (left) along with their segmented images (middle) produced by the pre-trained model, DeepC, and their corresponding velocity graphs (right). Four representative examples demonstrating how contours in original images are detected and classified in their types and sizes, showing the original image (first column) with the segmented image (second column), the overlapping image (third column) and the corresponding parameter values (fourth column) of each examples; (a) the case in which the original image is of a closed circular contour sized appropriately; (b) the case in which the original image is of a circular contour sized appropriately, but not wholly closed; (c) the case in which the original image is of an appropriately sized but neither closed nor circular contour; and (d) the case in which the original image is of a closed circular contour but sized not appropriately. * Total score/score of contour/score of numbers/score of hands/score of center.
Figure 7. Four cases of original drawings (left) along with their segmented images (middle) produced by the pre-trained model, DeepC, and their corresponding velocity graphs (right). Four representative examples demonstrating how contours in original images are detected and classified in their types and sizes, showing the original image (first column) with the segmented image (second column), the overlapping image (third column) and the corresponding parameter values (fourth column) of each examples; (a) the case in which the original image is of a closed circular contour sized appropriately; (b) the case in which the original image is of a circular contour sized appropriately, but not wholly closed; (c) the case in which the original image is of an appropriately sized but neither closed nor circular contour; and (d) the case in which the original image is of a closed circular contour but sized not appropriately. * Total score/score of contour/score of numbers/score of hands/score of center.
Sensors 21 05239 g007aSensors 21 05239 g007b
Figure 8. Eight representative examples demonstrating how numbers in original images are detected and classified in their orders and positions, showing the binarized image (left), the original image with the cropped areas for the numbers (middle), and the corresponding parameter values (right) of each example; (a) the case in which all the numbers without any additional numbers are present in the correct orders and the proper positions within the contour; (b) the case in which all the numbers without any additional numbers are present in the correct orders within the contour but not in proper positions; (c) the case in which all the numbers without any additional numbers are present in the correct orders but neither in the proper positions nor within the contour; (d) the case in which all the numbers without any additional numbers are present in the correct orders within the contour but not in the proper positions; (e) the case in which some numbers are missing and the presented numbers are not in proper positions but mostly in correct order within the contour; (f) the case in which there are additional numbers not belonging to a clock but the numbers are in correct orders within the contour; (g) the case in which some numbers are missing and the presented numbers are not in proper positions but mostly in correct order within the contour; and (h) the case in which many numbers are missing and the presented numbers are not in proper positions and correct order but within the contour. * Total score/score of contour/score of numbers/score of hands/score of center.
Figure 8. Eight representative examples demonstrating how numbers in original images are detected and classified in their orders and positions, showing the binarized image (left), the original image with the cropped areas for the numbers (middle), and the corresponding parameter values (right) of each example; (a) the case in which all the numbers without any additional numbers are present in the correct orders and the proper positions within the contour; (b) the case in which all the numbers without any additional numbers are present in the correct orders within the contour but not in proper positions; (c) the case in which all the numbers without any additional numbers are present in the correct orders but neither in the proper positions nor within the contour; (d) the case in which all the numbers without any additional numbers are present in the correct orders within the contour but not in the proper positions; (e) the case in which some numbers are missing and the presented numbers are not in proper positions but mostly in correct order within the contour; (f) the case in which there are additional numbers not belonging to a clock but the numbers are in correct orders within the contour; (g) the case in which some numbers are missing and the presented numbers are not in proper positions but mostly in correct order within the contour; and (h) the case in which many numbers are missing and the presented numbers are not in proper positions and correct order but within the contour. * Total score/score of contour/score of numbers/score of hands/score of center.
Sensors 21 05239 g008aSensors 21 05239 g008b
Figure 9. Eight representative examples demonstrating how numbers in original images are detected and classified in their orders and positions, showing the binarized image (left), the original image with the cropped areas for the numbers (middle), and the corresponding parameter values (right) of each examples; (a) the case in which two hands are present with the proper proportions and target numbers; (b) another example of the case in which two hands are present with the proper proportions and target numbers, where one of the target numbers is not in the proper position; (c) the case in which two hands are present with the proper proportions but one of them is not indicating the target number; (d) the case in which two hands are present with the proper target numbers but not the proper proportions; (e) the case in which two hands are present with the proper proportions but not the proper target numbers; (f) the case in which only one hand is present with the proper target number; (g) the case in which only one hand is present with neither the proper proportions nor the target numbers; and (h) the case in which no hands are present. * Total score/score of contour/score of numbers/score of hands/score of center.
Figure 9. Eight representative examples demonstrating how numbers in original images are detected and classified in their orders and positions, showing the binarized image (left), the original image with the cropped areas for the numbers (middle), and the corresponding parameter values (right) of each examples; (a) the case in which two hands are present with the proper proportions and target numbers; (b) another example of the case in which two hands are present with the proper proportions and target numbers, where one of the target numbers is not in the proper position; (c) the case in which two hands are present with the proper proportions but one of them is not indicating the target number; (d) the case in which two hands are present with the proper target numbers but not the proper proportions; (e) the case in which two hands are present with the proper proportions but not the proper target numbers; (f) the case in which only one hand is present with the proper target number; (g) the case in which only one hand is present with neither the proper proportions nor the target numbers; and (h) the case in which no hands are present. * Total score/score of contour/score of numbers/score of hands/score of center.
Sensors 21 05239 g009aSensors 21 05239 g009b
Figure 10. Four representative examples demonstrating how the centers in original images are detected to be present or inferred, showing the original image (left) and the corresponding parameter values (right) of each example; (a) the case in which two hands are present so the center is inferred; (b) the case in which only one hand is present so the center is inferred; (c) the case in which no hands are present but a data point exists near to the center of the contour; and (d) the case in which no center point is present or inferred. * Total score/score of contour/score of numbers/score of hands/score of center.
Figure 10. Four representative examples demonstrating how the centers in original images are detected to be present or inferred, showing the original image (left) and the corresponding parameter values (right) of each example; (a) the case in which two hands are present so the center is inferred; (b) the case in which only one hand is present so the center is inferred; (c) the case in which no hands are present but a data point exists near to the center of the contour; and (d) the case in which no center point is present or inferred. * Total score/score of contour/score of numbers/score of hands/score of center.
Sensors 21 05239 g010
Table 1. Statistics of age, gender, handedness and clinical status of the participants.
Table 1. Statistics of age, gender, handedness and clinical status of the participants.
Training SetTest Set
Volunteers
(n = 159)
Volunteers
(n = 79)
PD Patients
(n = 140)
Age24.78 ± 1.6322.81 ± 0.7975.09 ± 8.57
Male (female)112 (45)35 (44)76 (64)
Binary CDT score
Pass (Non-pass)
159 (0)75 (4)73 (67)
Table 2. Detailed list of the parameters for the scoring method.
Table 2. Detailed list of the parameters for the scoring method.
ParametersScoring Criteria
ContourContour is circular
Contour is closed
Contour size is appropriate
NumbersNumbers are all present without additional numbers
Numbers are in the corrected order
Numbers are in the correct positions
Numbers are within the contour
HandsTwo hands are present
One hand is present
Hour target number is indicated
Minute target number is indicated
Hands are in correct proportion
CenterA center is drawn or inferred
Total0–13 scores
Table 3. Formulas of the pre-estimated positions of the number digits from 0 to 12.
Table 3. Formulas of the pre-estimated positions of the number digits from 0 to 12.
Number Digit
k
Formula
x d [ k ] y d [ k ]
1 ( 2 x m i d c + x max c ) / 3 ( 2 y max c + y m i d c ) / 3
2 ( x m i d c + 2 x max c ) / 3 ( y max c + 2 y m i d c ) / 3
3 x max c y m i d c
4 ( x m i d c + 2 x max c ) / 3 ( y min c + 2 y m i d c ) / 3
5 ( 2 x m i d c + x max c ) / 3 ( 2 y min c + y m i d c ) / 3
6 x m i d c y min c
7 ( 2 x m i d c + x min c ) / 3 ( 2 y min c + y m i d c ) / 3
8 ( x m i d c + 2 x min c ) / 3 ( y min c + 2 y m i d c ) / 3
9 x min c y m i d c
10 ( x m i d c + 2 x min c ) / 3 ( y max c + 2 y m i d c ) / 3
11 ( 2 x m i d c + x min c ) / 3 ( 2 y max c + y m i d c ) / 3
12 x m i d c y max c
Table 4. Details for assignment of scores.
Table 4. Details for assignment of scores.
ParametersCriteriaConditions (Scoring Method) *
Contourcircular contour p f c θ c 1 1
closed contour d max c θ c 2 2
appropriately sized contour A c / W c θ c 3 3
Numbersall and no additional numbers N = 15 & 0 D [ j ] 9 for all j & n(D[j] = 1) = 5 & n(D[j] = 2) = 2
correct order of numbers R s e q θ n  4
correct position of numbers n ( d c n [ j ] d c 5 ) / n ( d c n [ j ] ) > θ d c 6
positioning of numbers within contour R c N < R c L ,
Handstwo hands p f h > θ h 1  7
one hand p f h > θ h 2 8
correct proportion of hands p f h > θ h 1 & a b s ( n ( S 1 ) n ( S 2 ) ) > θ p r  9
correct hour target number a b s ( P ( x [ n h ] , y [ n h ] ) P ( x [ h t ] , y [ h t ] ) ) < ε  10
correct minute target number a b s ( P ( x [ n m ] , y [ n m ] ) P ( x [ m t ] , y [ m t ] ) ) < ε
Centerexistence or inference of a center p f h > θ h 1 or a b s ( P ( x [ n ] , y [ n ] ) P ( x m i d c , y m i d c ) ) < ε c 11
Total sum13
* heuristic values of the thresholds: 1 θ c 1 = 75.00 pixels; 2 θ c 2 = 50.00 pixels; 3 θ c 3 = 0.1 ; 4 θ n = 65.00 pixels; 5 d c = 100.00 pixels; 6  θ d c = 0.65 ; 7  θ h 1 = 65.00 pixels; 8  θ h 2 = 50.00 pixels; 9 θ p r = 30.00 pixels; 10 ε = 200.00 pixels; 11 ε c = 75.00 pixels.
Table 5. Frequency of the ground truth of the 219 images in each criteria of the parameters of the scoring method of mCDT.
Table 5. Frequency of the ground truth of the 219 images in each criteria of the parameters of the scoring method of mCDT.
ParameterCriteriaFrequency
Count (%)
Errors in
Estimation
Count (%)
ContourContour is circular217(99.08)6(2.76)
Contour is closed178(81.27)13(7.30)
Contour size is appropriate215(98.17)1(0.46)
NumbersNumbers are all present without additional numbers153(69.86)11(7.18)
Numbers are in corrected order181(82.64)5(2.76)
Numbers are in the correct positions88(40.18)2(2.27)
Numbers are within the contour202(92.23)2(0.99)
HandsTwo hands are present171(78.08)13(7.60)
One hand is present181(82.64)6(3.31)
Hands are in correct proportion170(77.62)13(7.64)
Hour target number is indicated153(69.86)1(0.65)
Minute target number is indicated149(68.03)6(4.02)
CenterA center is drawn or inferred190(86.75)3(1.57)
Table 6. Distribution of the estimated scores for each scoring parameters in mCDT.
Table 6. Distribution of the estimated scores for each scoring parameters in mCDT.
ScoresContourNumbersHandsCenter
5--144
(70/74)
-
4-81
(53/28)
11
(3/8)
-
3175
(74/101)
71
(23/48)
15
(6/9)
-
242
(5/37)
30
(3/27)
4
(0/4)
-
11
(0/1)
27
(0/27)
7
(0/7)
190
(79/111)
01
(0/1)
10
(0/10)
38
(0/38)
29
(0/29)
Total219
(79/140)
219
(79/140)
219
(79/140)
219
(79/140)
Table 7. Performance of the scoring parameters in mCDT.
Table 7. Performance of the scoring parameters in mCDT.
ContourNumbersHandsCenter
TP159
(70/89)
77
(50/27)
130
(66/64)
187
(79/108)
FP3
(2/1)
5
(4/1)
3
(3/0)
4
(0/4)
FN19
(4/15)
19
(5/14)
25
(4/21)
3
(0/3)
TN38
(3/35)
118
(20/98)
61
(6/55)
25
(0/25)
Sensitivity89.33
(94.60/85.58)
80.21
(90.91/65.85)
83.87
(94.29/75.29)
98.42
(100.00/97.30)
Specificity92.68
(60.00/97.22)
95.93
(83.33/98.99)
95.31
(66.67/100.00)
86.21
(-/86.21)
Accuracy89.95
(92.41/88.57)
89.04
(88.61/89.29)
87.21
(91.14/85.00)
96.80
(100.00/95.00)
Precision98.15
(97.22/98.89)
93.90
(92.59/96.43)
97.74
(95.65/100.00)
97.91
(100.00/96.43)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, I.; Lee, U. Automatic, Qualitative Scoring of the Clock Drawing Test (CDT) Based on U-Net, CNN and Mobile Sensor Data. Sensors 2021, 21, 5239. https://doi.org/10.3390/s21155239

AMA Style

Park I, Lee U. Automatic, Qualitative Scoring of the Clock Drawing Test (CDT) Based on U-Net, CNN and Mobile Sensor Data. Sensors. 2021; 21(15):5239. https://doi.org/10.3390/s21155239

Chicago/Turabian Style

Park, Ingyu, and Unjoo Lee. 2021. "Automatic, Qualitative Scoring of the Clock Drawing Test (CDT) Based on U-Net, CNN and Mobile Sensor Data" Sensors 21, no. 15: 5239. https://doi.org/10.3390/s21155239

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop