Figures
Abstract
Background
Identifying bradykinesia is crucial for diagnosing Parkinson’s disease (PD). Traditionally, the finger-tapping test has been used, relying on subjective assessments by physicians. Computer vision offers a non-contact and cost-effective alternative for assessing Parkinson’s disease.
Objective
This study aimed to detect Parkinson’s disease by identifying bradykinesia using computer vision in the finger-tapping test and applying machine learning techniques for both hands.
Methods
We recruited 100 patients with PD and healthy controls. Four neurologists assessed bradykinesia, and 10-second smartphone-recorded finger-tapping movements were analyzed using Google MediaPipe Hands software. Six machine learning models were trained using a nested cross-validation framework.
Results
The differences in tapping scores between the left and right hands were significantly greater in the PD group (2.8 (5.0) vs 0.4 (0.7), p = 0.001) than in the healthy controls. Moreover, the tapping amplitude variation and all amplitude decremental parameters in the PD group differed significantly from those of the standard controls. The PD group had significantly lower tapping scores than the normal subjects (right: 17.9 (7.8)/ left: 17.9 (5.6) vs. right: 24.6 (7.3)/ left: 24.6 (7.2), p < 0.001). The support vector machine outperformed the other models. The most influential features were the tapping difference, followed by the tapping score (right hand) and tapping amplitude mean (right hand).
Citation: Kasemsap N, Tikkapanyo P, Wanjantuk P, Vorasoot N, Kongbunkiat K, Panitchote A (2025) Screening for Parkinson’s disease using “computer vision”. PLoS One 20(8): e0330373. https://doi.org/10.1371/journal.pone.0330373
Editor: John A. Thompson, University of Colorado Anschutz Medical Campus, UNITED STATES OF AMERICA
Received: September 28, 2024; Accepted: July 30, 2025; Published: August 12, 2025
Copyright: © 2025 Kasemsap et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: Our data are already publicly available. The code and all data underlying the results presented in this study are available on figshare • README. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28883756.v2 • File for setting up the environment. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28931441.v1 • Python code for the Jupyter notebook. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28883624.v2 • Parkinson_data.csv. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28883489.v1 • Example Video : Finger tapping for bradykinesia detection. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28930706.v1 Example extracted data from the finger-tapping task (1 case) for model prediction. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28930694.v1 • SVC model for bradykinesia detection. figshare. Dataset. https://doi.org/10.6084/m9.figshare.28883657.v2.
Funding: The study received funding from the Faculty of Medicine, Khon Kaen University, Thailand (Grant Number RU65203) and the Neurological Society of Thailand (no grant number). The respective URLs are https://resmd.kku.ac.th/web/ and http://www.neurothai.org. NK received these grants. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Parkinson’s disease (PD) is clinically defined by the presence of bradykinesia and at least one additional cardinal motor feature (rigidity or rest tremor), along with supporting and exclusionary criteria [1,2]. Accurate identification of bradykinesia is crucial for diagnosing PD, as it is a prerequisite for diagnosis [2]. Various tasks, such as finger tapping, pronation-supination movements, toe-tapping, and foot tapping, have been used subjectively to evaluate bradykinesia. Clinical rating scales, such as the Movement Disorder Society revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS), have been developed to assess both motor and non-motor symptoms of PD. These scales have shown high consistency and correlation with the original UPDRS [3]. However, they rely on subjective scoring based on a physician’s clinical experience, limiting their ability to detect mild bradykinesia. In order to address this, technology-based tools have been developed to quantify bradykinesia more objectively. These include a computer software program called BRAIN TEST© that assesses finger-tapping speed and accuracy [4]. However, it requires a keyboard and may not be helpful for patients with severe hand dysfunction. Another tool, the Coordination Ability Test System (CATSYS), utilizes sensors such as a tremor pen, touch recording plate, reaction time handle, and force plate for balance recording [5]. While CATSYS correlated with the UPDRS item for pronation/supination, it did not correlate with finger-tapping values and corresponding UPDRS bradykinesia items. Quantitative assessment of finger movement performance on a computer keyboard [6] and musical keyboard [7] demonstrates a correlation with clinical phenomena of repetitive movement and can identify motor signs in early PD [8]. A touchscreen-based tapping application was comparable to a conventional method for assessing motor dysfunction [9]. Gyrosensors have also been used to measure angular movement and have shown a moderate correlation with clinical finger-tap scores [10]. However, these tools require sensors that may not be widely available or specific computer programs with which patients need to interact.
Computer vision (CV) offers several advantages in assessing Parkinson’s disease (PD), including its non-contact nature and minimal instrumentation requirements. CV can be easily implemented using standard devices, such as computers with webcams or smartphones with cameras. Previous studies have utilized various computer vision techniques to track finger movement and normalize tapping signals in PD assessment. These techniques include motion history images and face height detection [11]. In a recent study, smartphone video recording at 60 frames per second (fps) and an optical flow time series were employed for finger detection. The researchers also utilized DeepLabCut, a tool for pose estimation, to quantify bradykinesia. This study demonstrated the potential of computer vision for assessing PD-related bradykinesia [12–14]. The Google MediaPipe Hands (GMH) framework [15], which is open source and compatible with multiple programming languages, including Python, is a powerful tool for extracting hand landmarks from videos. It provides accurate tracking of hand and finger movements, thereby enabling a detailed analysis.
To our knowledge, our study is the first to use computer vision techniques for screening Parkinson’s disease by detecting bradykinesia in the finger-tapping test for both hands.
Materials and methods
The study was approved by the Khon Kaen University Ethics Committee for Human Research (Khon Kaen, Thailand) (Approval #HE644021).
Participants
We recruited a convenience sample of 53 patients from Srinagarind Hospital, Khon Kaen University. The recruitment period was from February 22, 2022, to November 30, 2023. Participants were selected based on recent visits to the clinic in a hospital setting, without any specific selection criteria. A neurologist had previously diagnosed all patients with Parkinson’s disease according to the clinical diagnostic criteria of the Movement Disorder Society [2]. Healthy controls were excluded if they had bradykinesia. Controls without Parkinson’s disease or any other neurological diagnosis were recruited from the companions of patients or hospital/university staff members. Patients with motor weakness, joint stiffness, joint deformity, or inability to perform the finger-tapping test due to arthritis or pain were excluded. Written informed consent was obtained from all participants.
Four neurologists rated 100 videos. The raters were blinded to patient or healthy control status and each other’s scores. Each video was rated according to the presence of bradykinesia and MDS-UPDRS item 3.4 (finger tapping). For each video, the most common rating given by the clinician was used to measure how well it correlated with a computer vision program assessment. As assessed using Krippendorff’s alpha, the overall agreement was 0.72 (95% bootstrap confidence intervals 0.57–0.78) in all cases with at least two ratings. During video recording, the patients were not intentionally withheld from medication.
Video recording
Finger-tapping movements were recorded using an ordinary smartphone for 10 seconds (s). The recording was performed at a frame rate of 30 fps. The participants were instructed to perform a finger-tapping (FT) test. Before recording, they were asked to stretch their index finger vertically against the thumb as much as possible and then perform tapping as quickly and wide open as possible (i.e., between the thumb and index fingertip as possible). The camera was approximately 0.5–1 m from the hand without a tripod.
Pose estimation and accuracy
MediaPipe hands framework.
The Google MediaPipe Hands (GMH) framework is a real-time hand-tracking system that leverages machine learning to detect and predict 21 three-dimensional (3D) landmarks for each hand. The framework consists of two submodules: palm and hand landmark detection. Using a computationally efficient and robust model. This bounding box served as the input for the hand landmark detection submodule. A real-time regression model predicted the 3D coordinates of 21 key landmarks (e.g., fingertips, knuckles) within the detected bounding boxes. The coordinates of each landmark, including those of the hand and knuckles, consisted of x, y, and z values normalized to the range of 0.0 to 1.0 [16].
The GMH framework has been effectively validated for hand-tracking and motion analysis accuracy. It has been applied to detect and measure resting tremors in patients with Parkinson’s disease, and its performance has been confirmed against accelerometer-based systems [17,18]. The validation study compared the GMH with a gold standard system using three dynamic hand exercises, namely hand opening and closing, single-finger tapping, and multiple-finger tapping, which showed strong temporal and spectral consistency. This suggests a distance range of 60–100 cm., various motion speeds (slow, normal, and fast), and a lateral view angle [19].
Confidence thresholds and handling noisy data.
The GMH framework utilizes a palm detection model as an initial step designed to exclude non-hand objects and reduce the number of false positives. By incorporating bounding box localization and pre-trained models, the framework exhibits robustness across various settings, including challenging conditions such as low light and partial occlusions. These features make it highly applicable to clinical and real-world scenarios, particularly in video-based analysis. The precision of the palm detector has been reported to range between 86.2% and 95.7% [19]. Additionally, the framework ensures stable pose predictions during continuous hand movement by implementing a temporal smoothing mechanism. However, a key limitation is that hand landmarks become unreliable when hands are partially obscured, and detection accuracy declines considerably in environments characterized by significant noise or dynamic activity. For stabilized GMH, temporal filtering techniques stabilize landmark detection across video frames. This helps smooth out the noise and jittering effects that tremors exacerbate. We achieved this by adjusting the confidence thresholds for hand and landmark detection. In this study, we set the min_detection_confidence and min_tracking_confidence to the default value of 0.5.
Video extraction and pre-processing
The GMH framework [20] was used to extract the finger-tapping videos. This framework provides high-fidelity 3D hand landmarks. The pre-processing stage involved using the OpenCV library (version 4.6.0). We selected 3D data points corresponding to the thumb and index tips. We calculated the Euclidean distance from these data points, which was used as the finger-tapping amplitude. The amplitude values were standardized by dividing all values by the maximum distance in the corresponding time series for each hand. The standardized values were then normalized from 0 to 100, with 100 representing the maximum amplitude. We created time-series data from the finger-tapping amplitude values obtained from both hands of the participants.
Feature extraction
A peak and valley finder algorithm was used to analyze the tapping signal during the feature extraction process. We employed the find peaks function from Scipy.signal library version 1.7.3 [21], a signal processing package available in Python, to implement this algorithm. A prominence value of 0.3 was set to implement the algorithm. The algorithm identified the peaks (highest points) and valleys (lowest points) in the tapping signal, along with their respective magnitudes and locations. This analysis provided important information about the characteristics and variations of the tapping signal, which could be further used to extract relevant features for subsequent analyses and interpretations.
Finger-tapping (FT) features were derived from the peak amplitudes, which were subsequently used to calculate the features. The tapping signal yielded the following features:
- The tapping Score (TS) represented the total number of finger taps within a 10-s window.
- The tapping difference (TD) quantifies the discrepancy in the number of taps between the right and left hand.
- The mean tapping amplitude (TA) was calculated as the average peak amplitude for each hand.
- The tapping variation (TV) represents the standard deviation of the peak amplitudes of each hand, indicating the variability of the tapping intensity.
- Amplitude decrement 5 (AmpD-5) was calculated as the difference between the peak amplitudes of the first and fifth taps.
- Amplitude decrement 10 (AmpD-10) measures the difference between the peak amplitudes of the first and tenth taps.
An example plot is shown in Fig 1 To visually depict the finger-tapping movements of both hands and their corresponding timestamps. This plot illustrates the finger-tapping activity during the recorded session.
Model development, optimization, and comparison
We examined a diverse range of machine learning classifiers to identify the one with the highest categorical accuracy for clinical diagnosis of Parkinson’s disease, with an MDS-UPDRS part III finger tapping score of at least 1. The Python scikit-learn library (version 1.1.2) [22] was used to train and test the machine learning models using the Jupyter Notebook environment. The study involved the utilization and comparison of six machine learning algorithms: logistic regression (LR), support vector machine (SVM), random forest (RF), decision tree (DT), light gradient boosting model (LGBM), and K-nearest neighborhood (KNN).
We used a nested cross-validation framework to ensure a robust evaluation and prevent information leakage. This approach involved two levels of cross-validation: an outer loop to assess model performance and an inner loop for hyperparameter tuning.
Outer cross-validation.
The outer loop employed a stratified K-fold cross-validation with 5 splits to ensure that the class distribution of the target variable was preserved across the training and test sets. For each fold, the dataset was split into distinct training and test sets, and the final evaluation metrics were derived from the model performance on the unseen test data.
Inner cross-validation.
Within each training set generated by the outer loop, a 3-fold stratified K-fold cross-validation was conducted to perform hyperparameter tuning. Various methodologies were explored for tuning the hyperparameters of the six machine learning models using GridSearchCV to predict and analyze the presence of bradykinesia. The hyperparameter configuration that yielded the highest accuracy in the inner loop was selected for training the model on the corresponding outer training set. A similar approach was applied for tuning the hyperparameters of other machine learning algorithms.
Model evaluation and metrics.
The best model from the inner loop for each outer fold was used to predict the probabilities and class labels on the outer test set. Performance was evaluated using multiple metrics, including accuracy, precision, recall, F1-score, and area under the receiver operating characteristic curve (AUC). Additionally, the AUC was computed to train and test the datasets to assess potential overfitting.
Confidence interval estimation.
We calculated the mean, standard deviation, and 95% confidence intervals (CIs) for each evaluation metric. The CIs were derived using a t-distribution with degrees of freedom equal to the number of outer folds minus one. These CIs provide insights into the reliability of the reported performance metrics by considering the variability across folds.
This nested cross-validation framework [23], applied across all six machine learning algorithms, ensures that the reported model performance was unbiased, reflective of real-world application, and provides robust insights into the effectiveness of different classifiers.
Statistical analysis
Numbers and percentages are presented for categorical variables, whereas medians and interquartile ranges (IQRs) were reported for continuous variables that were not normally distributed. Overall agreement across raters was calculated using Krippendorff’s alpha [24], which is a helpful measure of overall agreement. Krippendorff’s alpha and bootstrap 95% confidence intervals (1,000 iterations, sampling subject with replacement) were calculated using scripts from Zapf A. et al. [25] in R version 4.0.2 (R Foundation for Statistical Computing, Vienna, Austria). The performance of machine learning in predicting the presence of bradykinesia was compared to that of a neurologist using receiver operating characteristics (ROC) analysis, and the area under the curve (AUC) was estimated. The interpretation of AUC was based on the scoring system as follows: 1.0 indicated a perfect test, 0.99–0.90 indicated an excellent test, 0.89–0.8 indicated a good test, 0.79–0.70 indicated a fair test, 0.69–0.51 indicated a poor test and 0.50 or lower indicated a failure [26]. All statistical tests were two-tailed, and P < 0.05 was considered significant.
Results
The study comprised 100 participants:53 patients and 47 healthy control subjects. The baseline characteristics of the patients and healthy controls are presented in Table 1. The two groups exhibited comparable mean ages, with the patients and controls having an average age of 60.6 ± 10.0 years and 59.5 ± 11.7 years, respectively (p = 0.53). However, there was a significant sex discrepancy (p = 0.003). Videos were recorded in both ON and OFF states. Most patients demonstrated Hoehn and Yahr (H&Y) scale scores of < 3 (83.1%). Furthermore, most of the patients exhibited abnormal finger-tapping to a mild degree (33 hands scored 1, and 40 hands scored 2 in the MDS-UPDRS-III sub-score). In contrast, the PD group had significantly lower tapping scores than the normal subjects (right: 17.9 (7.8)/ left: 17.9 (5.6) vs. right: 24.6 (7.3)/ left: 24.6 (7.2), p < 0.001). The differences in tapping scores between the left and right hands were significantly greater in patients with PD than in healthy controls (2.8 (5.0) vs 0.4 (0.7), p = 0.001). Moreover, the tapping amplitude variation and all amplitude decremental parameters in PD subjects differed significantly from those of the standard controls. No parameters were overtly affected by age, sex, or handedness.
Table 2 presents the predictive performance metrics of various machine learning models applied to the testing dataset. The evaluated models include LR, SVM, RF, DT, LGBM, and KNN. The performance metrics considered were accuracy, precision, recall, F1-score, area under the curve (AUC), and training AUC.
Among the models, the SVM achieved the highest accuracy (0.84 ± 0.07, 95% CI: 0.75–0.93), along with substantial precision (0.89 ± 0.10) and recall (0.81 ± 0.13), leading to an F1-score of 0.84 ± 0.08. Additionally, SVM attained the highest AUC in the testing dataset (0.91 ± 0.05), suggesting superior discriminatory ability. The RF and LR models demonstrated comparable performance, with accuracy values of 0.79 ± 0.08 and 0.79 ± 0.10, respectively, and AUC values of 0.87 ± 0.07. The DT model showed the lowest accuracy (0.74 ± 0.09) and AUC (0.73 ± 0.09), indicating lower generalization capability. LGBM and KNN demonstrated moderate performance with accuracy values of 0.76 ± 0.07 and 0.76 ± 0.06, respectively. Notably, the RF model exhibited overfitting, as indicated by a training AUC of 1.00 compared with a lower testing AUC of 0.87. The results suggest that the SVM model provides the best balance between accuracy, precision, recall, and AUC, making it the most suitable choice for this dataset.
Fig 2 shows the ROC curves for the multiple models using nested cross-validation. SVM achieved the highest AUC (0.91 ± 0.05), followed by LR and RF (0.87 ± 0.07), respectively. LGBM and KNN had moderate AUCs of 0.82 ± 0.08 and 0.80 ± 0.07, respectively, whereas the decision tree performed the worst (AUC = 0.73 ± 0.09). The dashed line represents a random chance (AUC = 0.50). Overall, SVM demonstrated the best discrimination ability among the models.
We employed the SHapley Additive exPlanations (SHAP) method [27] to identify and rank the contributing features in the model’s prediction. Shapley values (Fig 3) were used to estimate the importance of the features in the support vector machine model. Each dot represents the influence of tapping features on the prediction. Dots to the left decrease the chance of detecting bradykinesia, whereas dots to the right increase it. The features are listed in descending order of importance, with the tapping difference being the most important, followed by the tapping score (right hand) and tapping amplitude mean (right hand).
Discussion
The study introduced a novel approach utilizing computer vision techniques with MediaPipe hand to screen for Parkinson’s disease by analyzing bradykinesia through finger-tapping tests using smartphone video recordings. Uniquely, our study considered both hands simultaneously, unlike previous studies that often focused on single-hand measurements. This dual-hand assessment enhances the detection of asymmetries and subtle bradykinesia, potentially improving the sensitivity of early Parkinson’s disease detection. Notably, our machine learning model, developed from this data, has demonstrated strong predictability regarding bradykinesia, aiming to enhance diagnostic capabilities significantly. The MDS-UPDRS rating employs several crucial features, such as amplitude, speed, and regularity (including hesitations, halts, and decreasing amplitude). Our study assessed speed (tapping score), regularity (amplitude variation), and amplitude decrement. Previous research utilizing magnetic sensors to extract finger motion properties has identified motion amplitude and the variation of width and frequency as the most significant features [28]. William et al. [12] The peak-to-peak standard deviation, rhythm, and average amplitude were the most predictive measures. Additionally, a study analyzing finger tapping in patients with idiopathic rapid eye movement sleep behavior (RBD) found evidence of amplitude decrement in repetitive movements, a prodromal stage preceding the development of parkinsonism [29]. A study using infrared-emitting diodes found that finger-tapping trials showed slowness with a progressive reduction in amplitude and speed in patients with PD compared with those with PSP and healthy control [30]. This corresponds to our results for the tapping score, average amplitude, amplitude variation, and amplitude decrement. In addition to other studies, PD is an asymmetrical condition; thus, it would be appropriate to assess the performance of the finger-tapping test on both sides. We found significant differences in tapping scores between the two hands in patients with PD. The difference in the number of taps between the right and left hands exhibited the highest SHAP value, underscoring its strong influence on the model’s prediction. This metric likely captures asymmetry in finger-tapping performance, a hallmark motor symptom of Parkinson’s disease, particularly in the early stages. The second most influential feature, Tapping Score (right), reflects finger-tapping performance, with reduced speed and irregularity being well-established indicators of bradykinesia. These findings highlight the importance of assessing the magnitude and variability of motor actions across the hands. Using computer vision to analyze subtle movement patterns enhances the sensitivity of these measurements, making it particularly valuable for detecting early bradykinesia, where traditional clinical assessments may lack precision. Our results align with those of Yang et al., who used tapping rate and amplitude variation to train a model for assessing the severity of bradykinesia. One key difference is that Yang et al. included freeze time as an additional variable [31]. They then used computer vision to automatically classify bradykinesia through an analysis of finger-tapping performance (viz., the MDS-UPDRS score for multiclass prediction): their deep neural network (DNN) model achieved an average F1-score of 84–88%. Additionally, a convolutional neural network (CNN) model developed by Li et al. achieved an accuracy of 79.7% [31], aiming to classify MDS-UPDRS scores automatically. However, our study focused on screening for Parkinson’s disease by detecting bradykinesia and evaluating simultaneous finger tapping, assessing tapping performance in both hands and analyzing inter-hand differences. The study by Wong et al. also included binary classification of the presence of bradykinesia (UPDRS >1) using taping frequency, amplitude, and tap-to-tap variability to predict bradykinesia on the one hand, with an estimated accuracy of 79% [12]. Amprimo et al. utilized the MediaPipe Hand framework and specialized RGB-depth camera (Azure Kinect) to detect bradykinesia, achieving an accuracy of 94.4% and an F1 score of 98.4% in a leave-one-subject-out validation [32]. However, this approach requires a specialized RGB-depth camera for video recording purposes.
There are various publicly available open-source models for pose estimation, such as Openpose [33], DeepLabCut [13,34], AlphaPose [35], YOLOv7 [34], and Mediapipe. Mediapipe can also be utilized on standard hardware, including CPU, GPU, or TPU, and its framework is suitable for a wide range of platforms, such as iOS, Android, desktop, Edge, cloud, web, and IoT. Williams et al. found that DeepLabCut correlated well with a clinical rating for bradykinesia in patients with PD when the video was set at 60 fps [14]. However, the default video frame rates on smartphones are 30 fps. This study used a standard 30 fps, effectively tracking fingertip movement without motion blur. Morinan et al. utilized OpenPose to evaluate five items from the MDS-UPDRS to assess bradykinesia. Their investigation revealed that a binary model (categorizing bradykinesia as low-mild versus moderate-high) performed better in predicting than the five levels of severity: normal, slight, mild, moderate, and severe [36].
The principal benefit of our technique is its ease of deployment, low cost, and contactless nature. Moreover, it can utilize an ordinary smartphone without requiring any specialized settings, thus potentially assisting in early clinical diagnosis. Additionally, preparing video recordings is simple and does not require specific conditions, such as lighting, background, or distance. Our method employs a short recording time that minimizes fatigue. Furthermore, our PD patients did not require medication discontinuation before performing finger tapping, which is more convenient and relevant to clinical practice. In this study, we analyzed the data on a per-patient basis, in contrast to previous studies that analyzed data on a per-hand basis. We have also developed a web-based platform, pd.kku.ac.th/home, which allows users to screen for Parkinson’s disease by uploading a video. The website then processes the video, generates amplitude and time graphs, and provides the results for bradykinesia detection.
The current study had several limitations. First, identifying bradykinesia in patients with subtle symptoms (indicated by finger-tapping subscore of 1) can be challenging. Second, we did not quantitatively measure the absolute distance between the fingertip positions. This omission is noteworthy because the rotational movements of the thumb and finger in the horizontal plane can lead to inaccurate amplitude measurements. Nevertheless, it is worth noting that rotational movements are generally minimal during finger tapping, and we accounted for this by calibrating the maximum tapping distance to 100%. Third, our methodology exclusively focuses on detecting bradykinesia symptoms and may not detect other symptoms of Parkinson’s disease. Fourth, our method may encounter difficulties in cases where patients exhibit severe tremors, leading to motion blur that may interfere with accurate measurements. Unfortunately, our study did not include patients with severe tremors. The GMH framework can detect finger movement if the tremor amplitude is less than 10 cm. Finally, although we did not use a tripod for camera setup while recording the videos, most of the videos from patients were able to capture finger movements. Videos with poor recording quality, in which features could not be detected, were excluded. These limitations highlight areas for future research to improve our method and make it more accurate and useful for a wider range of Parkinson’s symptoms and patients with Parkinson’s disease.
Conclusions
In summary, in the present study, we demonstrated a computer vision method to detect bradykinesia using the tapping features from the finger-tapping method by simultaneously tapping both hands. The support vector machine algorithm yielded the highest average accuracy (0.84).
Acknowledgments
The North-Eastern Stroke Research Group, Khon Kaen University, Thailand, supported this study.
We thank Mr. Bryan Roderick Hamman, a native English speaker and member of the Council of Science Editors, for assistance with the English-language presentation of the manuscript under the Khon Kaen University’s Publication Clinic, Thailand.
References
- 1. Kalia LV, Lang AE. Parkinson’s disease. Lancet. 2015;386(9996):896–912. pmid:25904081
- 2. Postuma RB, Berg D, Stern M, Poewe W, Olanow CW, Oertel W, et al. MDS clinical diagnostic criteria for Parkinson’s disease. Mov Disord. 2015;30(12):1591–601. pmid:26474316
- 3. Goetz CG, Tilley BC, Shaftman SR, Stebbins GT, Fahn S, Martinez-Martin P, et al. Movement Disorder Society-sponsored revision of the Unified Parkinson’s Disease Rating Scale (MDS-UPDRS): scale presentation and clinimetric testing results. Mov Disord. 2008;23(15):2129–70. pmid:19025984
- 4. Homann CN, Suppan K, Wenzel K, Giovannoni G, Ivanic G, Horner S, et al. The bradykinesia akinesia incoordination test (BRAIN TEST), an objective and user-friendly means to evaluate patients with Parkinsonism. Mov Disord. 2000;15(4):641–7.
- 5. Papapetropoulos S, Katzen HL, Scanlon BK, Guevara A, Singer C, Levin BE. Objective quantification of neuromotor symptoms in Parkinson’s disease: implementation of a portable, computerized measurement tool. Parkinsons Dis. 2010;2010:760196. pmid:20976095
- 6. Bronte-Stewart HM, Ding L, Alexander C, Zhou Y, Moore GP. Quantitative digitography (QDG): a sensitive measure of digital motor control in idiopathic Parkinson’s disease. Mov Disord. 2000;15(1):36–47.
- 7. Taylor Tavares AL, Jefferis GSXE, Koop M, Hill BC, Hastie T, Heit G, et al. Quantitative measurements of alternating finger tapping in Parkinson’s disease correlate with UPDRS motor disability and reveal the improvement in fine motor control from medication and deep brain stimulation. Mov Disord. 2005;20(10):1286–98. pmid:16001401
- 8. Giancardo L, Sánchez-Ferro A, Arroyo-Gallego T, Butterworth I, Mendoza CS, Montero P, et al. Computer keyboard interaction as an indicator of early Parkinson’s disease. Sci Rep. 2016;6:34468. pmid:27703257
- 9. Lee CY, Kang SJ, Hong S-K, Ma H-I, Lee U, Kim YJ. A validation study of a smartphone-based finger tapping application for quantitative assessment of Bradykinesia in Parkinson’s disease. PLoS ONE. 2016;11(7):e0158852.
- 10. Kim J-W, Lee J-H, Kwon Y, Kim C-S, Eom G-M, Koh S-B, et al. Quantification of bradykinesia during clinical finger taps using a gyrosensor in patients with Parkinson’s disease. Med Biol Eng Comput. 2011;49(3):365–71. pmid:21052856
- 11. Khan T, Nyholm D, Westin J, Dougherty M. A computer vision framework for finger-tapping evaluation in Parkinson’s disease. Artif Intell Med. 2014;60(1):27–40.
- 12. Williams S, Relton SD, Fang H, Alty J, Qahwaji R, Graham CD, et al. Supervised classification of bradykinesia in Parkinson’s disease from smartphone videos. Artif Intell Med. 2020;110:101966. pmid:33250146
- 13. Nath T, Mathis A, Chen AC, Patel A, Bethge M, Mathis MW. Using DeepLabCut for 3D markerless pose estimation across species and behaviors. Nat Protoc. 2019;14(7):2152–76. pmid:31227823
- 14. Williams S, Zhao Z, Hafeez A, Wong DC, Relton SD, Fang H, et al. The discerning eye of computer vision: can it measure Parkinson’s finger tap bradykinesia? J Neurol Sci. 2020;416:117003. pmid:32645513
- 15.
Lugaresi C, Tang J, Nash H, McClanahan C, Uboweja E, Hays M. MediaPipe: framework for building perception pipelines. 2019. https://arxiv.org/abs/1906.08172
- 16.
google-ai-edge. MediaPipe Hands. [Accessed 2025 February 14]. https://github.com/google-ai-edge/mediapipe/blob/master/docs/solutions/hands.md
- 17. Wang X, Garg S, Tran SN, Bai Q, Alty J. Hand tremor detection in videos with cluttered background using neural network based approaches. Health Inf Sci Syst. 2021;9(1):30. pmid:34276971
- 18. Güney G, Jansen TS, Dill S, Schulz JB, Dafotakis M, Hoog Antink C, et al. Video-Based hand movement analysis of Parkinson patients before and after medication using high-frame-rate videos and MediaPipe. Sensors. 2022;22(20):7992.
- 19. Amprimo G, Masi G, Pettiti G, Olmo G, Priano L, Ferraris C. Hand tracking for clinical applications: Validation of the Google MediaPipe Hand (GMH) and the depth-enhanced GMH-D frameworks. Biomed Signal Process Control. 2024;96:106508.
- 20.
Zhang F, Bazarevsky V, Vakunov A, Tkachenka A, Sung G, Chang C. MediaPipe Hands: On-device Real-time Hand Tracking. 2020. http://arxiv.org/abs/2006.10214
- 21. Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, et al. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat Methods. 2020;17(3):261–72. pmid:32015543
- 22. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O. Scikit-learn: machine learning in Python. J Machine Learn Res. 2011;12(85):2825–30.
- 23. Vabalas A, Gowen E, Poliakoff E, Casson AJ. Machine learning algorithm validation with a limited sample size. PLoS One. 2019;14(11):e0224365. pmid:31697686
- 24. Krippendorff K. Estimating the reliability, systematic error and random error of interval data. Educ Psychol Measure. 1970;30(1):61–70.
- 25. Zapf A, Castell S, Morawietz L, Karch A. Measuring inter-rater reliability for nominal data - which coefficients and confidence intervals are appropriate? BMC Med Res Methodol. 2016;16:93. pmid:27495131
- 26. Hanley JA, McNeil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology. 1982;143(1):29–36. pmid:7063747
- 27. Lundberg SM, Lee SI. A unified approach to interpreting model predictions. Adv Neural Inform Process Syst. 2017;30.
- 28. Sano Y, Kandori A, Shima K, Yamaguchi Y, Tsuji T, Noda M, et al. Quantifying Parkinson’s disease finger-tapping severity by extracting and synthesizing finger motion properties. Med Biol Eng Comput. 2016;54(6):953–65. pmid:27032933
- 29. Krupička R, Krýže P, Neťuková S, Duspivová T, Klempíř O, Szabó Z, et al. Instrumental analysis of finger tapping reveals a novel early biomarker of parkinsonism in idiopathic rapid eye movement sleep behaviour disorder. Sleep Med. 2020;75:45–9.
- 30. Ling H, Massey LA, Lees AJ, Brown P, Day BL. Hypokinesia without decrement distinguishes progressive supranuclear palsy from Parkinson’s disease. Brain. 2012;135(4):1141–53.
- 31. Yang N, Liu D-F, Liu T, Han T, Zhang P, Xu X, et al. Automatic detection pipeline for accessing the motor severity of Parkinson’s disease in finger tapping and postural stability. IEEE Access. 2022;10:66961–73.
- 32.
Amprimo G, Rechichi I, Ferraris C, Olmo G. Objective assessment of the finger tapping task in Parkinson’s disease and control subjects using Azure Kinect and machine learning. In: 2023 IEEE 36th International Symposium on Computer-Based Medical Systems (CBMS). 2023.
- 33.
Cao Z, Simon T, Wei SE, Sheikh Y. Realtime multi-person 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017.
- 34.
Wang C, Bochkovskiy A, Liao HY. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. 2022. http://arxiv.org/abs/2207.02696
- 35. Fang HS, Li J, Tang H, Xu C, Zhu H, Xiu Y. Alphapose: Whole-body regional multi-person pose estimation and tracking in real-time. IEEE Transact Pattern Analysis Machine Intell. 2022.
- 36. Morinan G, Dushin Y, Sarapata G, Rupprechter S, Peng Y, Girges C, et al. Computer vision quantification of whole-body Parkinsonian bradykinesia using a large multi-site population. NPJ Parkinsons Dis. 2023;9(1):10. pmid:36707523