Realtime Motion Assessment For Rehabilitation Exercises: Integration Of Kinematic Modeling With Fuzzy Inference


This article describes a novel approach to realtime motion assessment for rehabilitation exercises based on the integration of comprehensive kinematic modeling with fuzzy inference. To facilitate the assessment of all important aspects of a rehabilitation exercise, a kinematic model is developed to capture the essential requirements for static poses, dynamic movements, as well as the invariance that must be observed during an exercise. The kinematic model is expressed in terms of a set of kinematic rules. During the actual execution of a rehabilitation exercise, the similarity between the measured motion data and the model is computed in terms of their distances, which are then used as inputs to a fuzzy interference system to derive the overall quality of the execution. The integrated approach provides both a detailed categorical assessment of the overall execution of the exercise and the degree of adherence to individual kinematic rules.

If the inline PDF is not rendering correctly, you can download the PDF file here.

  • [1] W. Zhao, H. Feng, R. Lun, D. D. Espy, and M. Reinthal, “A kinect-based rehabilitation exercise monitoring and guidance systems,” in Proceedings of the 5th IEEE International Conference on Software Engineering and Service Science. IEEE, 2014, pp. 762–765.

  • [2] C.-J. Su, “Personal rehabilitation exercise assistant with kinect and dynamic time warping,” International Journal of Information and Education Technology, pp. 448–454, 2013.

  • [3] L. Zhang, J.-C. Hsieh, and J. Wang, “A kinectbased golf swing classification system using hmm and neuro-fuzzy,” in Computer Science and Information Processing (CSIP), 2012 International Conference on, 2012, pp. 1163–1166.

  • [4] E. Velloso, A. Bulling, and H. Gellersen, “Motionma: motion modelling and analysis by demonstration,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 2013, pp. 1309–1318.

  • [5] W. Zhao, R. Lun, D. D. Espy, and M. A. Reinthal, “Rule based realtime motion assessment for rehabilitation exercises,” in Proceedings of the IEEE Symposium on Computational Intelligence in Healthcare and e-Health, December 2014, pp. 133–140.

  • [6] Y. Visell and J. Cooperstock, “Enabling gestural interaction by means of tracking dynamical systems models and assistive feedback,” in Systems, Man and Cybernetics, 2007. ISIC. IEEE International Conference on. IEEE, 2007, pp. 3373–3378.

  • [7] P. Hong, M. Turk, and T. S. Huang, “Gesture modeling and recognition using finite state machines,” in Automatic Face and Gesture Recognition, 2000. Proceedings. Fourth IEEE International Conference on. IEEE, 2000, pp. 410–415.

  • [8] P. Turaga, R. Chellappa, V. S. Subrahmanian, and O. Udrea, “Machine recognition of human activities: A survey,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 18, no. 11, pp. 1473–1488, 2008.

  • [9] S. Nomm and K. Buhhalko, “Monitoring of the human motor functions rehabilitation by neural networks based system with kinect sensor,” in Analysis, Design, and Evaluation of Human-Machine Systems, vol. 12, no. 1, 2013, pp. 249–253.

  • [10] R. Poppe, “A survey on vision-based human action recognition,” Image and vision computing, vol. 28, no. 6, pp. 976–990, 2010.

  • [11] J.-S. Lin and D. Kulic, “Online segmentation of human motion for automated rehabilitation exercise analysis,” Neural Systems and Rehabilitation Engineering, IEEE Transactions on, vol. 22, no. 1, pp. 168–180, 2014.

  • [12] S. Schaal, A. Ijspeert, and A. Billard, “Computational approaches to motor learning by imitation,” Philosophical Transactions of the Royal Society of London. Series B: Biological Sciences, vol. 358, no. 1431, pp. 537–547, 2003.

  • [13] N. Gordon, B. Ristic, and S. Arulampalam, “Beyond the kalman filter: Particle filters for tracking applications,” Artech House, London, 2004.

  • [14] R. A. Clark, Y.-H. Pua, K. Fortin, C. Ritchie, K. E. Webster, L. Denehy, and A. L. Bryant, “Validity of the microsoft kinect for assessment of postural control,” Gait and posture, vol. 36, no. 3, pp. 372–377, 2012.

  • [15] R. A. Clark, Y.-H. Pua, A. L. Bryant, and M. A. Hunt, “Validity of the microsoft kinect for providing lateral trunk lean feedback during gait retraining,” Gait & posture, vol. 38, no. 4, pp. 1064–1066, 2013.

  • [16] A. Bo, M. Hayashibe, P. Poignet et al., “Joint angle estimation in rehabilitation with inertial sensors and its integration with kinect,” in EMBC’11: 33rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2011, pp. 3479–3483.

  • [17] E. Akdoğan, E. Taçgın, and M. A. Adli, “Knee rehabilitation using an intelligent robotic system,” Journal of Intelligent Manufacturing, vol. 20, no. 2, pp. 195–202, 2009.

  • [18] Q. Wang, P. Turaga, G. Coleman, and T. Ingalls, “Somatech: an exploratory interface for altering movement habits,” in CHI’14 Extended Abstracts on Human Factors in Computing Systems. ACM, 2014, pp. 1765–1770.

  • [19] B. C. Bedregal, A. C. Costa, and G. P. Dimuro, “Fuzzy rule-based hand gesture recognition,” in Artificial Intelligence in Theory and Practice. Springer, 2006, pp. 285–294.

  • [20] T. Hachaj and M. R. Ogiela, “Rule-based approach to recognizing human body poses and gestures in real time,” Multimedia Systems, vol. 20, no. 1, pp. 81–99, 2014.

  • [21] W. Zhao, D. D. Espy, M. Reinthal, and H. Feng, “A feasibility study of using a single kinect sensor for rehabilitation exercises monitoring: A rule based approach,” in Computational Intelligence in Healthcare and e-health (CICARE), 2014 IEEE Symposium on. IEEE, 2014, pp. 1–8.


Journal + Issues