Eyeball Movement Detection To Control Smart Wheelchair Using Eye Aspect Ratio (EAR)

Gusti Pangestu

Abstract


Many developed technology's with an aim of helping the disabilities. One of them is a wheelchair. It is the most common stuff that used for helping disabilities as a tool for mobilization. There are two types of wheelchair. The first is the manual wheelchair, operated by hand. The second is an electrical wheelchair, that operated by joystick or other electric device. This research proposed a mechanism to control the wheelchair by using an eye movement. It could be used especially for people with multiple disabilities (hand and foot defects), so they can take an advantage of their eyeballs as a tool to control wheelchair movement. There are five options for controlling the wheelchair (leftward, rightward, upward, downward and center). Leftward, rightward and center used for control direction of smart wheelchair. Furthermore, upward and downward of eye movements used to control the speed of smart wheelchair. Upward command used to increase the speed. Meanwhile, down-ward used to decrease the speed (stop). The proposed method used EAR (Eye Aspect Ratio), which divided into three regions based on sector area, for determining the directions of the eyeball movement. EAR is the value that represents the ratio between the upper eyelid and lower eyelid. The result obtained high accuracy

Keywords


Eye ascpet ratio; eye; control

Full Text:

PDF

References


C. A. Okoro, N. D. Hollis, A. C. Cyrus, and S. Griffin-Blake, “Prevalence of Disabilities and Health Care Access by Disability Status and Type Among Adults — United States, 2016,” MMWR. Morb. Mortal. Wkly. Rep., vol. 67, no. 32, pp. 882–887, 2018, doi: 10.15585/mmwr.mm6732a3.

H. S. Jung, G. Park, Y. S. Kim, and H. S. Jung, “Development and evaluation of one-hand drivable manual wheelchair device for hemiplegic patients,” Appl. Ergon., vol. 48, pp. 11–21, 2015, doi: 10.1016/j.apergo.2014.10.020.

N. Peixoto, H. G. Nik, and H. Charkhkar, “Voice controlled wheelchairs: Fine control by humming,” Comput. Methods Programs Biomed., vol. 112, no. 1, pp. 156–165, 2013, doi: 10.1016/j.cmpb.2013.06.009.

F. Utaminingrum, T. A. Kumiawan, M. A. Fauzi, R. C. Wihandika, and P. P. Adikara, “Adaptive human tracking for smart wheelchair,” 5th Int. Symp. Comput. Bus. Intell. ISCBI 2017, no. August, pp. 10–13, 2017, doi: 10.1109/ISCBI.2017.8053535.

D. Orizaba, “A Hands Gesture System Of Control For An Intelligent Wheelchair.,” no. 852, pp. 68–71, 2007.

R. Hardiansyah, “The Electric Wheelchair Control Using Electromyography Sensor Of Arm Muscle,” Int. Conf. Information, Commun. Technol. Syst., pp. 129–134, 2016.

C. Ishii, S. Murooka, and M. Tajima, “Navigation of an Electric Wheelchair Using Electromyograms, Electrooculograms, and Electroencephalograms,” Int. J. Mech. Eng. Robot. Res., vol. 7, no. 2, pp. 143–149, 2018, doi: 10.18178/ijmerr.7.2.143-149.

F. Utaminingrum, P. P. Adikara, Y. A. Sari, D. Syauqy, and A. G. Hapsani, “Left-right head movement for controlling smart wheelchair by using centroid coordinates distance,” J. Theor. Appl. Inf. Technol., vol. 96, no. 10, pp. 2852–2861, 2018.

F. Utaminingrum, M. A. Fauzi, and Y. A. Sari, “Eye Movement as Navigator for Restricted Disabled Person in Handling Position,” in ICCIS (International Conference on Communication and Information System), 2017, pp. 1–5, doi: 10.475/123.

R. P. Prasetya and F. Utaminingrum, “Triangle similarity approach for detecting eyeball movement,” 5th Int. Symp. Comput. Bus. Intell. ISCBI 2017, pp. 37–40, 2017, doi: 10.1109/ISCBI.2017.8053540.

R. P. Prasetya, F. Utaminingrum, and W. F. Mahmudy, “Real time eyeball movement detection based on region division and midpoint position,” Int. J. Intell. Eng. Syst., vol. 11, no. 3, pp. 149–158, 2018, doi: 10.22266/IJIES2018.0630.16.

J. Cech and T. Soukupova, “Real-Time Eye Blink Detection using Facial Landmarks,” Cent. Mach. Perception, Dep. Cybern. Fac. Electr. Eng. Czech Tech. Univ. Prague, pp. 1–8, 2016.

P. Viola and M. Jones, “Rapid object detection using a boosted cascade of simple features,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2001, vol. 1, pp. I--511, doi: 10.1109/cvpr.2001.990517.

V. Kazemi and J. Sullivan, “One Millisecond Face Alignment with an Assemble of Regression Trees,” Conf. Comput. Vis. Pattern Recognit., 2017, doi: 10.13140/2.1.1212.2243.




DOI: https://doi.org/10.18860/mat.v13i2.12963

Refbacks

  • There are currently no refbacks.




Copyright (c) 2021 Gusti Pangestu

Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.

The journal is indexed by :

Dimensions Sinta CrossRef GoogleScholar
Index Copernicus Moraref Portal Garuda

 

_______________________________________________________________________________________________________________

Editorial Office:
Informatics Engineering Department
Faculty of Science and Technology
Universitas Islam Negeri Maulana Malik Ibrahim Malang
Jalan Gajayana 50 Malang, Jawa Timur, Indonesia 65144
Email: matics@uin-malang.ac.id
_______________________________________________________________________________________________________________

Creative Commons License
This work is licensed under a CC-BY-NC-SA 4.0.
© All rights reserved 2015. MATICS , ISSN : 1978-161X | e-ISSN :  2477-2550