Generic placeholder image

Current Medical Imaging

Editor-in-Chief

ISSN (Print): 1573-4056
ISSN (Online): 1875-6603

General Research Article

Ultrasonic Echolocation Device for Assisting the Visually Impaired

Author(s): Ben Mick, Nathan Reddmann, Rayyan Manwar and Kamran Avanaki*

Volume 16, Issue 5, 2020

Page: [601 - 610] Pages: 10

DOI: 10.2174/1573405615666190423141647

Price: $65

Abstract

Background: Echolocation is a technique whereby the location of objects is determined via reflected sound. Currently, some visually impaired individuals use a form of echolocation to locate objects and to orient themselves. However, this method takes years of practice to accurately utilize.

Aims: This paper presents the development of a sensory substitution device for visually impaired users, which gauged distances and the placement of objects.

Methods: Using ultrasonic technology, the device employed a method of echolocation to increase the user's independence and mobility. The main components of this device are an ultrasound transceiver and a miniaturized Arduino board. Through research and prototyping, this technology was integrated into a biomedical application in a watch form factor which provides feedback to the user regarding the measured distance by the ultrasonic transducer.

Results: The output of this process is a tactile feedback that varies in intensity proportional to the distance of the detected object. We tested the device in different scenarios including different distances from a different material. The difference between the device reading and the actual distance, from 0 to 400 cm was statistically insignificant.

Conclusion: It is believed this device will boost the confidence of the user in navigation.

Keywords: Echolocation, haptic feedback, ultrasonic transducer, visually impaired, blind, navigation.

Graphical Abstract

[1]
World Health Organization. Global data on visual impairments 2010, 2012. Google Scholar 2014.
[2]
Ballemans J. Kempen, Gertrudis IJM, and Zijlstra, GA Rixt, Orientation and mobility training for partially-sighted older adults using an identification cane: a systematic review. Clin Rehabil 2011; 25(10): 880-91.
[http://dx.doi.org/10.1177/0269215511404931]
[3]
Brault MW. Americans with disabilities: 2010. In: US Department of Commerce, Economics and Statistics Administration. Washington, DC: US Census Bureau 2012.
[4]
Galatas G, McMurrough C, Mariottini GL, et al. eyeDog: an assistive-guide robot for the visually impaired. In: 4th International Conference on Pervasive Technologies Related to Assistive Environments. Crete, Greece, New York: ACM 2011; pp. 1-8.
[http://dx.doi.org/10.1145/2141622.2141691]
[5]
Mahmoodkalayeh S, Jooya Ali Hariri, et al. low temperature-mediated enhancement of photoacoustic imaging depth. Sci Rep 2018; 8(1): 4873.
[http://dx.doi.org/10.1038/s41598-018-22898-2]
[6]
Manwar R, Hosseinzadeh M, Hariri A, et al. Photoacoustic Signal enhancement: towards utilization of low energy laser diodes in real-time photoacoustic imaging. Sensors (Basel) 2018; 18(10): 3498.
[http://dx.doi.org/10.3390/s18103498]
[7]
Nasiriavanaki M, Xia J, Wan H, et al. High-resolution photoacoustic tomography of resting-state functional connectivity in the mouse brain. Proc Natl Acad Sci USA 2014; 111(1): 21-6.
[http://dx.doi.org/10.1073/pnas.1311868111]
[8]
Zafar M, Manwar R, Kratkiewicz K, et al. Development of low-cost fast photoacoustic computed tomography: System characterization and phantom study. Appl Sci (Basel) 2019; 9(3): 374.
[http://dx.doi.org/10.3390/app9030374]
[9]
Avanaki MRN, Cernat R, Tadrous PJ, et al. Spatial compounding algorithm for speckle reduction of dynamic focus OCT images. IEEE Photonics Technol Lett 2013; 25(15): 1439-42.
[http://dx.doi.org/10.1109/LPT.2013.2266660]
[10]
Avanaki MRN, Hojjatoleslami A, Sira M, et al. Investigation of basal cell carcinoma using dynamic focus optical coherence tomography. Appl Opt 2013; 52(10): 2116-24.
[http://dx.doi.org/10.1364/AO.52.002116]
[11]
Avanki MRN, Hojjat A, Podoleanu AG. Investigation of computer-based skin cancer detection using optical coherence tomography. J Mod Opt 2009; 56(13): 1536-44.
[http://dx.doi.org/10.1080/09500340902990007]
[12]
Hojjatoleslami A, Avanaki MRN. OCT skin image enhancement through attenuation compensation. Appl Opt 2012; 51(21): 4927-35.
[http://dx.doi.org/10.1364/AO.51.004927]
[13]
Turani Z, Fatemizadeh E, Blumetti T, et al. Optical radiomic signatures derived from optical coherence tomography images to improve identification of melanoma. Canc Res 2019; pp. 2791-8.
[14]
Griffin DR. Listening in the dark: the acoustic orientation of bats and men. Ithaca: Cornell University Press 1958.
[15]
Juurmaa J. Analysis of orientation ability and its significance for the rehabilitation of the blind. Scand J Rehabil Med 1969; 1(2): 80-4.
[16]
Schenkman BN, Nilsson ME. Human echolocation: Blind and sighted persons’ ability to detect sounds recorded in the presence of a reflecting object. Perception 2010; 39(4): 483-501.
[http://dx.doi.org/10.1068/p6473]
[17]
Kish D. Echolocation: How humans can “see” without Sight. Part Three 2008.
[18]
Sohl-Dickstein J, Teng S, Gaub BM, et al. A device for human ultrasonic echolocation. IEEE Trans Biomed Eng 2015; 62(6): 1526-34.
[http://dx.doi.org/10.1109/TBME.2015.2393371]
[19]
Hikiya M. An introduction to ultrasonic sensors for vehicle parking. New Electronics 2010. Available from:. http://www. newelectronics.co.uk/electronics-technology/an-introduction-to-ultrasonic-sensors-for-vehicle-parking/24966/
[20]
Microdrives P. What is haptic / tactile feedback? 2015. Available from:. https://www.precisionmicrodrives.com/haptic-feedback/introduction-to-haptic-feedback/
[21]
MaxBotic Inc. High resolution, precision, low voltage ultrasonic range finder. HRLV-MaxSonar®- EZ™ Series 2014. Available from:. https://www.maxbotix.com/ documents/HRLV-MaxSonar-EZ_Datasheet.pdf
[22]
Liuyuan C, Zeynep B. SensCap is a device that guides the visually impaired around obstacles. Cornell University 2011.

Rights & Permissions Print Cite
© 2024 Bentham Science Publishers | Privacy Policy