Abstract

With a globally aging population, visual impairment is an increasingly pressing problem for our society. Visual disability drastically reduces quality of life and constitutes a large cost to the health care system. Mobility of the visually impaired is one of the most critical aspects affected by this disability, and yet, it relies on low-tech solutions, such as the white cane. Many avoid solutions entirely. In part, reluctance to use these solutions may be explained by their obtrusiveness, a strong deterrent for the adoption of many new devices. Here, we leverage new advancements in artificial intelligence, sensor systems, and soft electroactive materials toward an electronic travel aid with an obstacle detection and avoidance system for the visually impaired. The travel aid incorporates a stereoscopic camera platform, enabling computer vision, and a wearable haptic device that can stimulate discrete locations on the user’s abdomen to signal the presence of surrounding obstacles. The proposed technology could be integrated into commercial backpacks and support belts, thereby guaranteeing a discreet and unobtrusive solution.

1 Introduction

Visual impairment is a continually growing problem, affecting over 3.2 million Americans who are 40 years and older [1]. This number is expected to double by 2050 [1] due to the aging population [2]. Visual impairment is associated with reduced quality of life [3], along with increased risk of falls and hip fractures [4], depression [5], obesity [6], and death [7]. Such a debilitating condition brings about high costs for society at large. A study on the impact of vision loss in Canada claimed that direct and indirect costs associated with visual impairment summed up to $15.8 billion in 2007 (1.19% of the Country’s gross domestic product), constituting the most expensive disease category in terms of health care expenditures [8].

Affording increased, safe mobility and independence for the visually impaired is critical to improving quality of life and reducing financial burden. Despite the severity of the situation at a global scale, options for enhancing the mobility of the visually impaired remain limited, with little integration of new technologies. To date, the most common options are white canes and guide dogs. Less than 10% [9,10] of the visually impaired adopt these solutions, likely because of perceived social awkwardness, dependence, and anxiety [1113]. These facts speak to the need for low-profile, inconspicuous assistive devices [14].

Most electronic travel aids (ETAs) proposed in the literature present limitations that hinder the widespread application of these systems. Among the proposed solutions, one can find: devices that transmit information through acoustic signals in earphones, thereby occluding air conduction, an important transmission pathway for auditory cues of safe navigation [15]; solutions with hand-held devices that limit users’ capacity to protect themselves in the case of an unexpected fall [16] or devices that implement nonintuitive coding and interfacing with the surrounding environment [17,18]; smart white canes that enhance existing conspicuous solutions and are challenging to wield dynamically [19,20]; and bulky, obtrusive, conspicuous systems that may cause uneasiness in the user [21]. Overall, little to no attention has been paid toward the acceptance of these technologies by the visually impaired. However, technological advancements have the potential to provide alternative, cost-effective solutions for visually impaired immobility. While regenerative medicine could ultimately restore vision [22], research in this area is still in its nascent stages and requires a bevy of therapeutic algorithms and alternatives for a comprehensive approach. Sensory substitution systems that fuse information and communication technology with artificial intelligence could enable localization, obstacle avoidance, and quality of life improvements in the short term [23].

Here, we put forward a proof-of-concept ETA for the visually impaired that uses computer vision for obstacle detection with a novel wearable device—a bookbag that seamlessly incorporates sensors and a modified waist strap fashioned into a belt that provides haptic feedback to the user based on the computer vision output. The belt comprises a two-row, five-column array of piezoelectric-based tactors, which stimulate the abdomen based on the location of identified obstacles. More specifically, the computer vision software partitions the scene into an array of ten rectangular capture fields, resembling the tactor configuration on the torso; if an obstacle is identified in one of these rectangles, the corresponding actuator in the array of tactors starts vibrating. The proposed solution allows for an inconspicuous, highly customizable design, which can be incorporated into commercial backpack systems.

Computer vision and generally deep learning have contributed significant advancements that can be integrated in assistive technologies for the visually impaired [24]. The use of a computer vision system supports the integration of critical features, such as object identification [25] or access to pedestrian signals at intersection crossings [26]. The output from other types of sensors, such as ultrasonic or light detection and ranging devices [27,28], could also be integrated into our system through sensor fusion toward more accurate obstacle identification [25]. Sensor fusion allows for a robust estimate of the distance of obstacles, in spite of variable lighting conditions and harsh weather that are detrimental for computer vision, and also assists in the identification of large obstacles, such as walls and columns [25]. With respect to the wearable device, haptic feedback is preferred over auditory stimuli, since it provides faster information transfer for hazard negotiation, even if at a coarser level of detail [29], and it does not disrupt organic auditory cues from the environment. The choice of torso-based stimulation for a wearable device enables concealment and an egocentric re-display scheme that is intuitive, a spatiotopically preserving mapping of hazards in the surrounding environment. A binaural bone conduction headset may also be wirelessly connected to the system through Bluetooth, enabling an additional communication channel.

The design proposed in this article builds upon our previous belt prototype [30], which demonstrated the feasibility of relaying information through tactile feedback on the abdomen utilizing piezoelectric-based actuators. Compared to the previous device, the new prototype includes a higher spatial resolution with a larger number of tactors, along with improved vibration amplitude. In addition, our previous work did not include any sensing element, which is pursued herein through computer vision techniques.

In what follows, we detail the design of the ETA. First, we describe the wearable haptic device and computer vision system. Then, we illustrate the interface between these components, and we perform a delay analysis. Finally, we highlight potential areas of future research and experimental tests.

2 Wearable Haptic Device

The wearable haptic device consists of a belt with ten discrete tactors, capable of providing haptic feedback to the user. Similar to Ref. [30], actuation of the tactors is based on macro-fiber composites (MFCs). Developed by NASA in 1999 [31], these materials are composed of piezoceramic wafers with interdigitated electrodes on a polyimide film, along with structural epoxy layers. We utilize P1-type MFC actuators from Smart Material (Sarasota, FL). When a voltage on the order of kilovolts is applied across the electrodes, these MFCs elongate through the so-called d33 effect [32]. In other words, these materials exploit the nonzero coupling between the components of stress tensor and electric field in the poling direction [32].

The macroscopic deformations of MFCs are ultimately determined by boundary conditions. In our previous work [30], we utilized the nonlinear coupling between expansion and bending in a postbuckling configuration [33,34]. While this design demonstrated the feasibility of haptic stimulation of the abdomen to convey information to the user, the amplitude of the out-of-plane vibrations of the actuators did not enable robust discrimination of the vibration [30]. To improve the performance of the actuator in terms of vibration amplitude and blocking force, we opt for an alternative solution by bonding the MFC to a 54 × 20 × 0.25 (length × width × thickness) mm aluminum plate using epoxy. In this configuration [35,36], the asymmetric elongation with respect to the neutral axis of the overall structure elicits out-of-plane bending.

To protect the actuator and prevent electric short circuits, we encapsulate the aluminum-backed MFC in a 3D-printed case made of polylactic acid (PLA), as shown in Fig. 1(a). Specifically, the actuator is fixed in a double cantilever configuration inside the case. Four 0.77 mm thick spacers are mounted between the actuator and the case to grant space for vibration. A hollow cylindrical protrusion is positioned anteriorly at the center of the actuator to transmit vibrations to the user’s skin. The bottom part of the cylinder is 3D printed in thermoplastic polyurethane (TPU) and is connected to the case through two slender, flexible TPU rods. The top part of the cylinder, made of 3D-printed PLA, consists of a fixed cavity and a removable cap fitted through a threaded mechanism. The diameter of the area in contact with the abdomen is 14 mm. The internal cavity of the cylinder can accommodate additional masses to optimize the resonance frequency of the actuators.

Fig. 1
(a) Tactor with open cap. The MFC is visible below the central cylinder, which contains a tungsten alloy 8 g disc. The wires coming out of the case are the MFC terminals. A detailed schematic of tactors can be found in Ref. [37]. (b) Assembled belt within a sleeve. The wires connecting the tactors to the electronics in the backpack are placed within the wire comb spiral visible on the right.
Fig. 1
(a) Tactor with open cap. The MFC is visible below the central cylinder, which contains a tungsten alloy 8 g disc. The wires coming out of the case are the MFC terminals. A detailed schematic of tactors can be found in Ref. [37]. (b) Assembled belt within a sleeve. The wires connecting the tactors to the electronics in the backpack are placed within the wire comb spiral visible on the right.
Close modal

Preliminary mechanical tests show that adding a mass of 12 g reduces the resonance frequency to about 175 Hz [37], which is within the range of frequencies that skin mechanoreceptors are most sensitive to [29,38]. These tests [37] also suggest that the peak-to-peak displacement of the tactors can be increased by a tenfold factor throughout the frequency range of interest and more than a twentyfold factor at resonance with respect to our previous design [30], promising a significant improvement in vibration discrimination. The resulting amplitude would be at least twice the sensory threshold over the entire actuation frequency range [29]. Further details about the design and characterization of the electromechanical properties of the actuators can be found in Ref. [37].

Tactors are arranged in five columns and two rows on a commercial hiking belt (Granite Gear, Two Harbors, MN) that tethers to the backpack. They are mounted on a thin aluminum beam-based scaffold, which is later fastened to the belt, to maintain a minimum horizontal and vertical distance between the actuators of 130 and 110 mm, respectively. We expect this separation to improve users’ discrimination performance, as previously demonstrated [29]. A custom-made tissue sleeve is wrapped around the belt to cover electrical connections and improve users’ comfort, as shown in Fig. 1(b). Velcro® squares on the sleeve and tactors’ cases ensure accurate positioning and prevent damping of the vibration from the sleeve.

MFC actuators are driven by an electrical circuit composed of an arduino mega 2560 board (Arduino LLC, Boston, MA), five printed circuit boards (PCBs), and five high-voltage amplifiers (HVAs). Each PCB connects to an HVA, which can simultaneously drive two actuators. PCBs integrate a dual-channel operational amplifier (TLC2202CP, Texas Instruments, Dallas, TX) wired as an astable multivibrator. The switching frequency of the multivibrator is controlled by a dual-channel 10 kΩ digital potentiometer (AD8402ANZ10, Analog Devices Inc., Norwood, MA), whose variable resistance is selected by the arduino through a serial peripheral interface (SPI) protocol. The 0–5 V square wave generated by the multivibrator is fed to an HVA (AMT2012-CE3, Smart Material, Sarasota, FL). The HVA maps linearly the 0–2.5 V range to 0–0.5 kV, and the 2.5–5 V range to 0.5–2 kV. The ground pin of the HVA provides a 500 V offset, such that the output voltage will be in the –0.5 to 1.5 kV range.

Activation of PCBs and HVAs and modulation of the vibration frequency are achieved through a matlab® R2019a script that controls the arduino. PCB components are switched on by feeding a constant 5 V signal to their power pins. HVAs are turned on by shorting their enabling pin. The vibration frequency is modulated by varying the resistance of the digital potentiometer between 0 and 10 kΩ, corresponding to 10 Hz and 390 Hz, respectively, on 256 discrete levels. The desired level of resistance is selected by sending 10 bits through a SPI protocol. The arduino board is controlled by a laptop through a USB 2.0 cord, which supplies the power for the PCBs.

Power for the HVAs is provided by a TalentCell 3 12 V, 142.08 Wh lithium-ion battery (PB120B1, TalentCell, Shenzen, Guangdong Province, China), weighing 1 kg. This battery allows an operational period of three hours, assuming that all the actuators are continuously vibrating at resonance. Connection between the lithium-ion battery and HVAs is enabled by a break-out board through a standard 2.1 mm DC barrel-jack connector.

3 Computer Vision System

The computer vision system utilizes a ZED stereo camera (Stereolabs Inc., San Francisco, CA), which enables reconstruction of the depth field from the acquired images. The camera is encapsulated in a custom-made 3D-printed case that is secured to the shoulder straps of the bookbag and connected to a laptop through a USB 3.0 port.

We employ an Alienware 13 R3 (Dell, Round Rock, TX) with an Ubuntu 16.04.6 LTS operating system as the main processing unit. In this proof of concept, the choice of a laptop is motivated by the ease of programming and data accessibility. However, this solution would prove bulky in a commercial product, for which we will consider more compact options such as NVIDIA® Jetson™ modules.

Given the recent advancement in convolutional neural networks for object detection, in this article, we exploit such advances for real-time object detection. Our object detection method is adapted from the well-known real-time object detection approach You Only Look Once (YOLO) [39] with python 2.7 [40]. YOLO belongs to a family of computer vision algorithms that is currently able to identify over 9000 classes of objects [41]. YOLO was selected due to its small elaboration time per image, which is critical for our application. YOLO was developed based on a single neural network applied to the full image; then, the image is divided into regions with different sizes, and finally bounding boxes and probabilities are predicted for each region. The system outputs 3D locations from the point cloud associated with each identified object, along with one of the originally implemented semantic labels [39] and its detection confidence. From these data, we compute the centroid of the object as the average of the location of its points. Furthermore, we calculate the distance of the object as the Euclidean norm of the 3D vector at its centroid.

To spatiotopically map the obstacles to the tactors, we divide the acquired image in an array of ten equal rectangles, arranged in two rows and five columns, matching the spatial density of the belt. We assume that an object belongs to one of these rectangles when the projection of the centroid of the object on the image plane falls within the rectangle, see Fig. 2. This elaboration constitutes the activation signal for the tactors on the belt through the arduino, as described in Sec. 4. The algorithm can analyze images and provide these signals at a rate of 50 Hz, which is faster than the time in which typical saccadic eye movements would be able to acquire potential hazards [42].

Fig. 2
Example of the elaboration from the computer vision system. In the image acquired by the camera, a chair is identified with 59.0% confidence in the third column from the left and first row from the bottom, at a distance of 1.04 m from the camera (see bottom right of the computer vision image). The corresponding actuator in a spatiotopically preserved manner (framed in the figure) will vibrate on the belt to signal the presence of the obstacle.
Fig. 2
Example of the elaboration from the computer vision system. In the image acquired by the camera, a chair is identified with 59.0% confidence in the third column from the left and first row from the bottom, at a distance of 1.04 m from the camera (see bottom right of the computer vision image). The corresponding actuator in a spatiotopically preserved manner (framed in the figure) will vibrate on the belt to signal the presence of the obstacle.
Close modal

4 Interface Between the Wearable Haptic Device and Computer Vision System

All the electronic components and the laptop are stored in a commercial backpack (V6012, Victoriatourist, WandF, Gathersburg, MD), as shown in Fig. 3. The camera is mounted in front of the backpack, to maintain alignment with the torso and avoid possible confounds due to head motion; the torso is never misaligned with the intended travel path, whereas the head and other potential body areas/segments often are. The tactors are connected to the HVAs in the backpack through channeled wires, which are covered by the sleeve and by a binding coil, as shown in Fig. 1(b).

Fig. 3
Overall system on a person. The belt, here exposed for the sake of clarity, can be worn below personal garments. The camera, fixed on one of the two backpack straps, is locked in position with a buckle after the backpack is worn.
Fig. 3
Overall system on a person. The belt, here exposed for the sake of clarity, can be worn below personal garments. The camera, fixed on one of the two backpack straps, is locked in position with a buckle after the backpack is worn.
Close modal

The electronics of the haptic device are interfaced with the computer vision system through the main processing unit, as shown in Fig. 4. We utilize a transmission control protocol (TCP) to establish a connection between the computer vision algorithm in python 2.7 and the arduino control script in matlab® R2019b. When the system is turned on, as a first step, the matlab script initializes the arduino and opens the TCP, functioning as a server. Next, the python code establishes the connection as a client and begins to acquire images from the ZED camera, while the matlab script enters a spinning loop, waiting for data from the computer vision system. Once a frame is acquired, it is analyzed by the python code, which identifies all the objects present in the scene. For each object that is detected, the python code sends two numbers to the matlab script: the identifier number of the tactor to switch on, and the distance of the object. Finally, the matlab script processes the received information, sends signals to the arduino, and re-enters the spinning loop. To avoid saturating the connection and to limit power consumption, delays of 30 ms and 25 ms are added after each information transfer in python and after each check during the spinning loop in matlab, respectively.

Fig. 4
Schematics of the integrated ETA proposed in this article. The ZED camera acquires images, which are analyzed by the python 2.7 computer vision code running on the Alienware laptop. The output of the python code is sent through a TCP connection to a matlab script that controls the arduino. The arduino drives five dual-channel PCBs, each connected to a dual-channel HVA. Each HVA controls a column of two actuators on the belt.
Fig. 4
Schematics of the integrated ETA proposed in this article. The ZED camera acquires images, which are analyzed by the python 2.7 computer vision code running on the Alienware laptop. The output of the python code is sent through a TCP connection to a matlab script that controls the arduino. The arduino drives five dual-channel PCBs, each connected to a dual-channel HVA. Each HVA controls a column of two actuators on the belt.
Close modal

matlab reads all the information in the TCP queue. For each detected object, it first verifies the depth of the object from the end user and then maps the distance to a frequency, such that closer objects will have a higher frequency, with the closest objects at the resonance frequency. To facilitate frequency discrimination, we consider only discrete frequency steps.

Specifically, we set the frequency to a level in the 25–175 Hz range, with intervals of approximately 25 Hz. First, a peri-personal distance range of 1–10 m is mapped linearly to the 25–175 Hz frequency range, with 10 m mapped to 25 Hz and 1 m mapped to 175 Hz. Then, we round the obtained frequency to the closest admissible integer frequency. The frequency is then mapped to a resistance value, which is sent to the digital potentiometer on the PCBs through an SPI protocol. If the corresponding tactor is already on, no further action is taken; otherwise, it is switched on for 200 ms. Such an intermittent pulse allows for easier vibration discrimination compared to continuous stimulation, which would elicit habituation.

To assess performance of the connection between the computer vision system and the haptic device, we perform two experiments, through which we seek to quantify the elaboration and transmission delay in the python and matlab codes, respectively. These two tests are performed independently to minimize the delay associated to saving the results. For the duration of the two experiments, the experimenter carries the computer vision system connected to the arduino in the backpack to analyze an environment; the participant walks in a repeated back-and-forth loop along a corridor 5 m long. During ambulation, the ZED camera scans the ambient space. The obstacles, encompassing a desk, multiple chairs, computers, monitors, laptops, and other smaller objects, are at fixed and known locations. Since our focus is on the interface between the computer vision system and the wearable haptic device, we do not consider performance of the individual subsystems (such as accuracy in object classification or vibration intensity).

More specifically, in the first experiment, we run the computer vision code in python for about 3 min while measuring the time elapsed from image acquisition to transmission of data to the matlab script, along with the number of identified objects. We exclude from our analysis all the acquired frames in which no object was detected by the computer vision system. For each frame that contains at least one identified object, we consider three contributions of the delay in the python script: the time to acquire the image and identify all point clouds in the frame (“PtCl”), to classify the detected objects and compute their distances (“Clas”), and to determine the position of the centroid of each object in the image and send all data through the TCP connection (“Send”). We consolidate all these contributions in the total delay in the python code (“Tot”).

In Fig. 5, we show boxplots indicating the delay for each of these operations. The median value of the total delay is 20.9 ms. Identifying the position of the objects and sending the data through the TCP have a negligible contribution to this delay, which is mostly determined by the operations in YOLO (generating the point cloud and labelling the objects, PtCl and Clas). We observe that the distributions of delays are skewed, with most of the data around the median value and a significant tail toward larger delays. Thus, the system normally behaves as expected, but in several instances, the delays increase significantly. Interestingly, the Pearson correlation coefficient indicates that only the time to send data through TCP correlates with the number of detected objects in the frame (r = 0.93), whereas all the other contributions and the total delay do not (r < 0.22). This suggests that the performance of the python script is independent of the number of objects detected in the frames.

Fig. 5
Contributions to the delay in the python script, over 4745 object detections in 2662 frames, over a total of 3918 acquired frames. “PtCl” is the time required to reconstruct the point cloud of all the objects in the frame, “Clas” represents the time to assign a label to each object and determine its distance, and “Send” indicates the time required for determining the position of all the objects in the frame and sending all data through the TCP connection. “Tot” is the total delay, computed by summing all the previous contributions. Dash lines are the medians, boxes delimit the first and third quartiles, vertical lines indicate the 1.5-interquartile range, and crosses are outliers.
Fig. 5
Contributions to the delay in the python script, over 4745 object detections in 2662 frames, over a total of 3918 acquired frames. “PtCl” is the time required to reconstruct the point cloud of all the objects in the frame, “Clas” represents the time to assign a label to each object and determine its distance, and “Send” indicates the time required for determining the position of all the objects in the frame and sending all data through the TCP connection. “Tot” is the total delay, computed by summing all the previous contributions. Dash lines are the medians, boxes delimit the first and third quartiles, vertical lines indicate the 1.5-interquartile range, and crosses are outliers.
Close modal

In the second experiment, we systematically investigate the delay in the matlab script. Similar to the previous case, we run the computer vision code in python for about 5 min, which transmits data through TCP to the matlab script that elaborates and sends data downstream to arduino. In this case, we consider each reading of the TCP queue, which may contain multiple objects from several frames. We consider four contributions to the transmission delay in matlab: the time to receive the data from the TCP (“Read”), to elaborate on the numbers to find the corresponding pins and the resistance value for the potentiometer (“Elab”), to send data through the SPI protocol and modify the potentiometer resistance toward changing the vibration frequency (“Freq”), and to turn on electronic components of the PCBs, activate the corresponding actuators by shorting the enable pin of the HVAs, and provide the input voltages to the operational amplifiers (“On”). Furthermore, we calculate the total time (“Tot”) to obtain the data from the python code and send it to the arduino as the sum of the previous contributions. In addition, we record the number of objects in the TCP queue.

The results of this delay analysis are shown in Fig. 6. The median value of the total delay is 80.6 ms, with the largest contributions due to the operations of the arduino to modify the vibration frequency or activate the electronics. This value corresponds to a median time from acquisition to stimulation of 101.5 ms, an effective frequency of almost 10 Hz. Similar to the previous case, the dispersion of the data regarding the delay contributions in the matlab script is significant, with a non-Gaussian distribution skewed toward lower values. Large delays may be related to freezing of the SPI connection, which causes the arduino to reset. By computing the Pearson correlation coefficient, we find that delays are not correlated to the number of objects in the TCP queue (|r| < 0.04 for all contributions and total time). The significant delay introduced by arduino calls for the adoption of a more robust and reliable way to interface with the electronics. Given that the number of detected objects in the TCP queue does not continue increasing, we expect that the matlab script should be able to continue up with the transmission, emptying the queue accumulating in the TCP connection. This claim requires further integrated analyses during longer utilization periods, which will indicate the actual transmission delay between image acquisition and tactor activation.

Fig. 6
Contributions to the delay in the matlab script, over 9209 object detections, in 2942 TCP queue readings. “Read” indicates the reading from the TCP connection, “Elab” is the elaboration time of acquired data, “Freq” represents the time required to change the actuation frequency, “On” is the time to turn on the electronics, short the enable pin of the HVAs, and start generating the square wave. “Tot” is the total delay, computed by summing all the previous contributions. Dash lines are the medians, boxes delimit the first and third quartiles, vertical lines indicate the 1.5-interquartile range, and crosses are outliers.
Fig. 6
Contributions to the delay in the matlab script, over 9209 object detections, in 2942 TCP queue readings. “Read” indicates the reading from the TCP connection, “Elab” is the elaboration time of acquired data, “Freq” represents the time required to change the actuation frequency, “On” is the time to turn on the electronics, short the enable pin of the HVAs, and start generating the square wave. “Tot” is the total delay, computed by summing all the previous contributions. Dash lines are the medians, boxes delimit the first and third quartiles, vertical lines indicate the 1.5-interquartile range, and crosses are outliers.
Close modal

5 Conclusions

In this study, we developed a new, integrated ETA for visual impairment. The ETA comprises a computer vision system for obstacle detection and a bookbag-based belt with discrete tactors, providing haptic feedback relative to obstacle location in an egocentric, spatiotopically preserved reference frame. Compared to solutions proposed in the literature, our system constitutes a customizable, inconspicuous option, whereby the belt could be detached and worn under personal garments and the remaining electronics and sensing components could be integrated in commercial backpacks. This ETA and approaches fashioned in a similar way could address the needs of the visually impaired, delivering unobtrusive aesthetics and an ergonomic design, critical aspects for the adoption of new assistive technologies [13,14].

Future studies and improvements are necessary to ensure usability of the device and potential commercialization. While preliminary mechanical tests were performed on the tactors, systematic, rigorous testing is required to quantify their performance against the previous generation of actuators [30]. In this vein, discrimination tests [29] will provide a first benchmark to assess the capabilities of our new wearable device, from which we expect a higher performance compared to our previous studies. Experiments on obstacle avoidance are the cornerstone in the evaluation of our integrated system. We will conduct hypothesis-driven experiments to demonstrate the effectiveness of the haptic feedback, for example, by counting the number of collisions with obstacles during a navigation task with and without the belt. We will perform these experiments with healthy subjects under simulated impairment conditions and with visually impaired subjects.

When testing the performance of the integrated system, we expect that our current control system will not relay information to the user with high fidelity in all circumstances. For example, it is untenable to assume that the vibration of multiple actuators in the presence of multiple obstacles will provide meaningful data to the end user, rather it is likely to create sensory overload; this must be addressed through adaptive vibratory and/or audio messages. These observations warrant further investigation toward the development of a more sophisticated control system. After optimizing the control system, we will test the performance of our device against analogous systems in the literature. Finally, several technical aspects, including integration of other sensors in the prototype, miniaturization of the main processing unit, a single-integrated software script for both computer vision and control of the belt’s microprocessor, and incorporation of the tactors into customized bookbag-based waist straps, shoulder straps, and back straps and padding, should be considered toward a wearable device with significant commercial appeal.

Acknowledgment

This research was supported by Connect-the-Dots Pilot Award (New York University Langone Health and Tandon School of Engineering) and by the National Science Foundation (Grant Nos. CMMI-1433670 and CNS-1952180).

Conflict of Interest

J.-R. Rizzo discloses conflict interests as a result of intellectual property owned by New York University and related advisory positions with equity and ad hoc compensation. In the future, the aforementioned project may relate to multicomponent wearable technologies relevant to the stated interests.

References

1.
Varma
,
R.
,
Vajaranant
,
T. S.
,
Burkemper
,
B.
,
Wu
,
S.
,
Torres
,
M.
,
Hsu
,
C.
,
Choudhury
,
F.
, and
McKean-Cowdin
,
R.
,
2016
, “
Visual Impairment and Blindness in Adults in the United States: Demographic and Geographic Variations From 2015 to 2050
,”
JAMA Ophthalmol.
,
134
(
7
), pp.
802
809
. 10.1001/jamaophthalmol.2016.1284
2.
Stevens
,
G. A.
,
White
,
R. A.
,
Flaxman
,
S. R.
,
Price
,
H.
,
Jonas
,
J. B.
,
Keeffe
,
J.
,
Leasher
,
J.
,
Naidoo
,
K.
,
Pesudovs
,
K.
,
Resnikoff
,
S.
,
Taylor
,
H.
, and
Bourne
,
R. R.
,
2013
, “
Global Prevalence of Vision Impairment and Blindness: Magnitude and Temporal Trends, 1990–2010
,”
Ophthalmology
,
120
(
12
), pp.
2377
2384
. 10.1016/j.ophtha.2013.05.025
3.
Sherwood
,
M. B.
,
Garcia-Siekavizza
,
A.
,
Meltzer
,
M. I.
,
Hebert
,
A.
,
Burns
,
A. F.
, and
McGorray
,
S.
,
1998
, “
Glaucoma’s Impact on Quality of Life and Its Relation to Clinical Indicators: A Pilot Study
,”
Ophthalmology
,
105
(
3
), pp.
561
566
. 10.1016/S0161-6420(98)93043-3
4.
Legood
,
R.
,
Scuffham
,
P.
, and
Cryer
,
C.
,
2002
, “
Are We Blind to Injuries in the Visually Impaired? A Review of the Literature
,”
Injury Prevention
,
8
(
2
), pp.
155
160
. 10.1136/ip.8.2.155
5.
Zhang
,
X.
,
Bullard
,
K. M.
,
Cotch
,
M. F.
,
Wilson
,
M. R.
,
Rovner
,
B. W.
,
McGwin
,
G.
,
Owsley
,
C.
,
Barker
,
L.
,
Crews
,
J. E.
, and
Saaddine
,
J. B.
,
2013
, “
Association Between Depression and Functional Vision Loss in Persons 20 Years of Age Or Older in the United States, NHANES 2005–2008
,”
JAMA Ophthalmol.
,
131
(
5
), pp.
573
581
. 10.1001/jamaophthalmol.2013.2597
6.
Seddon
,
J. M.
,
Cote
,
J.
,
Davis
,
N.
, and
Rosner
,
B.
,
2003
, “
Progression of Age-Related Macular Degeneration: Association With Body Mass Index, Waist Circumference, and Waist-Hip Ratio
,”
Arch. Ophthalmol.
,
121
(
6
), pp.
785
792
. 10.1001/archopht.121.6.785
7.
McCarty
,
C. A.
,
Nanjan
,
M. B.
, and
Taylor
,
H. R.
,
2001
, “
Vision Impairment Predicts 5 Year Mortality
,”
British J. Ophthal.
,
85
(
3
), pp.
322
326
. 10.1136/bjo.85.3.322
8.
Access Economics Pty Limited
,
2009
, “
The Cost of Vision Loss in Canada
,”
CNIB and the Canadian Ophthalmological Society, Summary Report, Technical Report, Toronto, Canada
.
9.
Winter
,
B.
,
2020
, “
10 Fascinating Facts About the White Cane
,”
Perkins School for the Blind
, https://www.perkins.org/stories/10-fascinating-facts-about-the-white-cane, Accessed February 16, 2020.
10.
Guiding Eyes for the Blind
,
2020
, “
How Many People Use Guide Dogs?
https://www.guidingeyes.org/about/faqs/, Accessed February 16, 2020.
11.
Rogers
,
P.
,
2020
, “
The White Cane: Symbol of Dependence or Independence
,” VisionAware™, https://www.visionaware.org/blog/visually-impaired-now-what/the-white-cane-symbol-of-dependence-or-independence/12, Accessed February 16, 2020.
12.
13.
Ye
,
H.
,
Malu
,
M.
,
Oh
,
U.
, and
Findlater
,
L.
,
2014
, “
Current and Future Mobile and Wearable Device Use by People With Visual Impairments
,”
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
, pp.
3123
3132
.
14.
Shinohara
,
K.
, and
Wobbrock
,
J. O.
,
2011
, “
In the Shadow of Misperception: Assistive Technology Use and Social Interactions
,”
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
,
Toronto, Canada
,
Apr. 26–May 1
, pp.
705
714
.
15.
Bujacz
,
M.
, and
Strumiłło
,
P.
,
2016
, “
Sonification: Review of Auditory Display Solutions in Electronic Travel Aids for the Blind
,”
Arch. Acous.
,
41
(
3
), pp.
401
414
. 10.1515/aoa-2016-0040
16.
Vincent
,
C.
,
Routhier
,
F.
,
Martel
,
V.
,
Mottard
,
M.-E.
,
Dumont
,
F.
,
Côté
,
L.
, and
Cloutier
,
D.
,
2014
, “
Field Testing of Two Electronic Mobility Aid Devices for Persons Who Are Deaf-blind
,”
Disab. Rehabilit.: Assist. Tech.
,
9
(
5
), pp.
414
420
. 10.3109/17483107.2013.825929
17.
Bach-y Rita
,
P.
,
Kaczmarek
,
K. A.
,
Tyler
,
M. E.
, and
Garcia-Lara
,
J.
,
1998
, “
Form Perception with a 49-point Electrotactile Stimulus Array on the Tongue: a Technical Note
,”
J. Rehabilit. Res. Develop.
,
35
(
4
), pp.
427
430
.
18.
Ward
,
J.
, and
Meijer
,
P.
,
2010
, “
Visual Experiences in the Blind Induced by an Auditory Sensory Substitution Device
,”
Consciousness Cognition
,
19
(
1
), pp.
492
500
. 10.1016/j.concog.2009.10.006
19.
Rizzo
,
J.-R.
,
Conti
,
K.
,
Thomas
,
T.
,
Hudson
,
T. E.
,
Wall Emerson
,
R.
, and
Kim
,
D. S.
,
2018
, “
A New Primary Mobility Tool for the Visually Impaired: A White Cane–Adaptive Mobility Device Hybrid
,”
Assist. Technol.
,
30
(
5
), pp.
219
225
. 10.1080/10400435.2017.1312634
20.
Ulrich
,
I.
, and
Borenstein
,
J.
,
2001
, “
The GuideCane—Applying Mobile Robot Technologies to Assist the Visually Impaired
,”
IEEE Trans. Syst., Man, Cyber.-Part A: Syst. Humans
,
31
(
2
), pp.
131
136
. 10.1109/3468.911370
21.
Dakopoulos
,
D.
, and
Bourbakis
,
N. G.
,
2009
, “
Wearable Obstacle Avoidance Electronic Travel Aids for Blind: A Survey
,”
IEEE Trans. Syst., Man, Cyber., Part C (Appl. Rev.)
,
40
(
1
), pp.
25
35
. 10.1109/TSMCC.2009.2021255
22.
Stern
,
J. H.
,
Tian
,
Y.
,
Funderburgh
,
J.
,
Pellegrini
,
G.
,
Zhang
,
K.
,
Goldberg
,
J. L.
,
Ali
,
R. R.
,
Young
,
M.
,
Xie
,
Y.
, and
Temple
,
S.
,
2018
, “
Regenerating Eye Tissues to Preserve and Restore Vision
,”
Cell. Stem. Cell.
,
22
(
6
), pp.
834
849
. 10.1016/j.stem.2018.05.013
23.
Pissaloux
,
E.
, and
Velazquez
,
R.
,
2017
,
Mobility of Visually Impaired People. Fundamentals and ICT Assistive Technologies
,
Springer
,
New York
.
24.
Bhowmick
,
A.
, and
Hazarika
,
S. M.
,
2017
, “
An Insight Into Assistive Technology for the Visually Impaired and Blind People: State-of-the-Art and Future Trends
,”
J. Multi. User Interfaces
,
11
(
2
), pp.
149
172
. 10.1007/s12193-016-0235-6
25.
Rizzo
,
J.-R.
,
Pan
,
Y.
,
Hudson
,
T.
,
Wong
,
E. K.
, and
Fang
,
Y.
,
2017
, “
Sensor Fusion for Ecologically Valid Obstacle Identification: Building a Comprehensive Assistive Technology Platform for the Visually Impaired
,”
2017 7th International Conference on Modeling, Simulation, and Applied Optimization (ICMSAO)
,
Sharjah, United Arab Emirates
,
Apr. 4–6
,
IEEE
, pp.
1
5
.
26.
Li
,
X.
,
Cui
,
H.
,
Rizzo
,
J.-R.
,
Wong
,
E.
, and
Fang
,
Y.
,
2019
, “
Cross-Safe: A Computer Vision-based Approach to Make All Intersection-Related Pedestrian Signals Accessible for the Visually Impaired
,”
Science and Information Conference
,
Springer
, pp.
132
146
.
27.
Bhatlawande
,
S. S.
,
Mukhopadhyay
,
J.
, and
Mahadevappa
,
M.
,
2012
, “
Ultrasonic Spectacles and Waist-Belt for Visually Impaired and Blind Person
,” 2012
National Conference on Communications (NCC)
,
Kharagpur, India
,
Feb. 3–5
,
IEEE
, pp.
1
4
.
28.
Ton
,
C.
,
Omar
,
A.
,
Szedenko
,
V.
,
Tran
,
V. H.
,
Aftab
,
A.
,
Perla
,
F.
,
Bernstein
,
M. J.
, and
Yang
,
Y.
,
2018
, “
Lidar Assist Spatial Sensing for the Visually Impaired and Performance Analysis
,”
IEEE Trans. Neural Syst. Rehabilit. Eng.
,
26
(
9
), pp.
1727
1734
. 10.1109/TNSRE.2018.2859800
29.
Cholewiak
,
R. W.
,
Brill
,
J. C.
, and
Schwab
,
A.
,
2004
, “
Vibrotactile Localization on the Abdomen: Effects of Place and Space
,”
Perception Psychophys.
,
66
(
6
), pp.
970
987
. 10.3758/BF03194989
30.
Phamduy
,
P.
,
Rizzo
,
J.-R.
,
Hudson
,
T. E.
,
Torre
,
M.
,
Levon
,
K.
, and
Porfiri
,
M.
,
2018
, “
Communicating Through Touch: Macro Fiber Composites for Tactile Stimulation on the Abdomen
,”
IEEE Trans. Haptics
,
11
(
2
), pp.
174
184
. 10.1109/TOH.2017.2781244
31.
Wilkie
,
W. K.
,
Bryant
,
R. G.
,
High
,
J. W.
,
Fox
,
R. L.
,
Hellbaum
,
R. F.
,
Jalink
,
A., Jr.
,
Little
,
B. D.
, and
Mirick
,
P. H.
,
2000
, “
Low-Cost Piezocomposite Actuator for Structural Control Applications
,”
Smart Structures and Materials 2000: Industrial and Commercial Applications of Smart Structures Technologies, Proceedings of SPIE 3991
,
Newport Beach, CA
,
June 12
, Vol.
3991
, pp.
323
334
.
32.
Tiersten
,
H. F.
,
2013
,
Linear Piezoelectric Plate Vibrations: Elements of the Linear Theory of Piezoelectricity and the Vibrations Piezoelectric Plates
,
Springer
,
New York
.
33.
Humer
,
A.
,
2013
, “
Exact Solutions for the Buckling and Postbuckling of Shear-Deformable Beams
,”
Acta Mech.
,
224
(
7
), pp.
1493
1525
. 10.1007/s00707-013-0818-1
34.
Neukirch
,
S.
,
Frelat
,
J.
,
Goriely
,
A.
, and
Maurini
,
C.
,
2012
, “
Vibrations of Post-Buckled Rods: The Singular Inextensible Limit
,”
J. Sound. Vib.
,
331
(
3
), pp.
704
720
. 10.1016/j.jsv.2011.09.021
35.
Erturk
,
A.
, and
Delporte
,
G.
,
2011
, “
Underwater Thrust and Power Generation Using Flexible Piezoelectric Composites: An Experimental Investigation Toward Self-Powered Swimmer-Sensor Platforms
,”
Smart Mater. Struct.
,
20
(
12
), p.
125013
. 10.1088/0964-1726/20/12/125013
36.
Cha
,
Y.
,
Chae
,
W.
,
Kim
,
H.
,
Walcott
,
H.
,
Peterson
,
S. D.
, and
Porfiri
,
M.
,
2016
, “
Energy Harvesting From a Piezoelectric Biomimetic Fish Tail
,”
Renew. Energy
,
86
, pp.
449
458
. 10.1016/j.renene.2015.07.077
37.
Boldini
,
A.
,
Rizzo
,
J.-R.
, and
Porfiri
,
M.
,
2020
, “
A Piezoelectric-Based Advanced Wearable: Obstacle Avoidance for the Visually Impaired Built Into a Backpack
,”
Nano-Bio-, Info-Tech Sensors, and 3D Systems IV
,
Apr. 22
, Online, Vol.
11378
,
International Society for Optics and Photonics
, p.
1137806
.
38.
Bolanowski Jr
,
S. J.
,
Gescheider
,
G. A.
,
Verrillo
,
R. T.
, and
Checkosky
,
C. M.
,
1988
, “
Four Channels Mediate the Mechanical Aspects of Touch
,”
J. Acoust. Soc. Am.
,
84
(
5
), pp.
1680
1694
. 10.1121/1.397184
39.
Redmon
,
J.
,
Divvala
,
S.
,
Girshick
,
R.
, and
Farhadi
,
A.
,
2016
, “
You Only Look Once: Unified, Real-Time Object Detection
,”
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
,
Las Vegas, NV
,
June 27–30
, pp.
779
788
.
40.
StereoLabs - GitHub™
,
2020
, “
3D Object Detection Using Yolo and the ZED in Python and C++
, https://github.com/stereolabs/zed-yolo, Accessed February 16, 2020.
41.
Redmon
,
J.
, and
Farhadi
,
A.
,
2017
, “
Yolo9000: Better, Faster, Stronger
,”
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
,
Honolulu, HI
,
July 21–26
, pp.
7263
7271
.
42.
Leigh
,
R. J.
, and
Zee
,
D. S.
,
2015
,
The Neurology of Eye Movements
,
Oxford University Press
,
Oxford, UK
.