Académique Documents
Professionnel Documents
Culture Documents
1. INTRODUCTION
Most visually impaired people have many troubles when they
want to go outside or walk in a place that they are not familiar
with. Even if they have known the location well, they would
have many problems if there were obstacles on the ground or if
unexpected events occurred. The purpose of our study is to
develop a visual aid system that gives spatial information of
surroundings. We hope the system helps the visually impaired to
avoid such troubles mentioned above. We selected an auditory
sense as a sense of presentation information because the
processing faculty of hearing is next to the visual one. And we
opted for CCD camera for an input device to detect the
information of surroundings. A person who puts on the system
can get information actively, for the camera is mounted on the
head.
There are many computer visual aid systems that give visually
impaired surrounding information. However, most of them have
only verbal expressions or verbal interfaces by voice generator.
Verbal information is very important, but it is also important for
visually impaired people to get non-verbal information, like
spatial information that enables them to recognize their
0-7803-5731-0/99/$10.00
01999 IEEE
1-126
an image from the sensor, the system detects a direction and the
distance of an LED marker, and generates a three-dimensional
sound source from the position information. The subjects can
walk to a marker relying on the sound source generated by the
system.
LoudsDeakerl
LoudsDeakeR
2. PRINCIPLE OF 3D SOUND
In this chapter, the principle of three-dimensional sound
processors is simply described. Most three-dimensional sound
processors use Head Related Transfer Functions (HRTFs) to
produce three-dimensional sound. The basic principle is shown
in Fig.1. In the figure, HCjw) indicates transfer functions from a
sound source image produced by the processor to the subject's
ears, and
GCjw) indicates a transfer function from the
loudspeakers to the subject's ears, and SGw) indicates output
signals from the loudspeakers or a sound source image. The
subscripts R and L mean sides of the ears to which the sound
reaches, and the superscripts 1 and 2 mean the number of the
loud speakers. Here, the convolutions between the output signals
of sound source image and its transfer functions can be
described by the other functions:
S ( j w ) - H , ( j w ) = S ' ( j w ) * G A ( j w )+ S2(j w ) - G ; ( j w )
Sound source
@) Headphone mode
X ( j w ) = Gi(jw).G;(jw) -G&w)*GZ(jw)
Needless to say, by using a set of headphones, we can briefly
describe the output signals from them:
S ' ( J w ) = HL(jW).S(jO)
30 sound processor
Joystick G o r n w t e r
J
Position info.
1-127
35 p"-
-___
_I__ ___I_-___
30
25
.E 20
r" 1 5
10
5
0
LS
KK
XY
HT
TM
EN
MA
AM
Mean
Subects
I -128
100
200
300
i ......yjo.l
....................................................
100
200
300
@) Trajectories of subject MA
(a)Trajectories of subject EN
-100
400
0 1 0 0 2 0 0 3 0 0 4 0 0
50
100
150
200
250
300
350
4 3 Experimental procedure
The number of subjects who took part in the experiment was
three visually impaired, including one congenitally blind. Two
visually impaired subjects put on an eye mask for the experiment.
We equipped the head-mounted sensor to their head and
disoriented them before the experiment started. They were
requested to search for the sound source and to reach it. The
walking trajectory and the reaching time were recorded by a
digital video camera that was fixed on the field by a tripod. In
advance, we explained to them that a table was the goal and the
sound source represented it. The distance between the original
position where the subject stood and the table was 1.5m. In the
experiment, we tried seven patterns of directions for each subject.
They were separated by 45 degrees. We did not try directly
behind because a prop that supports a spherical mirror masks
that direction.
3D Scund Processor
I -129
trials with setting the direction of the table to the right. And right
trajectory of subject TM shows the case of the left position.
Those results show us that subject RK and TM are fast and
subject MA is a slow stepper. At the beginning of the experiment,
all of them rotated their head or body to sense the direction of
the sound source. Then, Subject RK and TM walked straight in
the direction that they determined to go without hesitation. On
the other hand, subject MA continued to rotate the head to sense
and confirm the sound position during locomotion. We think that
tendency is mainly caused by the individual characters of each
subject. Subject MA is a man of caution, so he walked slowly
and deliberately, and he was afraid of hitting the table. We think
another reason may be that MA is a congenitally blind subject.
We cant discuss it since the number of subjects was so small,
however, it was possible the results were influenced by the fact
that those eye-masked visually impaired subjects had some
information about the surroundings prior to the experiment and
the congenitally blind subject had less information than the
others. From the point of view that the original purpose for
developing the system is to avoid troubles when the visually
impaired person walks, we have to solve the problem of the
subject walking deliberately using the system.
After the experiment, we took subjects comments. They said
that they feel less accuracy about the distance of the sound
source and themselves in this experiment than in the virtual
space experiment. We think the reason why is that in the real
world experiment they could touch the table before they felt they
were close enough to the sound source. In the virtual space
experiment, they felt the sound source in their head when they
moved to the position of the sound source. Besides, they felt the
sound source is behind them when they went beyond the
position. Those factors might make them have the feelings of
distance to the sound. However, it is very important to express
an absolute distance of sound source and in general, hearing
sense about absolute distance is not linear and not accurate [7].
Therefore, to develop an effective or intuitive method to express
accurate distance from subject to a sound source is a crucial
point in our future research.
subject TM
subject MA
00:Left step
0:Right step
Fig.6: Example of walking trajectories
Table 1:Mean time to reach the table
Subject
RK
TM
MA
Mean
time [sec]
5.1
6.0
69.1
6. REFERENCES
M.Kobayashi, H.Suda, M.Ohta, On an acoustic visual aid
system which enables active acquisition of images, Proc.
ofAsian Control Conference ,1994, pp.957-960
Peter B. L. Meijer, An Experimental System for Auditory
Image Representations, IEEE Trans. on Biomedical Eng.,
V01.39,N0.2 1992,pp.112-121
Jack M. Loomis, C.Hebert, Jack.G.Cicinelli Active
localization of virtual sounds, J. Acoust. Soc. A m , VoI.88,
NO.4 1990 pp.1757-1764
5. CONCLUSION
I -130