вход по аккаунту



код для вставкиСкачать
Patent Translate
Powered by EPO and Google
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a
three-dimensional sound image control apparatus for three-dimensionally generating the sound
of an individual object in a three-dimensional image to a listener.
2. Description of the Related Art For example, Japanese Patent Application Laid-Open No. 3-9700
discloses a technique for controlling a sound image in space using three speakers, but the
obtained sound images are located in the same plane and in the plane. Although the position of
the sound image can be controlled, all the sounds can only be heard from the front of the
Also, Japanese Patent Laid-Open No. Hei 4-56500 discloses a technology for controlling a sound
image using 2n speakers for n listening positions, but an acoustic signal itself processed at the
front stage of 2n speakers is disclosed. The sound image emitted from this speaker is also flat for
each listener and the sound image can be placed at any position in the listening space because
There is a problem that localization can not be performed, and a three-dimensional sound image
in a true sense can not be obtained.
SUMMARY OF THE INVENTION The present invention has been made in view of the above-
mentioned problems of the prior art, in which each sound source is positioned at a position in
space corresponding to the spatial position of each object in a stereoscopic image. And sound
matching, and it is an object of the present invention to obtain a more realistic 3D stereoscopic
audiovisual apparatus.
According to the present invention, the angle θ n from the front to each object in the image with
reference to the position of the listener along with the three-dimensional video signal of still
image or moving image, distance rn, and Three-dimensional video data including a threedimensional audio signal consisting of an audio signal Sn generated by each object, a plurality of
speaker means arranged around the listener, and the three-dimensional audio signal (.theta.n, rn,
Sn) And a signal processing unit for distributing the sound signal Sn to the plurality of speaker
means in correspondence with the position of the object in each image, and dividing the plurality
of speakers into a plurality of areas, The signal processing unit specifies a region in which a
sound image corresponding to an object is located by the angle θn, selects a speaker to be used,
and calculates the distance attenuation of the sound signal Sn by the distance rn.
In the above configuration, the three-dimensional audio signal is taken into the signal processing
unit simultaneously with the video signal reproduction, and processing such as distance
attenuation of the acoustic signal, sound image area specification, output speaker selection, etc. is
performed here, and the selected speaker is selected. The processed three-dimensional sound
signal is output to obtain a three-dimensional sound image.
three-dimensional sound image system according to the present invention will be described in
detail with reference to the drawings.
FIG. 1 is a circuit block diagram showing an essential part of a three-dimensional sound imaging
apparatus using six speakers.
In the figure, 11 to 1n are signal processing units which input 3D audio signals (.theta.1, r1, S1)
to (.theta.n, rn, Sn) of each of the objects 1 to n in a 3D stereoscopic image and perform
processing to be described later 21 to 26 are adders for adding the outputs of the signal
processing units 11 to 1 n, and 31 to 36 are speakers A to F driven by the outputs of the adders
21 to 26.
In addition, although an amplifier etc. are required in the front | former stage of each speaker
A31-F36, it is abbreviate | omitting in this figure.
Here, the three-dimensional audio signals (.theta.1, r1, S1) to (.theta.n, rn, Sn) will be described
with reference to FIG.
In order to reflect each object in the 3D image in the space where the listener is present, for
example, with the listener O as the origin, it is as if the object (sound source) 1 was at this origin
O in polar coordinate display at the position of (θ1, r1) Since it is a condition that the sound
signal S1 is to be emitted, the audio signal input to the signal processing units 11 to 1n needs to
include such information.
Accordingly, such an audio signal is included in the original stereoscopic video signal, and the
audio signal is also taken out simultaneously with the reproduction of the video.
Next, the arrangement of the speakers A31 to F36 will be described with reference to FIG. The
six speakers A31 to F36 have an equal distance l from the listener O, and set the front face at 0
° -30 ° (speaker A31), 30 ° (speaker B32), 90 ° (speaker C33), 150 ° (speaker D34) , 210
° (speaker E35) and 270 ° (speaker F36). As for how to take the angle, the clockwise direction
is the positive direction. Thus, the speakers A31 to F36 are arranged around the listener O every
60 °. Then, a range (−30 ° ≦ θ <30 °) surrounded by the listener O and the speakers A31
and B32 is defined as a range I, and the respective ranges of the ranges II to VI are set in the
same manner. The signal processing units 11-1n calculate the distance attenuation with respect
to the acoustic signal Sn, specify the areas I-VI, select the speakers A31-F36 to be used, and
select the selected speakers A31-A36. The sound signal Sn is distributed to F36.
That is, first, as shown in FIG. 4, for the acoustic signal Sn input and the distance rn
The distance attenuation Sn ′ of the acoustic signal is determined using
Next, the regions I to IV are identified based on the input angle .theta.n.
Data of a pair of corresponding used speakers A31 to F36 is given in advance to each of the
regions I to IV, and if the region I is specified, the speakers to be used are A31 and B32, and if
the region II, B32 and C33, region III In this case, C33 and D34, D34 and E35 in the region IV,
E35 and F36 in the region V, and F36 and A31 in the region VI are selected.
And the angle θ n is
It is converted into .theta.n '(' < Thus, the result of the
process shown in FIG. 4 is output to the next process as Sn ′, θn ′ and information M of the
specified area.
The next process is the distribution of the acoustic signal. This distributes the acoustic signal to
the pair of speakers (speakers A31 and B32 in FIG. 5, and speakers C33 and D34 in FIG. 6) based
on the signals Sn 'and .theta.n' input as shown in FIG. To obtain distributed acoustic signals Sn
'(A), Sn' (B), Sn '(C), Sn' (D).
For example, when the area information M obtained by the processing of the previous stage of
the signal processing units 11 to 1 n is “area I, use speakers A 31 and B 32”,
The distributed acoustic signals Sn '(A) and Sn' (B) are determined so as to satisfy the above.
In the above equation 3, k is a level difference between the distributed acoustic signals Sn '(A)
and Sn' (B), and
It is given by If the unit of ΔI in Eq. 4 is dB, and the levels of the distributed acoustic signals Sn
′ (A) and Sn ′ (B) are A [dB] and B [dB], then
It is given by
Then, the distributed acoustic signals Sn '(A) and Sn' (B) obtained as a result of the above
processing are finally mixed in the signal processing units 11 to 1 n through the adders 21 to 26
and amplified through the amplifier for each By supplying the corresponding speakers A31 to
F36, it is possible to localize a sound image as if the object is at a position in the listening space
corresponding to the object in the image and the sound is emitted from the position.
In the three-dimensional sound image control apparatus having the above configuration, for
example, assuming that there are three objects (sound sources) in the image, these are the object
1, the object 2 and the object 3, and the object 1 is in the listening space corresponding to the
video space. It is assumed that the object 2 is in the region I, the object 2 is in the region II, and
the object 3 is in the region IV.
In this case, as shown in FIG. 7, the audio signals (.theta.1, r1, S1), (.theta.2, r2, S2) and (.theta.3,
r3, S3) of the objects 1 to 3 respectively correspond to the corresponding signal processing units
11, 12 , 13 to produce distributed acoustic signals S1 '(A), S1' (B), S2 '(B), S2' (C), S3 '(D), S3' (E),
S1 '. (A) to adder 21, S1 '(B) and S2' (B) to adder 22, S1 '(C) to adder 23, S3' (D) to adder 24, S3
'(E) is supplied to the adder 25.
Reference numerals 41 to 46 denote amplifiers, which are the distributed acoustic signals S1 '(A),
S1' (B), S2 '(B), S2' (C), S3 '(D) mixed by the respective adders 21 to 25. , S3 '(E) drive the
corresponding speakers A31 to E35 as desired acoustic signals, and obtain a three-dimensional
sound image as if the object were located in the regions I, II, and IV, and the sound could be
heard from there. Be
As described above, according to the present invention, the three-dimensionalization of audio
which has not been realized conventionally is realized, and it is possible for the viewer to see an
object appearing out of the screen in the three-dimensional image. In contrast, in the field of VR
(Virtual Reality) where the listener can be heard easily from the position where the sound comes
out or in the center of the image, the object is positioned around the viewer. Since it is easy to
create a situation where audio can be heard from there, it will be able to provide a simulated
experience of a three-dimensional space more effectively.
Без категории
Размер файла
14 Кб
Пожаловаться на содержимое документа