close

Вход

Забыли?

вход по аккаунту

?

JP2008131193

код для вставкиСкачать
Patent Translate
Powered by EPO and Google
Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
DESCRIPTION JP2008131193
An object of the present invention is to make it possible for a user to grasp the positional
relationship of conference participants while making it possible to easily hear the voice in the
same conference room. A control device (10A) stores position information of itself and receives
position information of control devices (10B to 10D). The control device 10A determines the
positional relationship of the control devices 10A to 10D based on the received position
information. When voice data including information indicating the position of the talker is
transmitted from control devices 10B to 10D, control device 10A determines the control device
from the information indicating the position of the talker and the positional relationship between
control devices 10B to 10D. The sound image localization position of the sound sent from 10B to
10D is set. Then, an audio signal is generated and supplied to the speaker array SPA so that the
sound image is localized at the set position. [Selected figure] Figure 1
Sound image position control device
[0001]
The present invention relates to a technology for reproducing the sound of a remote place.
[0002]
There is a system disclosed in Patent Document 1 as a conference system in which a plurality of
conference rooms perform a conference.
04-05-2019
1
In this conference system, terminals for performing voice collection in the conference room and
output of voices in other conference rooms are arranged in each conference room, and this
terminal has a function of acquiring the position of the terminal. Then, when the terminals in the
conference room acquire the position information of the other terminals, they calculate the
positional relationship between the own terminal and each terminal, and when outputting the
sound of the other conference room from the headphones, each conference according to the
calculated positional relationship The sound image of the sound of the other conference room is
localized so that the positional relationship of the room is reflected. According to this system,
when listening to the sound of a conference room at a remote location in a conference room, it is
possible to easily grasp the relative positional relationship of the conference participants in real
space.
[0003]
JP, 2005-341092, A
[0004]
By the way, in the system mentioned above, although the sound collected in the conference room
which is in a remote place is reproduced with the headphone, when there are a plurality of
conference participants in a certain conference room, the headphone is prepared for every
participant If there are many participants, it is difficult to arrange headphones.
Also, when using headphones, although the voices of the other conference rooms can be clearly
heard, the ears are blocked, and the voices of the conference participants in the same room are
difficult to hear.
[0005]
The present invention has been made under the above-mentioned background, and its object is to
make it possible for a user to grasp the positional relationship of conference participants while
making it possible to easily hear the voice in the same conference room. .
[0006]
In order to solve the problems described above, the present invention is a sound image position
control device for supplying an audio signal to a connected speaker array, including: direction
information indicating the direction of the speaker array; and the position of the sound image
04-05-2019
2
position control device Storage means for storing information and position information of
another sound image position control device connected to the sound image position control
device via the communication network, sound information representing sound, sound source
position representing the position of the sound source of the sound Receiving means for
receiving information from another sound image position control device connected to the
communication network, and audio signal generating means for generating an audio signal of
audio represented by the audio information according to the audio information received by the
receiving means Position information of the sound image position control device stored in the
storage means, and position information stored in the storage means; Based on the position
information of the sound image position control apparatus which is the transmission source of
the voice information received by the receiving means, the direction information stored in the
storage means, and the sound source position information received by the receiving means The
audio signal generated by the audio signal generation means so that the sound image localization
position of the sound represented by the audio information received by the reception means is
determined, and the sound image of the sound represented by the audio signal is localized at the
determined sound image localization position And an audio signal supply means for processing
the signal and supplying it to the speaker array.
[0007]
In a preferred aspect of the present invention, the sound image position control device detects
position of the sound image position control device, position detection means for detecting the
direction of the speaker array, and the position detection means. And transmission means for
transmitting position information representing a position to another sound image position
control device connected to the communication network, and the receiving means includes
another sound image position control device connected to the communication network. Position
information is received from another sound image position control device, and the storage means
detects position information representing the position detected by the position detection means,
position information received by the reception means, and the direction detection means Store
direction information indicating the selected direction
[0008]
Further, in another preferred aspect of the present invention, the position detection means
detects the position at a predetermined cycle, the direction detection means detects the direction
at a predetermined cycle, and the transmission means is detected by the position detection
means. The position information representing the position is transmitted at a predetermined
cycle, and the receiving means receives position information of another sound image position
control apparatus transmitted at a predetermined cycle from another sound image position
control apparatus connected to the communication network. .
04-05-2019
3
[0009]
According to the present invention, it is possible to make the user understand the positional
relationship of the conference participants while making it possible to easily hear the voice in the
same conference room.
[0010]
Configuration of Embodiment First, the configuration of a conference system 1 according to an
embodiment of the present invention will be described.
FIG. 1 is a view schematically showing the entire configuration of the conference system 1.
The conference rooms A to D shown in the figure are, for example, a conference room A, a
conference room B a district B, a conference room C a district C, a conference room D a district D,
etc. The distance is far away.
In each conference room, a microphone array, a speaker array, and a control device are arranged.
A conference system 1 is a voice communication system that enables a conference to be held
between conference rooms by collecting the sound in the conference room with a microphone
and outputting the collected sound from a speaker array of another conference room. It is.
[0011]
The microphone array, the speaker array, and the control device disposed in each conference
room all have the same configuration.
Therefore, in the following description, the configurations of the microphone array MA, the
speaker array SPA, and the control device 10A disposed in the conference room A will be
described, and the microphone array and the speaker array disposed in the conference rooms B
04-05-2019
4
to D Since the control device is the same as that disposed in the conference room A, the letters B,
C, and D representing each conference room are added to the end, and the description thereof is
omitted.
[0012]
The microphone array MA disposed in the conference room A is a microphone array in which a
plurality of microphones are arranged in a row. The microphone array MA is connected to the
control device 10A and picks up the sound in the conference room A. Then, each microphone
constituting the microphone array MA outputs an audio signal representing the collected sound
to the control device 10A. The speaker array SPA is a speaker array in which a plurality of
speaker units are arranged in a plurality of rows and a plurality of columns, and is connected to
the control device 10A. Each speaker unit of the speaker array outputs a sound corresponding to
the supplied audio signal. The communication network NET is a communication network that
relays communication performed between control devices arranged in each conference room,
and is, for example, the Internet. Control devices 10A to 10D are connected to the
communication network NET.
[0013]
The control device 10A exchanges various information and various signals with the control
devices 10B to 10D via the communication network NET. FIG. 2 is a block diagram showing the
hardware configuration of the control device 10A. The control device 10A includes a power
supply (not shown), and each part of the control device 10A operates with the power supplied
from the power supply. Each part of the control device 10A is connected to the bus 101 as
shown in the figure, and transmits and receives various signals and data between each part via
the bus 101.
[0014]
The antenna 20A is an antenna for receiving a signal for positioning the position output from the
artificial satellite in GPS (Global Positioning System). The antenna 20A receives signals from a
plurality of satellites in GPS, and supplies the received signals to the position analysis unit 106.
The position analysis unit 106 analyzes the signal supplied from the antenna 20A, and generates
position information indicating the position (longitude, latitude) of the control device 10A.
04-05-2019
5
[0015]
The direction sensor 30A is disposed in the speaker array SPA, and outputs a signal indicating
the direction in which the direction sensor 30A is directed to the direction analysis unit 107. In
the present embodiment, the direction sensor 30A is directed in the direction in which the
speaker unit of the speaker array SPA is directed. The direction analysis unit 107 analyzes the
signal supplied from the direction sensor 30A, and obtains the direction in which the speaker
unit of the speaker array SPA is directed. Then, azimuth information indicating the direction in
which the speaker unit is directed is generated. Although not shown in FIG. 1, in the control
device 10B, the antenna 20B and the direction sensor 30B, in the control device 10C, the
antenna 20C and the direction sensor 30C, and in the control device 10D, the antenna 20D and
the direction sensor 30D are connected. It is done.
[0016]
A microphone array MA is connected to the microphone interface 108. The microphone interface
108 converts audio signals supplied from the microphones constituting the microphone array
MA into digital data, and generates audio data indicating the sound represented by the audio
signal. In addition, the microphone interface 108 is a sound source of the collected voice based
on the difference in sound pressure level of the voice collected by each microphone and the
difference in arrival time of voice to each microphone (that is, if the received voice is a human
voice Location of the speaker). Then, sound source position information representing the
specified position is generated, and audio data including the sound source position information is
output.
[0017]
The communication unit 110 is connected to the communication network NET, and functions as
a communication interface when communicating with another control device via the
communication network NET. The communication unit 110 transmits various information such
as position information generated by the position analysis unit 106 and voice data generated by
the microphone interface 108 to other control devices connected to the communication network
NET. In addition, the communication unit 110 receives position information transmitted from
another control device and voice data transmitted from another control device.
04-05-2019
6
[0018]
The speaker control unit 109 includes a DSP (Digital Signal Processor), and a speaker array SPA
is connected. The speaker control unit 109 controls the sound image position of the sound
output from the speaker array SPA, and when receiving the supplied audio data, generates an
audio signal representing the received audio data. Then, the sound image position of the sound
output from the speaker array SPA is determined, processing such as time delay and amplitude
adjustment is performed on the generated audio signal, and the processed audio signal is output
to the speaker array SPA. .
[0019]
The storage unit 105 includes a non-volatile memory that stores various data. The storage unit
105 stores a device identifier that uniquely identifies the control device including the storage
unit 105, and position information received by the communication unit 110. For example, an
identifier “ID-A” is stored as a device identifier in the storage unit 105 of the control device
10A, “ID-B” in the control device 10B, “ID-C” in the control device 10C, In the control device
10D, a device identifier "ID-D" is stored. Further, the storage unit 105 stores information
indicating sound image positions corresponding to the positions of other control devices.
[0020]
A ROM (Read Only Memory) 103 stores a control program executed by the CPU 102. The CPU
102 reads a control program stored in the ROM 103, and executes a control program using a
random access memory (RAM) 104 as a work area. When the control program is executed by the
CPU 102, each part of the control device 10A is controlled by the CPU 102. Then, in the control
device 10A in which the control program is executed, a function of controlling the sound image
position of the sound output from the speaker array SPA in accordance with the positional
relationship of the other control devices 10B to 10D and the positional relationship of the
speaker Is realized.
[0021]
04-05-2019
7
[Operation of Embodiment] Next, the operation of the present embodiment will be described. In
the following description, as shown in FIG. 3, the latitude and longitude of the control device 10A
is (x0, y0), and the latitude and longitude of the control device 10B is (x1, y1). The operation will
be described on the assumption that the longitude is (x2, y2) and the latitude / longitude of the
control device 10D is (x3, y3).
[0022]
First, when the control device 10A is powered on, a signal indicating the direction in which the
direction sensor 30A is directed is output from the direction sensor 30A. When this signal is
input, the direction analysis unit 107 analyzes the input signal to obtain the direction in which
the direction sensor 30A is directed, and generates direction information indicating the obtained
direction (FIG. 10: step S1). . For example, when the direction sensor 30A is facing "south",
direction information representing "south" is generated. Then, the generated direction
information is stored in the storage unit 105. Since the direction in which the direction sensor
30A is facing is the same as the direction in which the speaker units of the speaker array SPA are
facing, the direction in which the direction sensor 30A is facing is the direction in which the
speaker unit is facing, in other words Then, it can be said that the direction in which the speaker
array SPA is facing.
[0023]
Further, the antenna 20A receives a signal transmitted from a GPS satellite, and the received
signal is supplied to the position analysis unit 106. When the signal is supplied, position analysis
unit 106 obtains the position (longitude and latitude) of control device 10A based on the
supplied signal, and generates position information indicating the obtained longitude and latitude
(x0, y0) (Step S2). When position information is generated in position analysis unit 106, the
generated position information (x 0, y 0) and the device identifier (ID-A) stored in storage unit
105 are controlled via communication unit 110. The information is transmitted to the devices
10B to 10D (steps S3 to S5).
[0024]
When the position information and the device identifier are received by the communication unit
04-05-2019
8
110 of the control device 10B, the received position information (x0, y0) and the device identifier
(ID-A) are associated with each other to store the storage portion of the control device 10B. It is
stored in 105 (step S6). Then, in the control device 10B, when storage of the received position
information and device identifier is completed, the device identifier (ID-B) of the control device
10B stored in the storage unit 105 and the position analysis unit 106 of the control device 10B.
The position information (x1, y1) representing the position of the control device 10B, which is
generated in the above, is transmitted to the control device 10A (step S9).
[0025]
When the device identifier and the position information are received by the communication unit
110 of the control device 10A, the received device identifier (ID-B) and the position information
(x1, y1) are associated with each other and stored in the storage unit 105. It is stored (step S10).
Also in the control devices 10C and 10D, the device identifier and the position information
transmitted from the control device 10A are stored (steps S7 and S8), and the device identifier of
the control device 10C transmitted from the control device 10C (ID-C And the position
information (x2, y2) (step S11), and the device identifier (ID-D) and position information (x3, y3)
(step S13) of the control device 10D transmitted from the control device 10D are also the control
device 10A. Are stored in the storage unit 105 (steps S12 and S14). As described above, when
each control device receives the device identifier and position information of another control
device, each control device transmits its own device identifier and position information to the
control device that has transmitted the device identifier. Can acquire position information of
another control device connected to the communication network NET.
[0026]
When the control device 10A receives position information from another control device, the
control device 10A obtains a positional relationship with the other control devices 10B to 10D
based on the received position information (step S15). Specifically, based on the position
information of the control devices 10B to 10D stored in the storage unit 105, the CPU 102 sets
the position of the control device 10B to (x1, y1) as shown in FIG. When the position of (x2, y2)
and the position of the control device 10D are (x3, y3), as shown in FIG. 4, the polar coordinate
position (r1, r1) of the control device 10B with the control device 10A at the center. The polar
coordinate position (r2, θ2) of the control device 10C and the polar coordinate position (r3, θ3)
of the control device 10D are obtained.
04-05-2019
9
[0027]
After obtaining the polar coordinate positions of the control devices 10B to 10D, the CPU 102
sets the sound image position of each sound when the sound of another meeting room is
reproduced in the meeting room A in which the control device 10A is disposed (step S16) .
Specifically, the speaker control unit 109 refers to the direction information generated by the
direction analysis unit 107 and, as shown in FIG. 5, proceeds from the center position of the
speaker array SPA in the horizontal direction toward the speaker array SPA. A predetermined
position P separated by a predetermined distance corresponds to the origin of polar coordinates
when the control device 10A is at the center. Then, since the position of the control device 10B is
(r1, θ1), the position (X1, Y1) in the direction of θ1 seen from the predetermined position P
corresponds to the control device 10B as shown in FIG. Become. Further, since the position of the
control device 10C is (r2, θ2), the position (X2, Y2) in the direction of θ2 viewed from the
predetermined position P is the position corresponding to the control device 10C, and the
position of the control device 10D is Since r 3 and θ 3), the position (X 3, Y 3) in the direction of
θ 3 viewed from the predetermined position P is a position corresponding to the control device
10 D.
[0028]
Next, the speaker control unit 109 normalizes the position of each control device while
maintaining the positional relationship of each control device so that the sound image of the
sound of another conference room is localized within the range where the sound image can be
localized. Do. Then, as shown in FIG. 6, the position after normalization is set as the basic position
of the sound image of each sound when the sound of another conference room is reproduced.
After setting the basic position of the sound image of the sound transmitted from the control
devices 10B to 10D, the CPU 102 supplies the device identifier of each control device and the set
basic position to the speaker control unit 109. Then, in the speaker control unit 109, as shown in
FIG. 7, the supplied basic position is stored in association with the device identifier of each
control device.
[0029]
Next, when a person in the conference room B emits a voice, this voice is collected by the
microphone array MB, and an audio signal representing the collected sound is converted into
voice data by the microphone interface 108 of the control device 10B. . In addition, the
04-05-2019
10
microphone interface 108 is a sound source of the collected voice based on the difference in
sound pressure level of the voice collected by each microphone and the difference in arrival time
of voice to each microphone (that is, if the received voice is a human voice Location of the
speaker). Then, sound source position information representing the specified position is
generated, and audio data including the sound source position information is generated (step
S17) and output (step S18).
[0030]
For example, as shown in FIG. 8, when the speaker B1 of the conference room B emits a voice,
the position of the speaker B1 with respect to the position of the microphone array MB is
determined, and a sound source representing the coordinates of this position (Xb1, yb1) is
generated as position information. Further, (xb2, yb2) is generated as sound source position
information when the speaker B2 emits a sound, and (xb3, yb3) is generated as sound source
position information when the speaker B3 emits a sound. When the sound source position
information is generated, the voice data including the sound source position information and the
device identifier (ID-B) stored in the storage unit 105 are sent from the communication unit 110
to the other control devices 10A, 10C, and 10D. Be
[0031]
When the voice data and the device identifier are received by the communication unit 110 of the
control device 10A, the received voice data and the device identifier are supplied to the speaker
control unit 109. When the audio data and the device identifier are supplied, the speaker control
unit 109 first generates an audio signal of the sound represented by the audio data (step S19).
Next, the speaker control unit 109 reads from the storage unit 105 the basic position (X10, Y10)
stored in association with the supplied device identifier (ID-B). Then, the read basic position is
corrected according to the sound source position information included in the audio data. For
example, when the sound source position information is (xb1, yb1) indicating the position of the
speaker B1, the coordinates of the sound source position information are added to the
coordinates of the basic position, and the sound image position is as shown in FIG. It is corrected
to the position of Y11). When the sound source position information is (xb2, yb2) representing
the position of the speaker B2, the sound image position is set to the position (X12, Y12) as
shown in FIG. 9, and the sound source position information is the speaker In the case of (xb3,
yb3) representing the position of B3, the sound image position is set to the position of (X13,
Y13) as shown in FIG.
04-05-2019
11
[0032]
After setting the sound image position according to the sound source position information, the
speaker control unit 109 performs processing such as time delay or amplitude adjustment at this
set position so that the sound image of the sound of the generated audio signal is localized. The
processed audio signal is supplied to the speaker array SPA (step S20). In the speaker array SPA
to which the audio signal is supplied, a sound corresponding to the audio signal is output.
Thereby, in the conference room A, the sound image of the sound of the conference room B is
localized at the position corresponding to the position of the conference room B and the
positions of the speakers B1 to B3 as shown in FIG.
[0033]
Further, when a person in the conference room D emits a voice, voice data including sound
source position information of the talker is generated (step S21). Then, the voice data including
the sound source position information of the talker and the device identifier (ID-D) stored in the
storage unit 105 are sent from the communication unit 110 to the other control devices 10A,
10B, and 10C (step S22). .
[0034]
When the voice data and the device identifier are received by the communication unit 110 of the
control device 10A, the received voice data and the device identifier are supplied to the speaker
control unit 109. When the audio data and the device identifier are supplied, the speaker control
unit 109 first generates an audio signal of the sound represented by the audio data (step S23).
Next, the speaker control unit 109 reads from the storage unit 105 the basic position (X30, Y30)
stored in association with the supplied device identifier (ID-D). Then, the read basic position is
corrected according to the sound source position information included in the audio data.
[0035]
After setting the sound image position according to the sound source position information, the
speaker control unit 109 performs processing such as time delay or amplitude adjustment at this
04-05-2019
12
set position so that the sound image of the sound of the generated audio signal is localized. The
processed audio signal is supplied to the speaker array SPA (step S24). In the speaker array SPA
to which the audio signal is supplied, a sound corresponding to the audio signal is output.
Thereby, the sound image of the sound of the meeting room D is localized at the position
corresponding to the position of the meeting room D and the position of the talker in the meeting
room A (position near the basic position (X30, Y30)).
[0036]
As described above, in the conference room A, the sound image positions of the voices of other
conference rooms are set according to the positional relationship between the conference room A
and the other conference rooms, so it is easy to distinguish the voices of the conference
participants become. Further, in the present embodiment, since the sound image position is
adjusted according to the position of the talker, it is possible to know the positional relationship
of a plurality of conference participants in the conference room. Further, in the present
embodiment, a person in the conference room can listen to the voices of other conference rooms
without using headphones, so it is necessary to prepare a plurality of headphones even if there
are a large number of conference participants. Absent. In addition, since the conference
participants do not use headphones, they can easily hear the conversation voice of the person in
the room.
[0037]
[Modifications] Although the embodiment of the present invention has been described above, for
example, the above-described embodiment may be modified as follows to carry out the present
invention.
[0038]
In the embodiment described above, the control devices 10A to 10D are provided with input
devices such as a keyboard, and information indicating the direction of the speaker array and the
position of each control device is input by the input device and stored in the storage unit 105.
May be
According to this aspect, there is no need to provide the antenna 20A, the position analysis unit
106, the direction sensor 30A, and the direction analysis unit 107, and therefore, the
04-05-2019
13
configurations of the control devices 10A to 10D can be simplified.
[0039]
In the embodiment described above, the control devices 10A to 10D may obtain their positions at
a constant cycle, and transmit the obtained positions to other control devices at a constant cycle.
[0040]
In the embodiment described above, the distance between the control devices is determined from
the position information stored in the storage unit 105. For example, the sound volume of the
conference room with a long distance is reduced and the sound of the conference room with a
short distance is The volume may be controlled according to the determined distance, such as
increasing the volume.
In the above-described embodiment, the audio signal may be subjected to delay processing to
delay the output of audio according to the distance between the control devices. Further, the
sound quality of the sound output from the speaker array SPA may be changed according to the
calculated distance between the control devices.
[0041]
It is a figure showing the whole composition of the meeting system concerning one embodiment
of the present invention. It is a block diagram showing the hardware constitutions of control
device 10A. It is the figure which illustrated the physical relationship of the control device. It is
the figure which illustrated the physical relationship of each control device. It is the figure which
illustrated the physical relationship of each control device. It is the figure which illustrated the
basic position of the sound image of the sound of each meeting room. It is the figure which
illustrated the stored device identifier and the basic position. FIG. 7 is a diagram illustrating the
positional relationship of speakers in the meeting room B. FIG. 6 is a view exemplifying a
localization position of a sound image in the conference room A. FIG. 6 is a diagram for
explaining the operation of the conference system 1;
Explanation of sign
04-05-2019
14
[0042]
1 ... conference system, A, B, C, D ... conference room, MA, MB, MC, MD ... microphone array, NET
... communication network, SPA, SPB, SPC, SPD ... Speaker array 10A to 10D: control device 20A:
antenna 30A: azimuth sensor 101: bus 102: CPU 103: ROM 104: RAM 105 ... Storage unit 106 ...
Position analysis unit 107 ... Direction analysis unit 108 ... Microphone interface 109 ... Speaker
control unit 110 ... Communication unit
04-05-2019
15
Документ
Категория
Без категории
Просмотров
0
Размер файла
25 Кб
Теги
jp2008131193
1/--страниц
Пожаловаться на содержимое документа