close

Вход

Забыли?

вход по аккаунту

?

JP2006003450

код для вставкиСкачать
Patent Translate
Powered by EPO and Google
Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
DESCRIPTION JP2006003450
An object of the present invention is to provide a user with a sense of realism when outputting a
voice from a system. A plurality of speakers 50a to 50e are provided. Based on the message, the
speaker that outputs the message is determined from among the plurality of speakers 50a to
50e. For this reason, a message can be output by voice also from speakers 50b-50e other than
speaker 50a provided in the driver's seat side. As a result, the sense of reality is improved as
compared with the case where a message is output by voice only from the speaker 50a provided
on the driver's seat side. This makes it possible to give the user a sense of realism when
outputting a message from the system by voice. [Selected figure] Figure 1
Message output control system
[0001]
The present invention relates to a message output control system.
[0002]
Conventionally, for example, in the navigation system, when voice guidance is performed on a
route to a destination, a message is output as voice from a speaker provided on the driver's seat
side.
This makes it easier for the driver to hear the voice guidance message from the navigation
system.
09-05-2019
1
[0003]
However, if any message is output only from the speaker provided on the driver's seat side, when
the message is output by voice, it will be unrealistic to the user.
[0004]
The present invention has been made in view of the above-described point, and it is an object of
the present invention to provide a message output control system capable of allowing the user to
have a sense of realism when outputting a message from the system by voice. Do.
[0005]
In order to achieve the above object, according to a first aspect of the present invention, there is
provided a message output control system comprising: a plurality of speakers provided at a
plurality of locations in a vehicle compartment; A speaker determining means for determining a
speaker to which a message is to be output from among a plurality of speakers based on the
message generated by the message generating means, and a voice output of the message from
the speaker determined by the speaker determining means And output control means for
[0006]
As described above, in the message output control system according to the first aspect of the
present invention, the speaker that outputs the message is determined from among the plurality
of speakers based on the message.
For this reason, a message can be output as voice from speakers other than the speaker provided
on the driver's seat side.
As a result, the sense of reality is improved as compared to the case where a message is output in
voice from only the speaker provided on the driver's seat side.
This makes it possible to give the user a sense of realism when outputting a message from the
system by voice.
09-05-2019
2
[0007]
As described in claim 2, the message generation means responds to the speech contents of the
user, the speech input means for inputting the speech uttered by the user, the speech recognition
means for recognizing the speech inputted to the speech input means, and A message storage
means for storing a plurality of messages and an extraction means for extracting a message
responding thereto from the message storage means based on the utterance content recognized
by the speech recognition means can be provided. Thereby, for example, when performing
various operations by voice, it is possible to confirm by a message whether or not the operation
intended by the user is performed.
[0008]
As described in claim 3, the speaker determination means includes storage means for storing in
advance correspondence relationships for speakers to be used for each message, and the
message is referred to by referring to the correspondence relationships stored by the storage
means. You can also decide which speakers to output. Thereby, when outputting each message as
an audio, an audio can be reliably output from the speaker considered most appropriate.
[0009]
As described in the fourth aspect, according to the utterance content recognized by the voice
recognition means, the control means is provided to execute predetermined control on the
predetermined control target device, and the message generation means In addition to
generating a message indicating that predetermined control is to be performed, the speaker
determination means preferably determines a speaker provided in the direction in which the
control target device is present as a speaker to which the message should be output. For
example, when the user inputs "open the sunroof" by voice, a message "open the sunroof"
corresponding to the content of the utterance is output by voice from a speaker provided on the
roof of the vehicle interior. Thereby, the user's consciousness can be easily directed in the
direction in which the control target device is present.
[0010]
09-05-2019
3
As described in claim 5, when the message generated by the message generating means includes
information on the direction, the speaker determining means determines a speaker present in the
direction included in the message as a speaker to which the message is to be output. It is good.
For example, when the message output control system according to the present invention is
applied to a navigation system, a voice guidance message "Please turn left at 200 m ahead"
regarding the route is output by voice from a speaker provided on the passenger side. This
makes it possible for the user to be strongly aware of what should be left.
[0011]
Hereinafter, preferred embodiments of the present invention will be described with reference to
the drawings. In this embodiment, an example in which the message output control system is
applied to a speech recognition device will be described. FIG. 1 is a block diagram showing a
schematic configuration of a speech recognition apparatus 100 according to the present
embodiment. Hereinafter, the speech recognition apparatus 100 according to the present
embodiment will be described in detail.
[0012]
As shown in FIG. 1, the speech recognition apparatus 100 includes a speech recognition unit 30
and an output control unit 40. Further, in the voice recognition device 100, a voice input unit 10
used for voice input, a talk switch 20, speakers 50a to 50e, and a control target device 60 are
connected.
[0013]
The voice input unit 10 and the talk switch 20 are devices used for voice input. In the case of
inputting a voice, for example, pressing a push button of the talk switch 20 transmits an input
trigger to a voice recognition unit 30 described later, and when the voice recognition unit 30
receives an input trigger, the voice input unit The mode is changed from 10 to a mode for
accepting voice input.
[0014]
09-05-2019
4
When the user utters, for example, "open the sunroof" in the mode for receiving the voice input,
the voice is converted into a voice signal by the voice input unit 10 and sent to the voice
recognition unit 30. The voice recognition unit 30 recognizes this voice, generates a message
corresponding to the voice, and gives the message to the output control unit 40. For example,
when the input voice is recognized as "open sunroof", a message "open sunroof" is generated in
response. This message is converted into an audio signal, and the output control unit 40 outputs
it from one of the speakers 50a to 50e. This message is composed of a message code to be
described later.
[0015]
The speakers 50a to 50e are provided on the driver's seat side, the front passenger's seat side,
the front, the rear, and the roof side in the vehicle interior, and are used to output the abovementioned message by voice. The control target device 60 is a device for which predetermined
control is performed according to the utterance content recognized by the voice recognition unit
30. The control target device 60 is mounted on a vehicle and includes all devices operable by the
user.
[0016]
Next, the speech recognition unit 30 of the speech recognition apparatus 100 will be described
with reference to FIG. FIG. 2 is a block diagram showing the configuration of the speech
recognition unit 30. As shown in FIG. As shown in the figure, the voice recognition unit 30
includes an AD conversion circuit 31, a recognition program processing unit 32, an acoustic
model storage unit 33, a recognition dictionary storage unit 34, a message code storage unit 35,
and a command code storage unit 36. Ru.
[0017]
The AD conversion circuit 31 receives an analog audio signal input via the audio input unit 10,
and converts this signal into a digitized signal. The converted digital audio signal is transmitted
to the recognition program processing unit 32.
09-05-2019
5
[0018]
The recognition program processing unit 32 uses the acoustic model storage unit 33 and the
recognition dictionary storage unit 34 to convert the digital speech signal into a recognition
word code. First, the recognition program processing unit 32 uses, for example, a method such as
a well-known hidden Markov model stored in the acoustic model storage unit 33, and the
utterance content corresponding to the digital voice signal (hereinafter referred to as recognition
word reading) Analyze
[0019]
The analyzed recognition word reading is compared with the recognition word reading stored in
the recognition dictionary storage unit 34, and the recognition word reading most likely to be
recognized and the recognition word code corresponding to the recognition word reading are
extracted. As shown in FIG. 4, the recognition dictionary storage unit 34 has a configuration in
which a recognition word reading and a recognition word code corresponding to the recognition
word reading in one-to-one correspondence are stored.
[0020]
Further, the recognition program processing unit 32 refers to the message code storage unit 35
and the command code storage unit 36, and extracts a message code and a command code
corresponding to the recognition code. As shown in FIG. 5, the message code storage unit 35 has
a configuration in which a recognition word code and a message code corresponding one-to-one
with the recognition word code are stored. Similarly, as shown in FIG. 6, the command code
storage unit 36 has a configuration in which a recognition word code and a command code
corresponding to the recognition word code in one-to-one correspondence are stored. The
message code is a code that can be identified by the message output unit 43 of the output control
unit 40. By giving the message code to the message output control unit 43, the message
corresponding to the message code can be output. The command code is a code that can be
identified by the control execution unit 44 of the output control unit 40, and outputs a control
signal to the control target device 60 so that control corresponding to the command code is
executed.
09-05-2019
6
[0021]
Then, the recognition program processing unit 32 outputs the message code and the command
code obtained by the above processing to the output control unit 40.
[0022]
Subsequently, the output control unit 40 of the speech recognition apparatus 100 will be
described with reference to FIG.
FIG. 3 is a block diagram showing the configuration of the output control unit 40. As shown in
FIG. As shown in the figure, the output control unit 40 includes a message storage unit 41, an
output destination speaker code storage unit 42, a message output unit 43, and a control
execution unit 44.
[0023]
The message output unit 43 refers to the message storage unit 41 and extracts a message
corresponding to the message code received from the voice recognition unit 30. As shown in FIG.
7, the message storage unit 41 has a configuration in which a message code and a message
corresponding to the message code in one-to-one correspondence are stored. Thereby, for
example, when performing various operations by voice, it is possible to confirm by a message
whether or not the operation intended by the user is performed.
[0024]
Further, the message output unit 43 refers to the output destination speaker code storage unit
42 and extracts an output destination speaker code corresponding to the message code received
from the voice recognition unit 30. As shown in FIG. 8, the configuration of the output
destination code storage unit 35 is a configuration in which a message code and an output
destination speaker code corresponding to the message code are stored. Thus, the speaker of the
output destination corresponding to each message is determined in advance.
[0025]
09-05-2019
7
The message output unit 43 outputs the extracted message via the speaker corresponding to the
extracted output destination speaker code. As shown in FIG. 9, the output destination speaker
code is uniquely associated with each of the plurality of speakers 50a to 50e, and the speakers
50a to 50e to be used can be specified by the speaker code.
[0026]
The control execution unit 44 controls the control target device 60 in response to the command
code received from the voice recognition unit 30. The control target device 60 is, for example, a
sunroof, a trunk, a wiper, or the like. As an example, the control performed on the sunroof is the
opening and closing of the sunroof.
[0027]
With the above-described configuration, according to the present embodiment, the speech
recognition apparatus determines a speaker that outputs a message from among a plurality of
speakers based on the message. For this reason, a message can be output as voice from speakers
other than the speaker provided on the driver's seat side. As a result, the sense of reality is
improved as compared to the case where a message is output in voice from only the speaker
provided on the driver's seat side. This makes it possible to give the user a sense of realism when
outputting a message from the system by voice.
[0028]
Next, processing of the above-described speech recognition apparatus according to this
embodiment will be described with reference to the flowcharts of FIGS. FIG. 10 is a flowchart
showing a main routine of the speech recognition apparatus, and FIG. 11 is a flowchart showing
a routine of speech recognition processing. FIG. 12 is a flowchart showing the message output
control process. The description will be made on the assumption that the voice "open trunk" is
input as a specific example.
[0029]
09-05-2019
8
First, in step S10 of FIG. 10, it is determined whether the talk switch 20 has been operated. If it is
determined that the operation has not been performed, the standby state is entered. On the other
hand, when it is determined that the operation is performed, the process proceeds to step S20. In
step S20, the speech recognition unit 30 is switched to the input mode, and is in a state of
receiving an input of speech. Here, for example, when the voice saying "open the trunk" is input,
the process proceeds to step S30.
[0030]
In step S30, speech recognition processing is performed. The speech recognition process will be
described with reference to the flowchart of FIG.
[0031]
In the flowchart of FIG. 11, first, in step S110, the recognition word reading is analyzed from the
input speech. Then, the recognition word code corresponding to the recognition word reading is
extracted from the recognition dictionary storage unit 34. Here, the speech "open trunk" is
analyzed to be judged as the recognition word reading "open trunk", and the recognition word
code R0002 corresponding to this recognition word reading is extracted. (See Figure 4).
[0032]
In step S120, the message code corresponding to the recognition word code extracted in step
S110 is extracted from the message code storage unit 35. Here, the message code M0002
corresponding to the recognition word code R0002 is extracted (see FIG. 5).
[0033]
In step S130, the command code corresponding to the recognition word code extracted in step
S110 is extracted from the command code storage unit 36. Here, the command code C0002
corresponding to the recognition word code R0002 is extracted (see FIG. 6). In step S140, the
09-05-2019
9
extracted message code and command code are transmitted to the output control unit 40.
[0034]
Returning to the flowchart of FIG. 10, in step S40, a message output control process is
performed. This message output control process will be described using the flowchart of FIG.
[0035]
In the flowchart of FIG. 12, first, in step S210, the message output unit 43 extracts a message
corresponding to the message code from the message storage unit 41. Here, since the message
code is M0002, the message "open trunk" is extracted (see FIG. 7).
[0036]
In step S220, the message output unit 43 extracts an output destination speaker code
corresponding to the message code from the output destination speaker code storage unit 42.
Here, the output destination speaker code S0004 corresponding to the message code M0002 is
extracted (see FIG. 8).
[0037]
In step S230, the message output unit 43 outputs the extracted message by voice via the speaker
corresponding to the extracted output destination speaker code. Here, since the output
destination speaker code corresponds to the speaker (rear) 50d (see FIG. 9), the voice “open the
trunk” is output through the speaker (rear) 50d. Thereby, when outputting each message as an
audio, an audio can be reliably output from the speaker considered most appropriate. In addition,
the user's consciousness can be easily directed backward in the presence of the trunk.
[0038]
09-05-2019
10
In step S240, the control execution unit 44 executes control corresponding to the command
code. Here, the trunk is opened based on command code C0002.
[0039]
The present invention is not limited to the above-described embodiment, and various
modifications can be made without departing from the spirit of the present invention.
[0040]
For example, in the above-described embodiment, an example has been described in which the
speaker of the output destination corresponding to the message from the system corresponding
to the voice input by the user is output as voice.
However, the present invention may be applied when outputting a message (e.g., speech
recognition guidance) from the system even if the user does not input speech. For example, a
message "Please press the talk switch 20 on the left side of the steering wheel and speak" from
the system is output as voice via the speaker (passenger seat) 50b. Thereby, the user's
consciousness can be easily directed to the left direction in which the talk switch 20 is present.
[0041]
Further, in the embodiment described above, the speech recognition apparatus 100 may be used
as an input device of a known navigation system. For example, the user says, "What is a
convenience store? When a voice “is input, a convenience store around the current position of
the vehicle is searched through the navigation system, and the search result is output. As an
example, the message of the search result "5 km ahead, there is a convenience store on the right
side" is output as voice via the speaker (driver's seat) 50a. This makes it possible for the user to
be strongly aware that there is a convenience store on the right side.
[0042]
Furthermore, in the embodiment described above, the speech recognition apparatus 100 may be
used as a speech output apparatus of a known navigation system. For example, a voice guidance
09-05-2019
11
message "Please turn right 100 m ahead" concerning the route from the navigation system is
voiced from the speaker 50a provided on the driver's side. This makes it possible for the user to
be strongly aware of what should be turned right.
[0043]
Furthermore, in the above-described embodiment, the message may be output with different
voice quality for each speaker. For example, the speakers on the driver's side and the passenger's
side output messages with different voice qualities. As a result, the user feels as if two people
talk. This allows the user to more realistically feel as compared to the case of outputting a
message with one voice quality with one fixed speaker.
[0044]
FIG. 1 is a block diagram showing a schematic configuration of a speech recognition apparatus
100 in the present embodiment. It is a block diagram which shows the structure of the speech
recognition part 30 in this embodiment. FIG. 2 is a block diagram showing a configuration of an
output control unit 40 in the present embodiment. It is a table | surface which shows the
correspondence of recognition word reading and a recognition word code in this embodiment. It
is a table | surface which shows the correspondence of a recognition word code and a message
code in this embodiment. It is a table | surface which shows the correspondence of a recognition
word code and a command code in this embodiment. It is a table | surface which shows the
correspondence of a message code and a message in this embodiment. It is a table | surface
which shows the correspondence of a message code | cord and an output destination speaker
code | cord in this embodiment. It is a table | surface which shows the correspondence of an
output destination speaker code | cord and an output destination speaker in this embodiment. It
is a flowchart which shows the main routine of the speech recognition apparatus in this
embodiment. In this embodiment, it is a flowchart which shows the routine of a speech
recognition process. In this embodiment, it is a flow chart which shows a routine of message
output control processing.
Explanation of sign
[0045]
DESCRIPTION OF SYMBOLS 10 ... Speech input part 20 ... Talk switch 30 ... Speech recognition
09-05-2019
12
part 40 ... Message output control part 50a ... Speaker (driver's seat) 50b ... Speaker (front
passenger seat) 50c ... Speaker (front) 50d ... Speaker (rear) 50e ... Speaker ( roof)
09-05-2019
13
Документ
Категория
Без категории
Просмотров
0
Размер файла
21 Кб
Теги
jp2006003450
1/--страниц
Пожаловаться на содержимое документа