close

Вход

Забыли?

вход по аккаунту

?

JP2017034502

код для вставкиСкачать
Patent Translate
Powered by EPO and Google
Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
DESCRIPTION JP2017034502
Abstract: PROBLEM TO BE SOLVED: To provide an apparatus, method and program capable of
providing an image and a sound intended by a communication apparatus on the other side. In a
control unit (101) of a communication apparatus (conference terminal), an acquisition unit (121)
acquires a photographed image. When the specifying unit 123 receives coordinate information
from another communication device, the specifying unit 123 specifies the coordinates of the
image corresponding to the coordinate information based on the correspondence information in
which the coordinates of the display video are associated with the coordinates of the image. The
cutout unit 124 cuts out an image of a region corresponding to the coordinates of the image
specified by the specifying unit 123 in the image. The directivity control unit 125 controls the
directivity of the microphone array 12 so as to direct the directivity to the microphone
corresponding to the coordinates of the image specified by the specifying unit 123 among the
plurality of microphones. The first transmission control unit 126 is a control that transmits
output information including a clipped image indicating the image clipped by the clipping unit
124 and a voice whose directivity is controlled by the directivity control unit 125 to another
communication device. I do. [Selected figure] Figure 5
Communication apparatus, communication method, program and communication system
[0001]
The present invention relates to a communication apparatus, a communication method, a
program and a communication system.
[0002]
03-05-2019
1
As one form of a communication system for transmitting and receiving data between a plurality
of communication devices, a video (television) conference system that realizes a remote
conference using a communication network is known.
In this video conference system, in a conference room where one party such as an attendee who
performs remote conference is present, an image and a speech etc of a conference room such as
a conference party using a communication device (terminal device) of the remote video
conference system Voice is converted to digital data and transmitted to the other party's
communication device, the image is displayed on the display of the other party's conference
room, and the voice is output from the speaker, so that the remote conference can be performed
in a state close to an actual conference. It can be performed.
[0003]
In such video conferencing systems, microphones are used to capture the audio of conference
participants, and cameras are used to capture video. Because the camera has an angle of view, it
is not possible to capture images of conference participants outside the angle of view of the
camera. In order to solve this problem, a method is known that uses a panoramic camera capable
of acquiring a 360-degree panoramic image. On the other hand, because ordinary microphones
are nondirectional, they will also collect surrounding sounds other than the participants'
comments. In order to solve this problem, it is known to use a microphone array to give
directivity to the sound collection range of the microphone, thereby suppressing the sound
collection of surrounding sounds and clearing the participants' speech. For example, Patent
Document 1 discloses a method in which a conference terminal forms a sound collection beam of
a microphone array in accordance with an arrangement pattern of a selected participant. Further,
Patent Document 2 discloses a technique of detecting the direction of a speaker using a
microphone array configured by arranging a plurality of microphones, and tracking the shooting
direction of the camera in that direction.
[0004]
The above panoramic camera is combined with a microphone array to shoot the entire
conference room with the panoramic camera, and when there is a speaker, the video of the
speaker is automatically closed up and displayed, and the direction of the speaker There is a way
to realize a more realistic video conference by pointing the sound collection range of the
03-05-2019
2
microphone to. However, with this combination, the speaker's video can be displayed close up
and displayed, and the sound collection range of the microphone can be directed to the direction
of the speaker. , I do not necessarily want to output that voice. Therefore, in this combination,
there is a problem that it is difficult for the other party's communication device to provide the
intended image and sound.
[0005]
The present invention has been made in view of the above, and it is an object of the present
invention to provide a communication apparatus, a communication method, a program and a
communication system capable of providing an image and a sound intended by a communication
apparatus on the other side. Do.
[0006]
In order to solve the problems described above and achieve the object, the present invention is a
communication device, wherein the other communication is performed from an acquisition unit
that acquires a captured image and another communication device that communicates with the
communication device. When coordinate information indicating the coordinates of a specified
range of the display video displayed on the output device connected to the device is received,
based on the correspondence information in which the coordinates of the display video are
associated with the coordinates of the image An identifying unit that identifies the coordinates of
the image corresponding to the coordinate information; a clipping unit that clips out an image of
a region of the image that corresponds to the coordinates of the image identified by the
identifying unit; Controlling the directivity of the microphone array including the plurality of
microphones so as to direct the directivity to the microphone corresponding to the coordinates of
the image specified by the specifying unit among the plurality of microphones mounted in a
distributed manner A control unit configured to transmit output information including a
flexibility control unit, a clipped image indicating an image clipped by the clipping unit, and
sound whose directivity is controlled by the directivity control unit to the other communication
device; And a transmission control unit.
[0007]
According to the present invention, it is possible to provide an image and a sound intended by
the communication apparatus on the other party side.
[0008]
03-05-2019
3
FIG. 1 is a diagram showing an example of the configuration of a communication system.
FIG. 2 is a diagram illustrating an example of a hardware configuration of the conference
terminal.
FIG. 3 is a diagram showing an example of installation of a microphone.
FIG. 4 is a diagram for explaining the details of the conference terminal. FIG. 5 is a diagram
illustrating an example of the function of the control unit. FIG. 6 is a schematic view showing an
example of the state of the video conference in the embodiment. FIG. 7 is a diagram for
describing a method of designating a cutout range of a display video displayed on the interactive
whiteboard. FIG. 8 is a diagram for explaining a specific example of the embodiment. FIG. 9 is a
diagram for explaining a specific example of the embodiment. FIG. 10 is a flowchart showing an
operation example of the interactive whiteboard. FIG. 11 is a flowchart showing an operation
example of the conference terminal. FIG. 12 is a flowchart showing an operation example of the
conference terminal. FIG. 13 is a flowchart showing an operation example of the conference
terminal.
[0009]
Hereinafter, embodiments of a communication apparatus, a communication method, a program,
and a communication system according to the present invention will be described in detail with
reference to the attached drawings.
[0010]
FIG. 1 is a diagram showing an example of the configuration of the communication system 1 of
the present embodiment.
In the example of FIG. 1, one conference terminal 2 which is an example of a communication
device is disposed at each of the base A and the base B. The conference terminals 2 disposed at
each of the base A and the base B are connected to the server 4 via the network 3 such as the
Internet. The present invention is not limited to this, and the number of conference terminals 2
(the number of bases) included in the communication system 1 can be arbitrarily changed.
03-05-2019
4
[0011]
The server 4 monitors whether or not each conference terminal 2 is connected to the server 4,
and performs necessary control such as call control of the conference terminal 2 at the start of
the conference. At the time of a conference, the conference terminal 2 transmits image and audio
data to the server 4 at the time of data transmission of its own device, and the server 4 transmits
image and audio data to another conference terminal 2 on the other side. Do. At the time of data
reception, image / voice data of the other conference terminal 2 on the opposite side is received
via the server 4. For example, when a meeting is held at the base A and the base B, the data
transmitted by the conference terminal 2 at the base A is transmitted to the conference terminal
2 at the base B via the server 4 and another conference terminal 2 Not sent to the conference
terminal 2) which is not Similarly, the data transmitted by the conference terminal 2 at the site B
is transmitted to the conference terminal 2 at the site A participating in the conference via the
server 4 and to the other conference terminals 2 not participating in the conference. Not sent By
performing the control as described above, it is possible to conduct a conference between a
plurality of conference terminals 2 (a plurality of bases).
[0012]
Next, the configuration of the conference terminal 2 will be described. In addition, since the
structure of the meeting terminal 2 arrange | positioned at each of the base A and the base B is
the same, in the following description, the one meeting terminal 2 is mentioned as an example,
and is demonstrated. FIG. 2 is a diagram showing an example of the hardware configuration of
the conference terminal 2. As shown in FIG. 2, the conference terminal 2 includes a panoramic
camera 10, a display unit 11, a microphone array 12, a speaker 13, a CPU 14, a storage device
15, a memory 16, a LAN I / F unit 17, and an operation unit 18.
[0013]
The panoramic camera 10 generates a panoramic image obtained by photographing 360 degrees
around the panoramic camera 10 (which may be considered as 360 degrees around the
conference terminal 2), and transmits the generated panoramic image to the CPU 14. For
example, the panoramic camera 10 is configured of a known omnidirectional camera or the like.
By capturing 360 degrees around the panoramic camera 10, it is possible to capture all the
conference participants present around the panoramic camera 10. Here, the panoramic image
03-05-2019
5
means an image generated by combining a plurality of images by moving a camera having a
plurality of imaging elements, moving one camera, or photographing using a plurality of
cameras. In addition, although the imaging | photography area | region of a panoramic image is
made into 360 degree, you may narrow than this. Further, although a panoramic image is used,
an image with a narrower angle of view may be used, but a panoramic image is more preferable.
[0014]
The display unit 11 has a function of displaying the video data received from the CPU 14 on the
screen. In this example, the display unit 11 is configured of a liquid crystal display device or the
like.
[0015]
The microphone array 12 includes a plurality of microphones distributedly installed in the
conference terminal 2 and has a function of acquiring voices of conference participants and
transmitting the voices to the CPU 14. The microphone array is composed of a plurality of
omnidirectional microphones (microphones) and forms directivity by a beam former. In general,
the beam former is a method of forming directivity by utilizing the time difference of the sound
reaching each microphone. The directivity formed by the beam former can be freely directed in
the vertical and horizontal directions by the arrangement of the microphones. Therefore, it is
possible to arbitrarily change the area (the target area) where sound collection is desired.
[0016]
FIG. 3 is a diagram showing an example in which six microphones a to f included in the
microphone array 12 are installed in the case of the conference terminal 2. The CPU 14 can
collect voices in an arbitrary range by switching on / off of the respective microphones and
integrating the voices of the respective microphones. Further, the correspondence between each
microphone and the installation location in the case of the conference terminal 2 is stored in
advance in the storage device 15 of the conference terminal 2.
[0017]
03-05-2019
6
The description of FIG. 2 is continued. The speaker 13 has a function of outputting audio data
received from the CPU 14. The CPU 14 centrally controls the operation of the entire conference
terminal 2. For example, the CPU 14 has a function to control a video conference, encodes video
acquired from the panoramic camera 10 and audio data acquired from the microphone array 12
and transmits it to the LAN I / F unit 17 and receives it by the LAN I / F unit 17 It has a CODEC
function or the like that decodes video and audio data of the other party of the conference and
transmits it to the display unit 11 and the speaker 13. As an example of the CODEC used by the
CPU 14, H. 2. H.264 / AVC and H.264. There are H.264 / SVC. Furthermore, the CPU 14 has a
function of controlling the directivity of the microphone array 12, a function of close-up and
displaying the speaker among the conference participants reflected in the panoramic image
acquired from the panoramic camera 10, and the like.
[0018]
The storage device 15 stores various control programs (video conference control and the like)
executed by the CPU 14, a conversion table described later, and the like. Examples of the storage
device 15 include non-volatile storage media such as flash memory and HDD.
[0019]
The memory 16 develops a program to be executed by the CPU 14 and temporarily stores
operation data. Examples of the memory 16 include volatile memory such as DDR memory. The
LAN I / F unit 17 is connected to another conference terminal 2 via the network 3 to transmit
and receive data (image / voice data). The LAN I / F unit 17 supports 10Base-T, 100Base-TX, and
1000Base-T, and is a wired LAN connected to Ethernet (registered trademark), and a wireless
LAN compatible with 802.11a / b / g / n / ac. Are mentioned as an example.
[0020]
The operation unit 18 is a device used for various operations by the user (various operations
related to device control of the conference terminal 2), and includes, for example, a keyboard,
buttons, and the like.
[0021]
03-05-2019
7
FIG. 4 is a diagram for explaining the details of the conference terminal 2.
The conference terminal 2 is connected to an interactive whiteboard 300 which is an example of
an output device having a coordinate detection function. In this example, one interactive
whiteboard 300 having the same configuration is disposed at each of the base A and the base B.
The display control unit 102 has a function of controlling display. In this example, the display
control unit 102 has a function of acquiring coordinate information described later from the
interactive whiteboard 300 and transmitting the acquired coordinate information to the control
unit 101 as well as the function related to screen drawing and output. . Here, although the
interactive whiteboard is used, any output device having a coordinate detection function such as
a tablet, a notebook computer, a touch panel display, etc. may be used.
[0022]
The panoramic image generated by the panoramic camera 10 is sent to the camera I / F receiver
111. The camera I / F receiver 111 is assumed to be a high-speed serial I / F such as VbyOne
(registered trademark) or HDMI (registered trademark). The distortion correction processing unit
113 corrects the distortion of the panoramic image that has been subjected to serial / parallel
conversion by the camera I / F receiver 111, and is output to the control unit 101. The lens
characteristic storage unit 112 stores conversion parameters for performing distortion
correction in accordance with the lens characteristics, and the distortion correction processing
unit 113 can correct distortion of a panoramic image using the conversion parameters.
[0023]
The network unit 100 transmits and receives data to and from another conference terminal 2
which is the other party of the conference.
[0024]
The control unit 101 is a unit that performs various controls and calculations, and corresponds
to the CPU 14.
FIG. 5 is a diagram illustrating an example of the function of the control unit 101. For
03-05-2019
8
convenience of explanation, FIG. 5 mainly illustrates the functions related to the present
embodiment, but the functions possessed by the control unit 101 are not limited to these.
[0025]
As shown in FIG. 5, the control unit 101 includes an acquisition unit 121, an output control unit
122, a specifying unit 123, a cutout unit 124, a directivity control unit 125, a first transmission
control unit 126, And a transmission control unit 127.
[0026]
The acquisition unit 121 acquires a captured image.
More specifically, the acquisition unit 121 acquires a panoramic image obtained by
photographing 360 degrees around the conference terminal 2. In this example, the acquisition
unit 121 acquires the panoramic image after correction input from the distortion correction
processing unit 113 described above.
[0027]
The output control unit 122 performs control to output the image / sound data received from the
other conference terminal 2 (hereinafter, may be simply referred to as “other conference
terminal 2”) as the conference partner. The output control unit 122 causes the display control
unit 102 to display an image received from another conference terminal 2 on the interactive
whiteboard 300 according to the current display mode of the interactive whiteboard 300 (the
display mode can be arbitrarily changed). Control to output audio data received from another
conference terminal 2 from the speaker 13.
[0028]
The specifying unit 123 specifies a range designated by the display video displayed on the
interactive whiteboard 300 connected to the other conference terminal 2 from the other
conference terminal 2 (in the following description, it is referred to as “cut-out range” When
the coordinate information indicating the coordinates of (a) is received, it corresponds to the
03-05-2019
9
received coordinate information based on correspondence information in which the coordinates
of the display image and the coordinates of the photographed image (a panoramic image in this
example) are associated. The coordinates of the image to be taken (in this example, a panoramic
image) are identified. In this example, the correspondence information is stored in advance in the
storage device 15. In addition, a normal video conference system can display only the video of its
own device, can only display the video of the other party, and can freely change the layout
(change the display mode), so it is displayed on the interactive whiteboard 300. The relationship
between the coordinates of the display image and the coordinates of the captured image (in this
example, a panoramic image) does not necessarily correspond one to one. Therefore, the
correspondence information in this example associates, for each display mode (layout
information) of the interactive whiteboard 300, the coordinates of the display video with the
coordinates of the photographed image (in this example, a panoramic image).
[0029]
The cutout unit 124 cuts out an image of a region corresponding to the coordinates of the image
(in this example, the panoramic image) specified by the specifying unit 123 among the images (in
this example, the panoramic image) acquired by the acquisition unit 121.
[0030]
The directivity control unit 125 directs directivity to the microphone corresponding to the
coordinates of the image (in this example, the panoramic image) specified by the specifying unit
123 among the plurality of microphones distributed and installed in the conference terminal 2 ,
Control the directivity of the microphone array 12.
In this example, the directivity control unit 125 is specified by the specifying unit 123 based on
position information indicating the relationship between the position of the microphone included
in the microphone array 12 and the coordinates of the captured image (a panoramic image in
this example). A microphone may be determined that corresponds to the coordinates of the
captured image (in this example, a panoramic image). The position information may be stored,
for example, in the storage device 15 or the like.
[0031]
The first transmission control unit 126 transmits, to another conference terminal 2, output
03-05-2019
10
information including a cutout image indicating the image cut out by the cutout unit 124 and a
voice whose directivity is controlled by the directivity control unit 125. Take control. When the
first transmission control unit 126 has not received the above-described coordinate information
from another conference terminal 2, the first transmission control unit 126 reflects the image
acquired in the acquisition unit 121 (a panoramic image in this example) and the image. Control
is performed to transmit, to another conference terminal 2, normal conference information
including a clipped image obtained by clipping out a speaker among the in-room conference
participants and audio data collected by the microphone array 12.
[0032]
Moreover, in this example, when the above-mentioned output control part 122 receives the
above-mentioned normal meeting information from the other meeting terminal 2, it performs
control which outputs the received normal meeting information. When the output control unit
122 receives the above-described output information from another conference terminal 2, the
output control unit 122 performs control to output the received output information.
[0033]
When the second transmission control unit 127 receives the above-described coordinate
information from the display control unit 102 of the own device, the second transmission control
unit 127 performs control to transmit the received coordinate information to another conference
terminal 2.
[0034]
The functions of the control unit 101 described above (acquisition unit 121, output control unit
122, identification unit 123, cutout unit 124, directivity control unit 125, first transmission
control unit 126, second transmission control unit 127) The CPU 14 may be realized by
executing a program stored in the storage device 15 or the like, or at least a part of the functions
of the control unit 101 may be realized by a dedicated hardware circuit.
[0035]
Moreover, in the example of FIG. 4, although the panoramic camera 10 and the speaker 13 are
contained in the meeting terminal 2, not only this but the form provided, for example out of the
meeting terminal 2 may be sufficient.
03-05-2019
11
[0036]
FIG. 6 is a schematic view showing an example of the state of the video conference according to
the present embodiment.
A conference terminal 2 is placed at a desk at the base 1.
Since the conference terminal 2 is equipped with the panoramic camera 10, it is assumed that it
is placed at the center of the desk.
As described above, the conference terminal 2 has the microphone array 12 including the
microphones a to f. Base 1 shows that five persons D to H participate in the video conference. In
addition, the above-mentioned interactive whiteboard 300 is provided at the site 1, and the
image on the site 2 side is displayed. Under the display image displayed on the interactive
whiteboard 300, a panoramic image in which all persons participating in the conference at the
base 2 (in this example, persons A to C) are displayed is displayed. Further, above the display
image, a cutout image in which the person A who is the speaker is enlarged is displayed. In the
case of a video conference system equipped with a normal speaker following function, when one
of the conference participants speaks, the speaker is displayed in a close-up on the cutout image.
[0037]
The conference terminal 2 is also placed on the desk at the site 2. As in the case of the base 1,
since the conference terminal 2 is equipped with the panoramic camera 10, it is assumed that it
is placed at the center of the desk, and the microphone array 12 including the microphones a to f
is mounted. . Base 2 indicates that three persons A to C participate in the video conference.
Similar to the base 1, the base 2 is equipped with the above-mentioned interactive whiteboard
300, and the image on the base 1 side is projected. Below the display video displayed on the
interactive whiteboard 300, a panoramic image in which all the persons participating in the
conference at the base 1 (in this example, the persons D to H) are displayed. Here, normally, the
speaker is displayed in the cutout image, but as shown in FIG. 6, any person participating in the
conference at the base 2 is surrounded by a dotted line using a stylus pen or the like. When the
selected area is designated, the cut-out image of the designated area is displayed on the display
video. That is, it is possible to display the cut-out image of the person on the interactive
whiteboard 300 on the base 2 side by designating the area where the person who wants to close
03-05-2019
12
up is reflected among the conference participants who are reflected in the panoramic image. . In
addition to the person, it is possible to display the cutout image of the subject by designating an
area in which the subject to be focused is included among the subjects included in the panoramic
image.
[0038]
A method of designating the cut-out range of the display video displayed on the interactive
whiteboard 300 will be described with reference to FIG. When the interactive whiteboard 300
receives the pressing of the start icon 310 for instructing the start of the specification of the cutout range, the interactive whiteboard 300 displays a pop-up screen for confirming the start of
the specification of the cut-out range as shown in FIG. Control to display. Then, when the
interactive whiteboard 300 receives the pressing of the “OK button” on the pop-up screen, the
interactive whiteboard 300 performs control to display a message prompting specification of a
cutout range as shown in FIG. 7B. After that, the user can perform an operation of specifying the
cutout range with a stylus pen or the like. Then, as shown in FIG. 7C, when the interactive
whiteboard 300 receives the pressing of the end icon 320 for instructing the end of the
specification of the cutout range, the display displayed on the interactive whiteboard 300 is
performed. Coordinate information indicating the coordinates of a designated clipping range (a
rectangle surrounded by a dotted line in the example of FIG. 7) in the image is transmitted to the
conference terminal 2. Then, the conference terminal 2 performs control to transmit the range
information received from the interactive whiteboard 300 to another conference terminal 2 as a
conference partner.
[0039]
Now, as shown in FIG. 8, it is assumed that a region 330 in which a person H is reflected in the
display video displayed on the interactive whiteboard 300 on the base 2 side is designated as the
cutout range. In this example, the area 330 is a rectangular area, and the coordinate information
of the area 330 is information indicating the coordinates of each of the four vertices (A, B, C, D)
of the area 330. In this example, the coordinates of the vertex A in the display image are (Xa, Ya),
the coordinates of the vertex B are (Xb, Yb), the coordinates of the vertex C are (Xc, Yc), and the
coordinates of the vertex D is (Xd, Yd) It is. This coordinate information is transmitted to the
conference terminal 2 on the base 1 side.
[0040]
03-05-2019
13
The conference terminal 2 on the base 1 side coordinates the coordinates of the panoramic
image corresponding to the coordinate information received from the conference terminal 2 on
the base 2 side based on the correspondence information in which the coordinates of the display
video and the coordinates of the panoramic image are associated. Identify. Here, among the
panoramic images acquired by the conference terminal 2 on the base 1 side, a point
corresponding to the vertex A of the display video displayed on the interactive whiteboard 300
on the base 2 side is A ′ and a point corresponding to the vertex B A point corresponding to B
′ and vertex C is represented as C ′ and a point corresponding to vertex D is represented as D
′. FIG. 9 is a diagram showing a correspondence relationship between each coordinate of
vertices A, B, C, and D in a display image and each coordinate of A ', B', C ', and D' in a panoramic
image. The conference terminal 2 on the base 1 side identifies the coordinates of the panoramic
image corresponding to the coordinate information received from the conference terminal 2 on
the base 2 side, and cuts out the image of the area corresponding to the identified coordinates in
the panoramic image. Cut out as an image. In this example, as shown in FIG. 8, the conference
terminal 2 on the base 1 side has coordinates (Xa ′, Ya ′) of the coordinates A ′ among the
panoramic images acquired from the panoramic camera 10 mounted on the own device. Cut out
an image of a rectangular area surrounded by the coordinates of B '(Xb', Yb '), the coordinates of
C' (Xc ', Yc'), and the coordinates of D '(Xd', Yd ') Cut out as an image. In addition, the conference
terminal 2 on the base 1 side is disposed at a position closest to the specified coordinates of the
panoramic image based on the position information indicating the relationship between the
position of the microphone included in the microphone array 12 and the coordinates of the
panoramic image. The directivity of the microphone array 12 is controlled so as to direct the
directivity to the microphone.
[0041]
Then, the conference terminal 2 on the base 1 side transmits, to the conference terminal 2 on the
base 2 side, output information including the clipped image cut out as described above and the
voice whose directivity is controlled. The conference terminal 2 on the base 2 side outputs the
output information received from the conference terminal 2 on the base 1 side. As a result, on
the base 2 side, the person H is closed up.
[0042]
FIG. 10 is a flowchart showing an example of the operation of the interactive whiteboard 300 in
03-05-2019
14
the case of specifying the cutout range. When pressing of the start icon 310 is received (step S1:
Yes), an operation for specifying the clipping range is received (step S2). Then, when pressing of
the end icon 320 is received (step S3: Yes), coordinate information indicating the coordinates of
the designated clipping range is transmitted to the conference terminal 2 (step S4).
[0043]
FIG. 11 is a flowchart showing an operation example of the conference terminal 2 in the case
where coordinate information is received from the interactive whiteboard 300 connected to the
apparatus itself. When the coordinate information is received (step S5: Yes), the second
transmission control unit 127 performs control to transmit the received coordinate information
to the other conference terminal 2 (step S6).
[0044]
FIG. 12 is a flowchart showing an operation example of the conference terminal 2 when
coordinate information is received from another conference terminal 2. When the coordinate
information is received from the other conference terminal 2 (step S10: Yes), the specifying unit
123 specifies the coordinates of the panoramic image corresponding to the received coordinate
information based on the correspondence information described above (step S11). ). Next, the
cutout unit 124 cuts out an image of a region corresponding to the coordinates of the panoramic
image identified in step S11 among the panoramic images acquired from the panoramic camera
10 mounted on the own device (step S12). Next, the directivity control unit 125 sets the
directivity of the microphone array 12 to direct the directivity to the microphone corresponding
to the coordinates of the panoramic image specified in step S11 among the plurality of
microphones dispersedly mounted on the conference terminal 2. The directivity is controlled
(step S13). Then, the first transmission control unit 126 transmits to the other conference
terminal 2 output information including the cut-out image indicating the image cut out in step
S12 and the voice whose directivity is controlled as a result of step S13. (Step S14).
[0045]
FIG. 13 is a flowchart showing an operation example of the conference terminal 2 when output
information is received from another conference terminal 2. When the output information is
received from another conference terminal 2 (step S20: Yes), the output control unit 122
03-05-2019
15
performs control to output the received output information (step S21). The output control unit
122 performs control to display the cut-out image included in the output information on the
interactive whiteboard 300, and performs control to output sound included in the output
information from the speaker 13.
[0046]
As described above, when the conference terminal 2 of the present embodiment receives the
above-mentioned coordinate information from another conference terminal 2 as a conference
partner, the coordinates of the display video are associated with the coordinates of the
panoramic image. The coordinates of the panoramic image corresponding to the received
coordinate information are specified based on the correspondence information, and an image of
the region corresponding to the coordinates of the specified panoramic image in the panoramic
image is cut out as a cutout image. Further, the directivity of the microphone array 12 is
controlled to direct the directivity to the microphone corresponding to the coordinates of the
specified panoramic image among the plurality of microphones dispersedly mounted on the
conference terminal 2. Then, the conference terminal 2 transmits the output information
including the cutout image and the voice whose directivity is controlled to the other conference
terminal 2, and the other conference terminal 2 outputs the received output information. The
video and audio intended by the other conference terminal 2 can be output. In the present
embodiment, as the panoramic camera, the range of the imaging region is 360 degrees around.
However, the gist of the present invention is to designate a part of a photographed image as a
cutout region and control the directivity of the region and the microphone. Therefore, as an
imaging area, the angle of view of the camera may be about 80 degrees.
[0047]
As mentioned above, although embodiment concerning this invention was described, this
invention is not limited to the above-mentioned each embodiment as it is, In an execution phase,
in the range which does not deviate from the summary, a component is changed and can be
materialized. . In addition, various inventions can be formed by appropriate combinations of a
plurality of components disclosed in the above-described embodiments. For example, some
components may be deleted from all the components shown in the above-described
embodiments.
[0048]
03-05-2019
16
The program executed by the above-described conference terminal 2 is a file of an installable
format or an executable format, and is a CD-ROM, a flexible disk (FD), a CD-R, a DVD (Digital
Versatile Disk), and a USB (Universal) It may be configured to be recorded and provided in a
computer readable recording medium such as Serial Bus), or may be configured to be provided or
distributed via a network such as the Internet. In addition, various programs may be configured
to be provided by being incorporated in a ROM or the like in advance.
[0049]
Reference Signs List 1 communication system 2 conference terminal 3 network 4 server 10
panoramic camera 11 display unit 12 microphone array 13 speaker 14 CPU 15 storage device
16 memory 17 LAN I / F unit 18 operation unit 100 network unit 101 control unit 102 display
control unit 111 camera I / F receiver 112 lens characteristic holding unit 113 distortion
correction processing unit 121 acquisition unit 122 output control unit 123 identification unit
124 cutout unit 125 directivity control unit 126 first transmission control unit 127 second
transmission control unit 300 interactive whiteboard 300
[0050]
Patent document 1: JP 2007-274463 A Patent document 5029844
03-05-2019
17
Документ
Категория
Без категории
Просмотров
0
Размер файла
30 Кб
Теги
jp2017034502
1/--страниц
Пожаловаться на содержимое документа