JP2010200212

Patent Translate
Powered by EPO and Google
Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
DESCRIPTION JP2010200212
An object of the present invention is to allow a viewer to view suitable video and audio regardless
of their own location. According to a client signal measurement result supplied from a wireless
node WN, a server 1 detects the position of a client unit CU. The server 1 variably sets a
parameter value for converting at least one of the audio signal and the video signal based on the
position of the client unit CU, and converts the signal using the parameter value. Among the
converted signals, the video signal is output to the very large screen monitor 2, and the audio
signal of each channel is output to each of the speakers 3 to 7. The present invention is
applicable to a video and audio distribution system. [Selected figure] Figure 1
INFORMATION PROCESSING APPARATUS AND METHOD, AND PROGRAM
[0001]
The present invention relates to an information processing apparatus and method, and a
program, and more particularly, to an information processing apparatus and method, and a
program that allow a viewer to view suitable video and audio regardless of their own location.
About.
[0002]
Conventionally, in order to output video and audio in a wide range such as an event hall, there
have been cases in which a super large screen monitor and a multi-channel speaker have been
arranged.
09-05-2019
1
In such a case, multi-channel audio signals are converted to audio signals of relatively small
channels such as 2 channels or 5.1 channels. And the audio ¦ voice corresponding to the audio ¦
voice signal of each channel is each output from the speaker of a corresponding channel. (See,
for example, Patent Document 1).
[0003]
JP, 2006-108855, A
[0004]
However, in a wide range such as an event venue, the viewer may not be able to view suitable
video and audio depending on the location.
[0005]
The present invention has been made in view of such a situation, and enables a viewer to view
suitable video and audio regardless of the position where the viewer is present.
[0006]
An information processing apparatus according to one aspect of the present invention is a
position detection unit that detects the position of the client unit based on a signal output by a
client unit held by a user, and the client unit detected by the position detection unit. Based on the
position, a parameter value for converting at least one of the audio signal and the video signal is
variably set, and conversion means for converting the signal using the parameter value; And
output means for outputting the signal.
[0007]
The conversion means variably sets parameter values for determining the mixing ratio of multichannel audio signals, and converts the audio signals using the parameters.
[0008]
The position detection means detects information indicating a divided area where the client unit
is located among a plurality of divided areas divided from a predetermined area, and the
conversion means detects the information detected by the position detection means. The
parameter value is variably set based on the information.
09-05-2019
2
[0009]
The conversion means variably sets a parameter value for determining an enlargement ratio of
an image corresponding to the image signal itself or a character related to the image, and
converts the image signal using the parameter.
[0010]
The position detection means detects the position of the client unit as a time variable based on
the time transition of the signal output by the client unit.
[0011]
The conversion means maintains the setting of the parameter value when the position detection
means detects that the position of the client unit has not changed.
[0012]
Each of an information processing method and a program according to an aspect of the present
invention is a method and a program corresponding to the information processing apparatus
according to the aspect of the present invention described above.
[0013]
In an information processing apparatus, method, and program according to one aspect of the
present invention, an information processing apparatus that outputs at least one of an audio
signal and a video signal as an output signal, or at least one of an audio signal and a video signal.
The position of the client unit is detected based on the signal output by the client unit held by the
user by a computer that controls an output device that outputs an output signal as the output
signal, and the position of the client unit is detected based on the detected position. A parameter
value for converting a signal from which an output signal is generated is variably set, the signal is
converted using the parameter value, and the converted signal is output as the output signal.
[0014]
As described above, according to the present invention, a viewer can view and listen to suitable
video and audio regardless of the position where the viewer is present.
[0015]
09-05-2019
3
1 is a configuration example of an information processing system to which the present invention
is applied.
FIG. 1 is a block diagram showing a configuration of an embodiment of an information
processing system to which the present invention is applied.
It is a flow chart explaining sound signal output processing.
It is a figure for demonstrating the audio ¦ voice signal output process in the audio ¦ voice signal
output device to which this invention is applied.
It is a figure which shows the structural example of one Embodiment of the client unit CU of the
audio ¦ voice signal output device to which this invention is applied.
FIG. 16 is a block diagram showing a configuration example of a computer included in an audio
signal control device to which the present invention is applied or that controls the drive thereof.
[0016]
Hereinafter, as an embodiment of an information processing system to which the present
invention is applied, a first embodiment and a second embodiment will be described.
The description will be made in the following order.
1.
First Embodiment (Client Unit CU: Example Composed of Only Wireless Tags)
Second Embodiment (Client Unit CU: Example Composed of Wireless Tag Headphones and a
09-05-2019
4
Wireless Tag Monitor)
[0017]
〈1.
First embodiment> [Configuration example of information processing system to which the
present invention is applied] Fig. 1 is a view showing a configuration example of an information
processing system to which the present invention is applied.
[0018]
The information processing system includes a server 1, a super large screen monitor 2, speakers
3 to 7, wireless nodes WN 1 to WNK (K is an integer value of 1 or more).
In the example of FIG. 1, K = 9) and client units CU1 to CUM (M is an integer value indicating the
number of users).
In the example of FIG. 1, M = 4) is included.
[0019]
Information processing systems are built in a wide range of areas, such as event venues.
[0020]
In the example of FIG. 1, the server 1 and the super large screen monitor 2 are disposed at the
upper side in FIG.
Hereinafter, the upward direction in FIG. 1, that is, the direction in which the user views the
super large screen monitor 2 is referred to as the forward direction. Further, the lower direction
09-05-2019
5
in FIG. 1 is referred to as the rear direction, the left direction in FIG. 1 is referred to as the left
direction, and the right direction in FIG. 1 is referred to as the right direction. As a matter of
course, the arrangement position of the server 1 is not limited to the position of the example of
FIG. 1 and may be any position.
[0021]
For example, a circular area α formed facing the front of the super large screen monitor 2
(display surface of the super large screen monitor 2) is a range where the user can view the
image displayed on the super large screen monitor 2 It is assumed that Hereinafter, such a
region α is referred to as a target region. It is needless to say that the target area α is a design
item that the builder of the information processing system may freely determine, and is not
limited to the target area α in FIG. The speakers 3 to 7 are respectively arranged on the
boundary (circumferential) of the target area α. Specifically, the speaker 3 is disposed at the
front left, the speaker 4 at the front right, the speaker 5 at the rear right, the speaker 6 at the
center rear, and the speaker 7 at the left rear. ing.
[0022]
The radio nodes WN1 to WN9 are arranged at equal intervals in three rows vertically and three
rows horizontally from the front.
[0023]
The wireless nodes WN1 to WN9 may be disposed at a plurality of locations in the target area α,
and are not limited to the arrangement and the number of the example in FIG.
[0024]
Client unit CUK (K is 1 to M).
M is the maximum number of viewers) is held by a user (not shown).
For example, in the example of FIG. 1, M = 4. That is, in the example of FIG. 1, one of the client
units CU1 to CU4 is held by each of the four viewers. As described later, when the client unit CUK
09-05-2019
6
is located in the target area α, the server 1 detects the position. This detected position indicates
the existing position of the user holding the client unit CUK.
[0025]
The server 1 outputs the input video signal to the very large screen monitor 2. The super large
screen monitor 2 displays a video corresponding to the video signal. The viewer present in the
target α visually recognizes the video displayed on the super large screen monitor 2.
[0026]
Further, multi-channel audio signals are input to the server 1. In the present embodiment, the
server 1 converts the input multi-channel audio signal into a 5.1-channel audio signal. In the
present embodiment, the 5.1 channel audio signal is composed of a stereo signal LO, a stereo
signal R0, a right surround signal Rs, a center channel signal C, and a left surround signal Ls.
[0027]
In the initial state, the 5.1 channel audio signal is supplied as follows. The stereo signal LO is
supplied to the speaker 3, the stereo signal R0 to the speaker 4, the right surround signal Rs to
the speaker 5, the center channel signal C to the speaker 6, and the surround signal Ls to the
speaker 7.
[0028]
That is, in the initial state, the speaker 3 outputs a sound corresponding to the stereo signal LO,
and the speaker 4 outputs a sound corresponding to the stereo signal R0. The speaker 5 outputs
sound corresponding to the right surround signal Rs, the speaker 6 outputs sound corresponding
to the center channel signal C, and the speaker 7 outputs sound corresponding to the left
surround signal Ls. .
[0029]
09-05-2019
7
Thus, in the initial state, the speakers 3 to 7 merely output the conventional 5.1 channel sound.
Therefore, when the viewer is present at an optimal listening point near the center of the target
area α, the user can listen to the optimal audio. Here, "optimum" in the "optimum listening point"
means that it is optimum when the conventional 5.1 channel audio is simply output. That is, as
described later, it should be noted that when the present invention is applied, any position of the
target area α is an optimal listening point for the user. Therefore, hereinafter, the optimal
listening point when the conventional 5.1 channel audio is simply output is referred to as the
conventional optimal listening point.
[0030]
By the way, since the target area α is a wide area such as an event hall, the viewer is not
necessarily positioned at the conventional optimum listening point. Therefore, when the viewer is
not positioned at the conventional optimum listening point, as described above in the section of
the "Problems to be solved by the invention", it is not possible to listen to a suitable voice.
[0031]
Therefore, in the present embodiment, the server 1 performs control to change the state of each
sound output from the speakers 3 to 7 in accordance with the presence position of the viewer.
That is, when the user's existing position is other than the conventional optimum listening point
position, control is performed to shift the state of each sound output from the speakers 3 to 7 to
a state different from the initial state. In order to realize this control, first, the server 1 needs to
detect the presence position of the viewer. Therefore, the server 1 has a function of detecting the
position of the client unit CUK, in other words, a function of detecting the existing position of the
viewer holding the client unit CUK. Hereinafter, this function is referred to as a client unit
position detection function. Further, information indicating the detection result of the client unit
CUK is referred to as client unit position information.
[0032]
In order to realize the client unit position detection function, each of the client units CU1 to CU4
has a wireless tag. Each wireless tag of the client units CU1 to CU4 transmits a signal.
09-05-2019
8
[0033]
Hereinafter, when there is no need to distinguish the client units CU1 to CU4 individually, they
will be collectively referred to as a client unit CU. Also, a signal transmitted from the client unit
CU is referred to as a client unit signal.
[0034]
Each of the wireless nodes WN1 to WN9 receives a client unit signal. Each of the wireless nodes
WN1 to WN9 measures the radio wave intensity, the delay characteristic, and the like of the
client unit signal. The measurement results are hereinafter referred to as client signal
measurement results. The client signal measurement result is output to the server 1.
[0035]
The server 1 generates client unit position information based on each client signal measurement
result from the wireless nodes WN1 to WN9. That is, the presence position of the user holding
the client unit CU is detected. Then, the server 1 performs control to change the state of each
sound output from the speakers 3 to 7 in accordance with the position where the user is present.
A detailed example of this control will be described later. Also, hereinafter, the wireless nodes
WN1 to WN9 are simply referred to as wireless nodes WN when there is no need to distinguish
them individually.
[0036]
FIG. 2 is a block diagram showing a detailed configuration example of the server 1.
[0037]
The server 1 is configured to include a system interface unit 21 to a position detection unit 26.
[0038]
09-05-2019
9
Further, for example, the tuner 11, the network 12, and the recording device 13 are connected to
the server 1.
The tuner 11 to the recording device 13 may also be understood as one component of the
information processing system of FIG. 1.
Furthermore, the functions of the tuner 11 and the recording device 13 may be installed in the
server 1.
[0039]
The tuner 11 receives a broadcast program from a broadcast station, and supplies it to the
system interface unit 21 in the form of compression-coded video and audio signals.
[0040]
The video signal and the audio signal compressed and encoded by another device are output
from the other device and supplied to the system interface unit 21 via the network 12.
[0041]
The recording device 13 compresses and records the video signal and the audio signal of the
content.
The recording device 13 supplies the system interface unit 21 in the form of compression-coded
video and audio signals.
[0042]
The system interface unit 21 supplies the video signal and the audio signal supplied from the
tuner 11, the network 12 or the recording device 13 to the system decoding unit 22.
[0043]
09-05-2019
10
As described above, the video signal and the audio signal supplied from the system interface unit
21 to the system decoding unit 22 are compressed and encoded in a predetermined format.
Therefore, the system decoding unit 22 performs expansion decoding processing on the
compression-coded video signal and audio signal.
Among the video signal and the audio signal obtained as a result of the decoding /
decompression processing, the video signal is supplied to the video processing unit 23, and the
audio signal is supplied to the audio processing unit 24.
[0044]
The video processing unit 23 appropriately performs image processing on the video signal from
the system decoding unit 22, and then supplies the video signal to the network interface unit 25.
[0045]
The audio signal supplied to the audio processing unit 24 is an audio signal of multiple channels
as described above.
Therefore, the audio processing unit 24 converts multi-channel audio signals into 5.1-channel
audio signals. Further, using the client unit position information from the position detection unit
26 and the 5.1 channel audio signal, the audio processing unit 24 generates an audio signal of
each channel to be supplied to each of the speakers 3 to 7. Hereinafter, the audio signal of each
channel supplied to each of the speakers 3 to 7 will be referred to as an audio signal S̲out3, an
audio signal S̲out4, an audio signal S̲out5, an audio signal S̲out6, and an audio signal S̲out7.
A series of processes until generating the audio signals S̲out3 to S̲out7 will be referred to as
audio signal output processing. Details of the audio signal output process will be described later
with reference to FIG.
[0046]
The network interface unit 25 outputs the video signal from the video processing unit 23 to the
09-05-2019
11
very large screen monitor 2. Also, the network interface unit 25 outputs the audio signals S̲out3
to S̲out7 from the audio processing unit 24 to the speakers 3 to 7, respectively.
[0047]
The position detection unit 26 receives the client signal measurement result of the wireless node
WN, and generates client unit position information based on the reception result. The client unit
position information refers to information indicating the presence position of the user holding
the client unit CU, as described above. The client unit position information is provided from the
position detection unit 26 to the voice processing unit 24.
[0048]
[Example of Processing Method of Audio Signal Output Device to Which Present Invention is
Applied] FIG. 3 is a flowchart for explaining an example of audio signal output processing.
[0049]
In step S1, the position detection unit 26 of the server 1 determines whether a client unit signal
measurement result has been received from the wireless node WN.
[0050]
In the example of FIG. 1, the case where the client unit signal measurement result is not
transmitted from any of the wireless nodes WN1 to WN9 means the case where the client unit
CU does not exist in the target area.
Therefore, in such a case, it is determined that the answer is NO in step S1, and the process
proceeds to step S7.
However, the process after step S7 will be described later.
[0051]
On the other hand, when the client unit signal measurement result is transmitted from at least
09-05-2019
12
one of the wireless nodes WN1 to WN9 and received by the position detection unit 26, it is
determined as YES in step S1, and the process is performed Go to S2.
[0052]
In step S2, the position detection unit 26 attempts to receive the client unit signal measurement
results from the other wireless node WN.
[0053]
In step S3, the position detection unit 26 determines whether a predetermined time has elapsed.
If the predetermined time has not yet elapsed, it is determined as NO in step S3, and the process
is returned to step S2, and the subsequent processes are repeated.
That is, each time a client unit signal measurement result from another wireless node WN is
transmitted until the predetermined time elapses, the position detection unit 26 receives the
result.
[0054]
Thereafter, when the predetermined time has elapsed, it is determined as YES in step S3 and the
process proceeds to step S4.
[0055]
In step S4, the server 1 generates client unit position information based on the client unit signal
measurement results from one or more wireless nodes WN.
The client unit position information is supplied from the position detection unit 26 to the voice
processing unit 24.
[0056]
09-05-2019
13
Specifically, for example, in the present embodiment, the target area α is divided into a plurality
of areas (hereinafter, referred to as group areas). The position detection unit 26 detects in which
group area the client unit CU is located, based on the client unit signal measurement result
received from the wireless node WN. Then, the position detection unit 26 generates, as client unit
position information, information indicating a group area to which the client unit CU belongs.
Note that a specific example of the client unit position information will be described later with
reference to FIG.
[0057]
Also, the number of client units CU is not limited to one, and there are as many as the number of
viewers present in the target area α. For example, in the example of FIG. 1, four client units CU1
to CU4 exist in the target area α. In such a case, in the process of step S4, client unit position
information is generated for each of the plurality of client units CU.
[0058]
In step S5, the voice processing unit 24 determines whether or not the client unit CU to be
detected is located in the same group area.
[0059]
Here, the client unit CU to be detected refers to the client unit CU in which the position
information of the client unit is generated in the process of step S4.
[0060]
If at least one of the plurality of client units CU exists in a different group area in the target area
α, it is determined as NO in step S5, and the process proceeds to step S7.
However, the process after step S7 will be described later.
[0061]
09-05-2019
14
On the other hand, if only one client unit CU exists in the target area α, or if a plurality of client
units CU exist in the same group, it is determined as YES in step S5, and the process is
performed. The process proceeds to step S6.
[0062]
In step S6, the audio processing unit 24 sets the output state of the audio signal according to the
group area in which the client unit CU is located.
That is, the audio processing unit 24 generates audio signals S̲out 3 to S̲out 7 according to the
group region, and outputs the audio signals S̲out 3 to S̲out 7 to the speakers 3 to 7 through the
network interface unit 25.
[0063]
On the other hand, if the client unit CU does not exist in the target area α or if a plurality of
client units CU exist across two or more group areas, it is determined as NO in step S1 or S5. The
process proceeds to step S7.
In step S7, the audio processing unit 24 sets the output state of the audio signal to the initial
state. That is, the audio processing unit 24 transmits each of the stereo signal LO, the stereo
signal R0, the right surround signal Rs, the center channel signal C, and the left surround signal
Ls to each of the speakers 3 to 7 through the network interface unit 25. Output.
[0064]
When a plurality of client units CU exist across two or more group areas, that is, when it is
determined as NO in S5, the audio processing unit 24 determines the output state of the audio
signal as the initial state. As another different state, for example, a non-directional state can be
used.
[0065]
09-05-2019
15
The audio signal output process is repeated at regular time intervals.
That is, the client unit signal measurement results from the plurality of wireless nodes WN
arranged at multiple points are transmitted to the position detection unit 26 of the server 1 at
fixed time intervals. As a result, if there is no movement of the client unit CU, the output state of
the audio signal by the process of step S6 of each time becomes the same. That is, if the client
unit CU does not move, the output state of the audio signal is maintained. On the other hand, if
the client unit CU is moving, the output state of the audio signal by the process of step S6 of each
time changes every moment according to the movement position of the client unit CU. In this
case, the position detection unit 26 can calculate each client unit position information as a time
variable, and configure the center offset distance table based on the calculation result.
[0066]
FIG. 4 is a diagram showing an example of client unit position information.
[0067]
The client unit position information shown in FIG. 4 is indicated by the combination of the
respective distances of the target client unit CU and the speakers 3 to 7.
[0068]
The first line (initial setting) in FIG. 4 indicates a basic example of the client unit position
information in the initial state.
When such client unit position information (initial setting) is supplied from the position detection
unit 26 to the audio processing unit 24, the output state of the audio signal becomes the initial
state.
That is, the stereo signal LO, the stereo signal R0, the right surround signal Rs, the center
channel signal, and the left surround signal Ls are output from the speakers 3 to 7, respectively.
[0069]
09-05-2019
16
For example, it is assumed that only the client unit CU1 of FIG. 1 exists in the target area α. In
this case, the client unit CU1 is near to the speaker 3, near to the speaker 4, far to the speaker 5,
far to the speaker 5, mid to the speaker 6, and The speaker 7 belongs to a group area near (near).
Accordingly, the client unit position information No 1 shown in FIG. 4 is generated by the
position detection unit 26 and supplied to the voice processing unit 24.
[0070]
In this case, the audio processing unit 24 generates the audio signals S̲out 3 to S̲out 7 by
respectively operating the following equations (1) to (5), and the speakers 3 to 7 are generated
via the network interface unit 25. Output to each.
[0071]
Speaker 3: S̲out3 = L0 * CL + R0 * CS + C * CS + Rs * CS + Ls * CM (1) Speaker 4: S̲out4 = L0 *
CL + R0 * CL + C * CS + Rs * CM + Ls * CS (2) Speaker 5: S̲out5 = L0 * CL + R0 * CL + C * CS + Rs
* CM + Ls * CS (3) Speaker 6: S̲out6 = L0 * CL + R0 * CL + C * CS + Rs * CM + Ls * CS (4) Speaker
7: S̲out7 = L0 * CS + R0 * CL + C * CS + Rs * CM + Ls * CS (5)
[0072]
Here, CL, CM, and CS indicate coefficients for weighting the audio signal (hereinafter, referred to
as downmix coefficients).
The downmix coefficients CS, CM, and CL have large values in that order.
[0073]
That is, the audio signal S̲outM supplied to the speaker M (M is an integer value of 3 to 7) is
calculated as in the following equation (6).
That is, a linear combination of each channel signal weighted by multiplying downmix
coefficients C1 to C5 to each of stereo signal LO, stereo signal R0, right surround signal Rs,
center channel signal C, and left surround signal Ls Is the audio signal S̲outM.
09-05-2019
17
[0074]
Speaker M: S̲out M = L0 * C1 + R0 * C2 + C * C2 + Rs * C4 + Ls * C5 (6)
[0075]
Each of these downmix coefficients C1 to C5 is varied to any one of the downmix coefficients CL,
CM, and CS according to the group area in which the client unit M exists.
[0076]
For example, it is assumed that a combination of downmix coefficients C1 to C5 is determined in
advance for each of the speakers 3 to 7 in accordance with the group area indicated by the client
unit position information No2.
In this case, if only the client unit CU2 of FIG. 1 exists in the target area α, the client unit
position information No. 2 is acquired, so the downmix coefficients C1 to C5 determined for the
client unit position information No. 2 are obtained. Is adopted for each of the speakers 3 to 7.
Then, the downmix coefficients C1 to C5 respectively adopted for the speakers 3 to 7 are
substituted into the equation (6) and calculated, and as a result, the audio signals S̲out3 to
S̲out7 corresponding to the client unit position information No3 are obtained. It is generated.
[0077]
Further, for example, it is assumed that combinations of downmix coefficients C1 to C5 are
determined in advance for each of the speakers 3 to 7 in accordance with the group area
indicated by the client unit position information No. 3. In this case, if only the client unit CU3 of
FIG. 1 exists in the target area α, client unit position information No 3 is acquired. Therefore, the
combination of the downmix coefficients C1 to C5 determined for the client unit position
information No. 3 is adopted for each of the speakers 3 to 7. Then, the downmix coefficients C1
to C5 respectively adopted for the speakers 3 to 7 are substituted into the equation (6) and
calculated, and as a result, the audio signals S̲out3 to S̲out7 corresponding to the client unit
position information No3 are obtained. It is generated.
09-05-2019
18
[0078]
Further, for example, it is assumed that combinations of downmix coefficients C1 to C5 are
determined in advance for each of the speakers 3 to 7 in accordance with the group area
indicated by the client unit position information No 4. In this case, if only the client unit CU4 of
FIG. 1 exists in the target area α, the client unit position information No 4 is acquired. Therefore,
the combination of the downmix coefficients C1 to C5 determined for the client unit position
information No 4 is adopted for each of the speakers 3 to 7. Then, the downmix coefficients C1
to C5 respectively adopted for the speakers 3 to 7 are substituted into the equation (6) and
calculated, and as a result, the audio signals S̲out3 to S̲out7 corresponding to the client unit
position information No 4 are obtained. It is generated.
[0079]
The audio signal S̲out3 to S̲out7 generated so as to be suitable for the location where the
viewer holding the client unit CU is present in any position of the target area α by the abovedescribed audio signal output processing is a speaker It is supplied to each of 3 to 7. Therefore,
the sound of each channel suitable for the viewer's location is output from the speakers 3 to 7.
As a result, the viewer can listen to suitable audio.
[0080]
The audio signal processing when the client unit position information No. 5 shown in FIG. 4 is
acquired will be described.
[0081]
Client unit position information No 5 means Near for the speaker 3, Far for the speaker 4, Far for
the speaker 5, Near for the speaker 6, and the speaker 7. Is a collection of information called
Near.
[0082]
However, in the example of FIG. 1, it can not be considered that the client unit position
09-05-2019
19
information No. 5 is acquired in a state where any one of the client units CU1 to CU4 is
stationary.
Therefore, in the example of FIG. 1, the following two possibilities can be considered as the
possibility that the client unit position information No. 5 is obtained.
[0083]
As a first possibility, it is conceivable that a plurality of client units CU may exist in different
group areas.
For example, in the example of FIG. 1, when the client unit CU1 and the client unit CU3
simultaneously exist at the positions shown in FIG. 1, the client unit position information No. 5 is
acquired.
[0084]
A second possibility is that a single client unit CU may be moving during the process of acquiring
client unit position information. For example, in the example of FIG. 1, when the client unit CU1
moves from the position shown in FIG. 1 to the position shown as the position of the client unit
CU2 in FIG. 1, the client unit information No. 5 is acquired.
[0085]
When such client unit information No. 5 is acquired, the audio processing unit 24 can set the
output state of the audio signal to a general-purpose state without directivity (for example, an
initial state).
[0086]
If it is necessary to determine whether it is the first possibility or the second possibility, the
center offset distance table configured based on each client unit position information as a time
variable may be used. .
09-05-2019
20
This is because the history of the client unit position information acquired before the client unit
position information No. 5 can be easily determined as to whether it is the first possibility or the
second possibility.
[0087]
As described above, the server 1 can variably set the parameter (the downmix coefficient in the
above example) of the audio signal based on the client unit position information of the client unit
CU. Furthermore, the server 1 can also change various parameters of the video signal based on
the client unit position information of the client unit CU. For example, when the position of the
client unit CU is at a position far from the super-large screen monitor 2, the server 1 variously
displays the image itself or character information (such as subtitles) related to the image.
Parameters can be set. 〈2. Second Embodiment [Configuration Example of Client Unit CU]
FIG. 5 is an example of another embodiment of the client unit CU described above with reference
to FIGS. 1 and 2.
[0088]
The client unit CUa shown in FIG. 5 is a portable monitor with a wireless tag. The client unit CUb
is a headphone with a wireless tag.
[0089]
The client unit CUa receives the video signal and the audio signal from the server 1, displays the
video corresponding to the video signal, and outputs the audio corresponding to the audio signal.
[0090]
In this case, the server 1 can, of course, variably set the parameters (for example, the downmix
coefficient) of the above-described audio signal based on the client unit position information of
the client unit CUa.
09-05-2019
21
Furthermore, the server 1 can change various parameters of the video signal based on the client
unit position information of the client unit CUa. For example, according to the existing position of
the client unit CUa, the server 1 displays the video itself displayed on the super large screen
monitor 2 or character information (subtitles etc.) regarding the video on the client unit CUa.
Thus, various parameters can be set.
[0091]
The client unit CUb receives an audio signal from the server 1 and outputs the received audio.
[0092]
For example, the server 1 can variably set the parameters (for example, the downmix coefficient)
of the above-described audio signal based on the client unit position information of the client unit
CUb.
The voice signal generated by the server 1 after parameter setting, in the above example, the
voice signals S̲out3 to S̲out7 are wirelessly transmitted to the client unit CUb.
[0093]
That is, in the first embodiment, it is premised that the audio of each channel is output from the
speakers 3 to 7. Therefore, for example, when a plurality of client units CU exist in different
group areas, etc., the server 1 does not have directivity as a parameter of the audio signal, for
example, a general setting (for example, a parameter value for setting an initial state). It was set).
[0094]
On the other hand, in the second embodiment, audio is output from the client unit CUb.
Therefore, even if, for example, a plurality of client units CUb exist in different group areas, etc.,
the server 1 individually responds to each of the plurality of client units CUb as a parameter of
the audio signal according to each existing position Setting (eg, setting of separate downmix
coefficients) can be performed. This allows each client unit CUb to hear an audio signal suitable
09-05-2019
22
for its location.
[0095]
The viewer may hold both the client unit CUa and the client unit CUb, or may hold only one of
them.
[0096]
Further, the method of detecting the client position in the information processing apparatus to
which the present invention is applied is not limited to the method described with reference to
FIGS. 1 to 4, and any method may be used.
[0097]
In the information processing apparatus to which the present invention is applied, suitable video
and audio can be output according to the presence position of the viewer.
As a result, in the case where a viewer views video or audio in a wide range such as an event
venue, for example, the viewer can easily view suitable video or audio regardless of the location
thereof. .
[0098]
Furthermore, in the information processing apparatus to which the present invention is applied,
each client unit position information can be calculated as a time variable.
As a result, even when the viewer moves, for example, in the event hall, the information
processing apparatus to which the present invention is applied can prepare a suitable viewing
environment according to the position.
[0099]
By the way, the series of processes described above can be performed by hardware or can be
performed by software.
09-05-2019
23
[0100]
When the series of processes described above are performed by software, an information
processing apparatus to which the present invention is applied can be configured to include, for
example, a computer shown in FIG.
Alternatively, the robot hand apparatus to which the present invention is applied may be
controlled by the computer of FIG.
[0101]
In FIG. 6, a central processing unit (CPU) 101 executes various processes in accordance with a
program stored in a read only memory (ROM) 102 or a program loaded from a storage unit 108
to a random access memory (RAM) 103. Do. The RAM 103 also stores data necessary for the
CPU 101 to execute various processes.
[0102]
The CPU 101, the ROM 102, and the RAM 103 are mutually connected via a bus 104. An input /
output interface 105 is also connected to the bus 104.
[0103]
The input / output interface 105 includes an input unit 106 including a keyboard and a mouse,
an output unit 107 including a display and the like, a storage unit 108 including a hard disk and
the like, and a communication unit 109 including a modem and a terminal adapter. It is
connected. The communication unit 109 controls communication performed with another device
(not shown) via a network including the Internet.
[0104]
09-05-2019
24
A drive 110 is also connected to the input / output interface 105 as necessary, and removable
media 111 composed of a magnetic disk, an optical disk, a magneto-optical disk, or a
semiconductor memory are appropriately attached, and a computer program read from them is ,
And installed in the storage unit 108 as necessary.
[0105]
When a series of processes are to be executed by software, various functions are executed by
installing a computer in which a program configuring the software is incorporated in dedicated
hardware or various programs. Can be installed, for example, on a general-purpose personal
computer or the like from a network or a recording medium.
[0106]
A recording medium containing such a program, as shown in FIG. 6, includes a magnetic disk (a
floppy disk having a program recorded therein) distributed to provide the program to the viewer
separately from the apparatus main body. Removable media (packages) consisting of optical
disks (including CD-ROMs (Compact Disk-Read Only Memory) and DVDs (Digital Versatile Disks)),
magneto-optical disks (including MD (Mini-Disks)), or semiconductor memory etc. In addition to
being configured by the media 111, the ROM 102 is provided to the viewer in a state of being
incorporated in advance in the apparatus main body, and is configured by the ROM 102 in which
the program is recorded and a hard disk included in the storage unit 108.
[0107]
In the present specification, in the step of describing the program to be recorded on the
recording medium, the processing performed chronologically along the order is, of course,
parallel or individually not necessarily necessarily chronologically processing. It also includes the
processing to be performed.
[0108]
Further, in the present specification, the system represents the entire apparatus configured by a
plurality of apparatuses and processing units.
[0109]
Reference Signs List 1 server, 2 super large screen monitor, 3, 4, 5, 6, 7 speakers, CU1, CU2,
CU3, CU4, CU5, CUa, CUb Client unit, WN1, WN2, WN3, WN4, WN5, WN6, WN7, WN8 , WN9
wireless node, 11 tuner, 12 network, 13 recording device, 21 system interface unit, 22 system
decoding unit, 23 video processing unit, 24 audio processing unit, 25 network interface unit, 26
09-05-2019
25
position detection unit, 101 CPU, 102 ROM , 103 RAM, 104 bus, 105 input / output interface,
106 input unit, 107 output unit, 108 storage unit, 109 communication unit, 110 drives, 111
removable media
09-05-2019
26