close

Вход

Забыли?

вход по аккаунту

?

JP2018121310

код для вставкиСкачать
Patent Translate
Powered by EPO and Google
Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
DESCRIPTION JP2018121310
Abstract: To easily realize a multi-channel audio reproduction environment. An audio data
generation unit (111) of an audio processing device (11) includes a sound bar (12) in which a
plurality of speaker groups for outputting audio data for each output channel are arranged in the
vertical direction; The audio data for each of the plurality of output channel speaker groups
including the sound bar (13) in which the speaker groups are arranged in the vertical direction is
generated by downmixing the audio data of the plurality of input channels. [Selected figure]
Figure 1
Audio processing apparatus, audio output apparatus, television receiver, audio processing
method, program, and recording medium for program
[0001]
The present invention relates to an audio processing device, an audio output device, a television
receiver, an audio processing method, a program, and a recording medium for the program.
[0002]
In advanced BS broadcasting, where practical broadcasting is scheduled to start, not only high
resolution of video but also 22.2ch of audio (hereinafter, ch indicates "channel") is transmitted as
to audio.
In order to reproduce audio signals of 22.2 ch and other chs transmitted by these broadcasts, 22
03-05-2019
1
SPs (hereinafter, SP indicates a "speaker") and 2 low-pass SPs are predetermined. 24ch amplifier
is required to install in position and drive SP.
[0003]
In order to more easily reproduce the above three-dimensional sound signal, a method of
converting it into a general surround signal such as 5.1 ch has been presented. For example,
Non-Patent Document 1 discloses a conversion equation for downmixing a 22.2 ch signal to a
general 5.1 ch signal.
[0004]
Patent No. 5604365 (disclosed on December 6, 2012) Patent No. 5852325 (disclosed on
December 6, 2012)
[0005]
ARIB SRTD-B32 Part 2 (revised on December 9, 2016)
[0006]
The method of Non-Patent Document 1 has a problem that it is not a conversion method that
places more emphasis on stereophonic sound effects.
That is, although the planar surround sound image is reproduced by merging the signals of the
upper layer channel transmitting a large amount of information in 22.2 ch and the front channel
signal into a smaller channel, the three-dimensional effect, The thickness etc. of the frontal sound
image will be reduced.
[0007]
Further, in Patent Documents 1 and 2, while emphasizing the information of the front channel
which is important data of 22.2 ch, the signal of the surround channel is also divided into the SP
of the upper layer and the SP of the middle layer. There is disclosed a multi-channel speaker
03-05-2019
2
device which realizes audio reproduction of (three-dimensional effect).
However, in the techniques of Patent Documents 1 and 2, dedicated speakers are disposed
around the television screen, and audio processing is performed according to the dedicated
speakers.
[0008]
On the other hand, in recent years, reproduction of a three-dimensional sound image such as
Dolby Atmos has been required. Since the original 22.2 ch has sufficient information of the threedimensional sound image, there is a need for a method for easily realizing the three-dimensional
sound image in ordinary homes without using a dedicated speaker while utilizing the information
of the three-dimensional sound image .
[0009]
An aspect of the present invention aims to easily realize a multi-channel audio reproduction
environment.
[0010]
In order to solve the above problems, an audio processing apparatus according to an aspect of
the present invention is an audio processing apparatus that generates audio data of a plurality of
output channels by downmixing audio data of a plurality of input channels, A first sound bar in
which a plurality of speaker groups for outputting audio data for each output channel are
arranged in the vertical direction, and a plurality of speaker groups for outputting audio data for
each output channel are arranged in the vertical direction An audio data generation unit is
provided that generates audio data for each of a plurality of output channel speakers including
two sound bars by downmixing audio data of the plurality of input channels.
[0011]
In order to solve the above problems, an audio processing method according to an aspect of the
present invention is an audio processing method for generating audio data of a plurality of
output channels by downmixing audio data of a plurality of input channels. A first sound bar in
which a plurality of speaker groups for outputting audio data for each output channel are
arranged in the vertical direction, and a plurality of speaker groups for outputting audio data for
each output channel are arranged in the vertical direction Audio data generation step of
03-05-2019
3
generating audio data for each of the plurality of output channel speakers including the second
sound bar by downmixing the audio data of the plurality of input channels.
[0012]
According to one aspect of the present invention, it is possible to easily realize a multi-channel
audio reproduction environment.
[0013]
It is a block diagram showing composition of an audio output device concerning Embodiment 1
of the present invention.
BRIEF DESCRIPTION OF THE DRAWINGS It is a front view which shows the external appearance
of the television receiver which concerns on Embodiment 1 of this invention.
It is a figure which shows the speaker position which the audio | voice data of 22.2ch which the
audio | voice output device concerning Embodiment 1 of this invention acquires acquires.
It is a figure which shows the speaker position which the audio | voice data of 14.2 ch which the
audio | voice output device based on Embodiment 1 of this invention produces | generates
assume.
It is a figure which shows the calculation formula of the downmix which concerns on
Embodiment 1 of this invention. It is a figure which shows typically the sound field produced |
generated when the sound data of 14.2 ch which concerns on Embodiment 1 of this invention
arrange | positions two sound bars which have 7.1 ch on the left and right, and reproduces |
regenerates. It is a figure which shows the speaker position which the audio | speech data of
10.2 ch which the audio | voice output device based on Embodiment 1 of this invention produces
| generates assumes. It is a figure which shows the calculation formula of the downmix which
concerns on Embodiment 2 of this invention. It is a figure which shows the speaker position
which the audio | speech data of 8.1ch which the audio | voice output device based on
Embodiment 3 of this invention produces | generates assumes. It is a figure which shows the
calculation formula of the downmix which concerns on Embodiment 3 of this invention. It is a
figure which shows typically the sound field produced | generated when it reproduces |
03-05-2019
4
regenerates with two sound bars corresponding to the audio data of 8.1ch which concerns on
Embodiment 3 of this invention. It is a figure which shows the speaker position which the audio |
speech data of 6.1ch which the audio | voice output device based on Embodiment 4 of this
invention produces | generates assumes. It is a figure which shows the calculation formula of the
downmix which concerns on Embodiment 4 of this invention.
[0014]
Embodiment 1 Hereinafter, Embodiment 1 of the present invention will be described in detail.
[0015]
(Configuration of Audio Output Device 1) FIG. 1 is a block diagram showing a configuration of
the audio output device 1 according to the present embodiment.
The audio output device 1 includes an audio processing device 11, a sound bar (first sound bar)
12, and a sound bar (second sound bar) 13.
[0016]
The audio processing device 11 generates audio data of a plurality of output channels by
downmixing audio data of a plurality of input channels. The voice processing device 11 includes
a voice data generation unit 111. The audio data generation unit 111 generates audio data for
each of a plurality of output channel speakers including the sound bars 12 and 13 by
downmixing audio data of a plurality of input channels.
[0017]
The sound bar 12 is formed by arranging a plurality of speaker groups which output audio data
for each output channel in the vertical direction. The sound bar 13 is formed by arranging a
plurality of speaker groups which output audio data for each output channel in the vertical
direction.
03-05-2019
5
[0018]
(External Appearance of Television Receiver 2) FIG. 2 is a front view showing the external
appearance of the television receiver 2 according to the present embodiment. The television
receiver 2 includes an audio output device 1. As shown in FIG. 2, sound bars 12 and 13 and a
screen 14 are disposed on the front of the television receiver 2. The sound bar 12 is disposed on
the left side of the screen 14. The sound bar 13 is disposed on the right side of the screen 14.
[0019]
(Process of audio processing device 11) The audio processing device 11 acquires audio data of
22.2 ch and downmixes to generate audio data of 14.2 ch.
[0020]
In the following description, each channel of audio data and the speaker position assumed by the
channel will be described using the same symbol.
[0021]
FIG. 3 is a diagram showing a speaker position assumed by the audio data of 22.2 channels
acquired by the audio processing device 11 according to the present embodiment.
As shown in FIG. 3, the audio data of 22.2 ch is arranged in the middle of 11 SPs (TpFL, TpFC,
TpFR, FL, FLC, FC, FRC, FR, BtFL, BtFC, BtFR) arranged in the front. It is assumed that the output
is from 5 SPs (TpSiL, TpC, TpSiR, SiL, SiR) arranged and 6 SPs (TpBL, TpBC, TpBR, BL, BC, BR)
arranged behind There is.
In addition, each channel of audio data of 22.2 ch is 9 ch (TpFL, TpFC, TpFR, TpSiL, TpC, TpSiR,
TpBL, TpBC, TpBR) in the upper layer, and 10 ch (FL, FLC, FC, FRC, FR) in the middle layer. It is
classified into FR, SiL, SiR, BL, BC, BR), and 5ch (BtFL, BtFC, BtFR, LFE1, LFE2) which is the lower
layer.
[0022]
The lower 2ch channels (LFE1, LFE2) correspond to the audio data of the woofer 0.2ch.
03-05-2019
6
[0023]
FIG. 4 is a view showing a speaker position assumed by audio data of 14.2 ch generated by the
audio processing device 11 according to the present embodiment.
As shown in FIG. 4, the audio data of 14.2 ch is composed of six SPs (R_L, C_L, L_L, L_R, C_R,
R_R) arranged in front, and four virtual SPs (SR_L arranged in the middle). , SL_L, SL_R, SR_R),
and four virtual SPs (SBL_L, SBL_R, SBL_R, SBR_R) arranged behind are output as output
positions. In FIG. 4 the virtual SP is indicated by the outline of the dashed line.
[0024]
The audio output device 1 can output audio data of 14.2 ch by arranging two sound bars having
7.1 ch vertically and arranging them on the left and right and arranging two pairs of woofers as a
pair in the lower layer. . Here, of the two sound bars, the sound bar disposed in the left layer is
referred to as SoundBar_L, and the sound bar disposed in the right layer is referred to as
SoundBar_R.
[0025]
As shown in FIG. 4, the sound bar (SoundBar_L) has a speaker group that reproduces audio
corresponding to each of seven channels SBR_L, SR_L, R_L, C_L, L_L, SL_L, and SBL_L. Here, each
channel (SBR_L, SR_L, SL_L, SBL_L (total four)) shown by a broken line in FIG. 4 has virtual SP
(SBR_L, SR_L, SL_L, SBL_L) as an output position, and a sound bar It has a speaker group for
reproducing the corresponding virtual speaker of (SoundBar_L).
[0026]
Similarly, as shown in FIG. 4, the sound bar (SoundBar_R) has a speaker group that reproduces
audio corresponding to each of the seven channels SBL_R, SL_R, L_R, C_R, R_R, SR_R, and SBR_R.
There is. FIG. 4 shows an example in which speakers corresponding to each channel are
03-05-2019
7
individually provided. Here, each channel (SBL_R, SL_R, SR_R, SBR_R (a total of four)) shown by a
broken line in FIG. 4 has virtual SP (SBL_R, SL_R, SR_R, SBR_R) as an output position, and a
sound bar It has a speaker group for reproducing a corresponding virtual speaker of
(SoundBar_R).
[0027]
The audio output device 1 includes an amplifier for driving the speaker configured as described
above.
[0028]
In the audio processing device 11, the audio data generation unit 111 downmixes audio data of
the left layer of a plurality of input channels to audio data of the left layer of a plurality of output
channels, and generates audio data of middle layers of a plurality of input channels. And merging
the audio data of the left layer and the right layer of the plurality of output channels, and
downmixing the audio data of the right layer of the plurality of input channels into the audio data
of the right layer of the plurality of output channels.
[0029]
When downmixing audio data of 22.2 ch to audio data of 14.2 ch, the audio data generation unit
111 outputs audio data of the left layer 7 ch of 22.2 ch (TpFL, TpSiL, TpBL, FL, SiL, BL, BtFL) is
downmixed to 14.2 ch left layer 7 ch audio data (SBR_L, SR_L, R_L, C_L, L_L, SL_L, SBL_L).
The audio data generation unit 111 is configured to transmit audio data (TpFC, TpC, TpBC, FLC,
FC, FRC, BC, BtFC) of 22.2 middle layers in the middle layer 8 ch to audio data (SBR_L, SR_L, R_L
in the left layer 6 ch) , C_L, L_L, SBL_L) and the right layer 6ch (SBL_R, SL_R, L_R, C_R, R_R,
SBR_R).
The audio data generation unit 111 generates audio data (TpFR, TpSiR, TpBR, FR, SiR, BR, BtFR)
of the 22.2 right layer 7ch into audio data (SBL_R, SL_R, L_R, 14.2ch of the right layer 7ch).
Downmix to C_R, R_R, SR_R, SBR_R).
[0030]
FIG. 5 is a diagram showing a formula of downmix according to the present embodiment.
03-05-2019
8
[0031]
The voice data generation unit 111 adjusts the distribution coefficient of the middle layer to
express the sense of voice spread.
a, k, v3, v4, v5 are common coefficients.
[0032]
The coefficient a is a coefficient set to prevent data overflow during downmix processing, and a
value between 0 and 1 is arbitrarily set. The coefficient k is a surround channel gain coefficient,
which is a coefficient sent by broadcasting. The coefficients v3 and v4 are vertical coefficients
(vertical distribution), and the coefficient v5 is a vertical coefficient (middle distribution), and
values of 0 to 1 are arbitrarily set. Further, the relationship between the vertical direction
coefficients is v3 + v4 + v5 = 1. These coefficients can be arbitrarily set in order to optimize the
feeling of expansion in accordance with the position of the upper and lower SP (sound bar
installation position, television screen size).
[0033]
Formulas 1 to 7 are formulas for calculating audio data of the left layer of 14.2 ch. As shown in
Equation 1, the audio data generation unit 111 generates L_L using BtFL, BtFC, FL, FLC, and FC.
As shown in Equation 2, the audio data generation unit 111 generates R_L using TpFL, TpFC, FL,
FLC, and FC. As shown in Equation 3, the audio data generation unit 111 generates C_L using FL,
FLC, and FC.
[0034]
As shown in Equation 4, the audio data generation unit 111 generates SL_L using SiL. As shown
in Expression 5, the audio data generation unit 111 generates SR_L using TpSiL and TpC. As
shown in Equation 6, the audio data generation unit 111 generates SBL_L using BL and BC. As
03-05-2019
9
shown in Equation 7, the audio data generation unit 111 generates SBR_L using TpBL, TpBC, and
TpC.
[0035]
As shown in Expression 8, the audio data generation unit 111 generates 14.2 ch LFE 1 using
22.2 ch LFE 1.
[0036]
Equations 9 to 15 are equations for calculating audio data of the 14.2 ch right layer.
As shown in Equation 9, the audio data generation unit 111 generates L_R using TpFR, TpFC, FR,
FRC, and FC. As shown in Expression 10, the audio data generation unit 111 generates R_R using
BpFR, BpFC, FR, FRC, and FC. As shown in Expression 11, the voice data generation unit 111
generates C_R using FR, FRC, and FC.
[0037]
As shown in Expression 12, the audio data generation unit 111 generates SL_R using TpSiR and
TpC. As shown in Expression 13, the audio data generation unit 111 generates SR_R using SiR.
As shown in Expression 14, the audio data generation unit 111 generates SBL_R using TpBR,
TpBC, and TpC. As shown in Formula 15, the audio data generation unit 111 generates SBR_R
using BR and BC.
[0038]
As shown in Expression 16, the audio data generation unit 111 generates 14.2 ch LFE 2 using
22.2 ch LFE 2.
[0039]
FIG. 6 is a view schematically showing a sound field generated when audio data of 14.2 ch
according to the present embodiment is reproduced by arranging two sound bars having 7.1 ch
to the left and right.
03-05-2019
10
[0040]
In FIG. 6, SF_SoundBar_L schematically shows a sound field generated by the sound bar
(SoundBar_L), and SF_SoundBar_R schematically shows a sound field generated by the sound bar
(SoundBar_R).
[0041]
[Effects of First Embodiment] According to the above configuration, the audio processing device
11 can easily realize a multi-channel audio reproduction environment by using the two sound
bars 12 and 13.
In particular, when downmixing 22.2ch audio data to 14.2ch audio data, downmix the 22.2ch left
layer audio data to 14.2ch left layer audio data, and Since the middle layer audio data is merged
with the 14.2ch left layer and right layer audio data, and the 22.2ch right layer audio data is
downmixed to the 14.2ch right layer audio data, the 22.2ch The audio data of three layers (left,
middle and right) can be easily reproduced by two sound bars 12, 13 (left and right).
[0042]
EMBODIMENT 2 It will be as follows if Embodiment 2 of this invention is demonstrated based on
FIG.7 and FIG.8.
In addition, about the member which has the same function as the member demonstrated in the
said embodiment for convenience of explanation, the same code | symbol is appended and the
description is abbreviate | omitted.
[0043]
The audio processing apparatus 11 generates audio data of 10.2 ch by acquiring audio data of
22.2 ch and downmixing.
[0044]
In the following description, each channel of audio data and the speaker position assumed by the
03-05-2019
11
channel will be described using the same symbol.
[0045]
FIG. 7 is a view showing a speaker position assumed by audio data of 10.2 ch generated by the
audio processing device 11 according to the present embodiment.
As shown in FIG. 7, the audio data of 10.2 ch is composed of six SPs (R_L, C_L, L_L, L_R, C_R,
R_R) placed before, and four virtual SPs placed behind (SR_L). , SL_L, SL_R, SR_R) are output as
output positions.
In FIG. 7, the virtual SP is indicated by the outline of the dashed line.
[0046]
The audio output device 1 can output audio data of 10.2 ch by vertically arranging two sound
bars having 5.1 ch and arranging them on the left and right.
Here, of the two sound bars, the sound bar disposed in the left layer is referred to as SoundBar_L,
and the sound bar disposed in the right layer is referred to as SoundBar_R.
[0047]
As shown in FIG. 7, the sound bar (SoundBar_L) has a speaker group that reproduces audio
corresponding to each of the five channels SR_L, R_L, C_L, L_L, and SL_L. FIG. 7 shows an
example in which speakers corresponding to each channel are individually provided. Here, each
channel (SR_L, SL_L (total 2)) shown by a broken line in FIG. 7 has virtual SP (SR_L, SL_L) as an
output position, and two surround waves of a sound bar (SoundBar_L) It is output from the
speaker group.
[0048]
03-05-2019
12
Similarly, as shown in FIG. 7, the sound bar (SoundBar_R) has a speaker group that reproduces
audio corresponding to each of the five channels SL_R, L_R, C_R, R_R, and SR_R. FIG. 7 shows an
example in which speakers corresponding to each channel are individually provided. Here, each
channel (SL_R, SR_R (two in total)) indicated by a broken line in FIG. 7 has virtual SP (SL_R, SR_R)
as an output position, and two surround of a sound bar (SoundBar_R) It is output from the
speaker group.
[0049]
The audio output device 1 includes an amplifier for driving the speaker configured as described
above.
[0050]
When downmixing audio data of 22.2 ch to audio data of 10.2 ch, the audio data generation unit
111 converts the left layer 7 ch of 22.2 ch to the left layer 5 ch of 10.2 ch (SR_L, R_L, C_L, L_L ,
Downmix to SL_L).
The audio data generation unit 111 merges 8 channels of 22.2 middle layers into 5 left channels
(SR_L, R_L, C_L, L_L, SL_L) and right channel 5 ch (SL_R, L_R, C_R, R_R, SR_R) of 10.2 channels. .
The audio data generation unit 111 merges the 22.2 ch right layer 7 ch with the 10.2 ch right
layer 5 ch (SL_R, L_R, C_R, R_R, SR_R).
[0051]
FIG. 8 is a diagram showing a formula of downmix according to the present embodiment.
[0052]
The voice data generation unit 111 expresses the sense of depth of voice by adjusting the
distribution coefficient of the middle layer.
a, k, v3, v4, v5 are common coefficients.
03-05-2019
13
[0053]
The coefficient a is a coefficient set to prevent data overflow during downmix processing, and a
value between 0 and 1 is arbitrarily set. The coefficient k is a surround channel gain coefficient,
which is a coefficient sent by broadcasting. The coefficients v3 and v4 are vertical coefficients
(vertical distribution), and the coefficient v5 is a vertical coefficient (middle distribution), and
values of 0 to 1 are arbitrarily set. Further, the relationship between the vertical direction
coefficients is v3 + v4 + v5 = 1. These coefficients can be arbitrarily set in order to optimize the
feeling of expansion in accordance with the position of the upper and lower SP (sound bar
installation position, television screen size).
[0054]
Equations 17 to 21 are equations for calculating audio data of the left layer of 10.2 ch. As shown
in Expression 17, the audio data generation unit 111 generates L_L using BtFL, BtFC, FL, FLC,
and FC. As shown in Expression 18, the audio data generation unit 111 generates R_L using
TpFL, TpFC, TpSiL, TpC, FL, FLC, and FC.
[0055]
As shown in Expression 19, the audio data generation unit 111 generates C_L using FL, FLC, FC,
and SiL. As shown in Expression 20, the audio data generation unit 111 generates SL_L using BL,
BC, and SiL. As shown in Expression 21, the audio data generation unit 111 generates SR_L using
TpBL, TpBC, TpSiL, and TpC.
[0056]
As shown in Expression 22, the audio data generation unit 111 generates a 10.2 ch LFE 1 using
the 22.2 ch LFE 1.
[0057]
03-05-2019
14
Formulas 23 to 27 are formulas for calculating audio data of the 10.2 ch right layer.
As shown in Expression 23, the audio data generation unit 111 generates L_R using TpFR, TpFC,
TpSiR, TpC, FR, FRC, and FC. As shown in Expression 24, the audio data generation unit 111
generates R_R using BtFR, BtFC, FR, FRC, and FC.
[0058]
As shown in Expression 25, the audio data generation unit 111 generates C_R using FR, FRC, FC,
and SiR. As shown in Expression 26, the audio data generation unit 111 generates SL_R using
TpBR, TpBC, TpSiR, and TpC. As shown in Expression 27, the audio data generation unit 111
generates SR_R using BR, BC, and SiR.
[0059]
As shown in Expression 28, the audio data generation unit 111 generates 10.2 ch LFE 2 using
22.2 ch LFE 2.
[0060]
[Effects of the Second Embodiment] According to the above configuration, by using the two
sound bars 12 and 13, the audio processing device 11 can easily realize a multi-channel audio
reproduction environment.
In particular, when downmixing 22.2ch audio data to 10.2ch audio data, downmix the 22.2ch left
layer audio data to 10.2ch left layer audio data, and Since the middle layer audio data is merged
with the 10.2ch left layer and right layer audio data, and the 22.2ch right layer audio data is
downmixed to the 10.2ch right layer audio data, the 22.2ch The audio data of three layers (left,
middle and right) can be easily reproduced by two sound bars 12, 13 (left and right).
[0061]
EMBODIMENT 3 It will be as follows if Embodiment 3 of this invention is demonstrated based on
03-05-2019
15
FIGS. 9-11. In addition, about the member which has the same function as the member
demonstrated in the said embodiment for convenience of explanation, the same code | symbol is
appended and the description is abbreviate | omitted.
[0062]
The audio processing apparatus 11 generates audio data of 8.1 ch by acquiring audio data of
22.2 ch and downmixing.
[0063]
In the following description, each channel of audio data and the speaker position assumed by the
channel will be described using the same symbol.
[0064]
FIG. 9 is a view showing a speaker position assumed by the 8.1ch audio data generated by the
audio processing device 11 according to the present embodiment.
As shown in FIG. 9, the voice data of 8.1 ch is composed of six SPs (R_L, C_L, L_L, L_R, C_R, R_R)
placed before, and two virtual SPs placed behind (S_L). , S_R) are output as output positions.
[0065]
The audio output device 1 can output 8.1 channel audio data by disposing the SPs on the front
left, right, upper, lower, and the surround SP on the rear left and right in the television receiver 2.
Here, the sound bar arranged in the left layer is represented as SoundBar_L, and the sound bar
arranged in the right layer is represented as SoundBar_R.
[0066]
03-05-2019
16
As shown in FIG. 9, the sound bar (SoundBar_L) has a speaker group that reproduces audio
corresponding to each of four channels of R_L, C_L, L_L, and S_L. FIG. 9 shows an example in
which speakers corresponding to each channel are individually provided. Here, the channel (S_L)
indicated by a broken line in FIG. 9 has the virtual SP (S_L) as an output position, and is output
from one surround speaker group possessed by the sound bar (SoundBar_L).
[0067]
Similarly, as shown in FIG. 9, the sound bar (SoundBar_R) has a speaker group that reproduces
audio corresponding to each of the four channels L_R, C_R, R_R, and S_R. FIG. 9 shows an
example in which speakers corresponding to each channel are individually provided. Here, a
channel (S_R) indicated by a broken line in FIG. 9 has a virtual SP (S_R) as an output position, and
is output from one surround speaker group included in the sound bar (SoundBar_R).
[0068]
The audio output device 1 includes an amplifier for driving the speaker configured as described
above.
[0069]
When downmixing audio data of 22.2 ch to audio data of 8.1 ch, the audio data generation unit
111 converts the left layer 7 ch of 22.2 ch to the left layer 4 ch of 8.1 ch (R_L, C_L, L_L, S_L
Downmix to
The audio data generation unit 111 merges 8 ch of the 22.2 middle layers into the left 4 ch (R_L,
C_L, L_L, S_L) and the right 4 ch (L_R, C_R, R_R, S_R). The audio data generation unit 111
downmixes the 22.2ch right layer 7ch to the 8.1ch right layer 4ch (L_R, C_R, R_R, S_R).
[0070]
FIG. 10 is a diagram showing a formula of downmix according to the present embodiment.
[0071]
03-05-2019
17
The voice data generation unit 111 adjusts the distribution coefficient of the middle layer to
express the sense of voice spread.
a, k, v3 and v4 are general coefficients.
[0072]
The coefficient a is a coefficient set to prevent data overflow during downmix processing, and a
value between 0 and 1 is arbitrarily set. The coefficient k is a surround channel gain coefficient,
which is a coefficient sent by broadcasting. The coefficients v3 and v4 are vertical direction
coefficients (vertical distribution), and values of 0 to 1 are arbitrarily set. Further, the relationship
between the vertical direction coefficients is v3 + v4 = 1. These coefficients can be arbitrarily set
in order to optimize the feeling of expansion in accordance with the position of the upper and
lower SP (sound bar installation position, television screen size).
[0073]
Formulas 29 to 33 are formulas for calculating sound data of the left layer of 8.1 ch. As shown in
Expression 29, the audio data generation unit 111 generates L_L using BtFL, BtFC, FL, FLC, and
FC. As shown in Expression 30, the audio data generation unit 111 generates R_L using TpFL,
TpFC, TpSiL, TpC, FL, FLC, and FC.
[0074]
As shown in Expression 31, the audio data generation unit 111 generates C_L using FL, FLC, FC,
and SiL. As shown in Expression 32, the audio data generation unit 111 generates S_L using BL,
BC, SiL, TpFR, TpFC, TpSiR, and TpC.
[0075]
As shown in Expression 33, the audio data generation unit 111 generates LFE using LFE 1 and
LFE 2.
03-05-2019
18
[0076]
Equations 34 to 37 are equations for calculating the sound data of the 8.1 ch right layer.
As shown in Expression 34, the audio data generation unit 111 generates L_R using TpFR, TpFC,
TpSiR, TpC, FR, FRC, and FC. As shown in Expression 35, as shown in Expression 35, the audio
data generation unit 111 generates R_R using BtFR, BtFC, FR, FRC, and FC. As shown in
Expression 36, the audio data generation unit 111 generates C_R using FR, FRC, FC, and SiR. As
shown in Expression 37, the audio data generation unit 111 generates S_R using BR, BC, SiR,
TpBR, TpBC, TpSiR, and TpC.
[0077]
FIG. 11 is a view schematically showing a sound field generated in the case of reproduction by
two right and left sound bars corresponding to 8.1 ch audio data according to the present
embodiment.
[0078]
In FIG. 11, SF_SoundBar_L schematically shows a sound field generated by the sound bar
(SoundBar_L), and SF_SoundBar_R schematically shows a sound field generated by the sound bar
(SoundBar_R).
[0079]
[Effect of Third Embodiment] According to the above configuration, the audio processing device
11 can easily realize a multi-channel audio reproduction environment by using the two sound
bars 12 and 13.
In particular, when downmixing 22.2ch audio data to 8.1ch audio data, downmix the 22.2ch left
layer audio data to the 8.1ch left layer audio data, and Since the middle layer audio data is
merged with the 8.1ch left layer and right layer audio data, and the 22.2ch right layer audio data
is downmixed to the 8.1ch right layer audio data, the 22.2ch The audio data of three layers (left,
middle and right) can be easily reproduced by two sound bars 12, 13 (left and right).
03-05-2019
19
[0080]
EMBODIMENT 4 It will be as follows if Embodiment 3 of this invention is demonstrated based on
FIG. 12, FIG.
In addition, about the member which has the same function as the member demonstrated in the
said embodiment for convenience of explanation, the same code | symbol is appended and the
description is abbreviate | omitted.
[0081]
The audio processing apparatus 11 generates audio data of 6.1 ch by acquiring audio data of
22.2 ch and downmixing.
[0082]
In the following description, each channel of audio data and the speaker position assumed by the
channel will be described using the same symbol.
[0083]
FIG. 12 is a diagram showing a speaker position assumed by 6.1-ch audio data generated by the
audio processing device 11 according to the present embodiment.
As shown in FIG. 12, the audio data of 6.1 ch includes four SPs (R_L, L_L, L_R, R_R) arranged in
front, and two virtual SPs (SL_L, SR_R) arranged in the rear. Output as an output position.
[0084]
The audio output device 1 can output 6.1 channel audio data by arranging the SP on the front
left, right, upper and lower and the surround SP on the rear left and right in the television
receiver 2.
03-05-2019
20
Here, of the two sound bars, the sound bar disposed in the left layer is referred to as SoundBar_L,
and the sound bar disposed in the right layer is referred to as SoundBar_R.
[0085]
As shown in FIG. 12, the sound bar (SoundBar_L) has a speaker group that reproduces audio
corresponding to each of the four channels of R_L, L_L, and SL_L. FIG. 12 shows an example in
which speakers corresponding to each channel are individually provided. Here, the channel
(SL_L) indicated by a broken line in FIG. 12 has the virtual SP (SL_L) as an output position, and is
output from one surround speaker group possessed by the sound bar (SoundBar_L).
[0086]
Similarly, as shown in FIG. 12, the sound bar (SoundBar_R) has a speaker group that reproduces
audio corresponding to each of the four channels L_R, R_R, and SR_R. FIG. 12 shows an example
in which speakers corresponding to each channel are individually provided. Here, a channel
(SR_R) indicated by a broken line in FIG. 12 has a virtual SP (SR_R) as an output position, and is
output from one surround speaker group included in the sound bar (SoundBar_R).
[0087]
The audio output device 1 includes an amplifier for driving the speaker configured as described
above.
[0088]
When downmixing audio data of 22.2 ch to audio data of 6.1 ch, the audio data generation unit
111 converts 7 ch of left layer of 22.2 ch into 3 ch of left layer of 6.1 ch (R_L, L_L, SL_L)
Downmix.
The audio data generation unit 111 merges the 8 channels of the 22.2 middle layers into the 3
left channels (R_L, L_L, SL_L) and the right 3 channels (L_R, R_R, SR_R). The audio data
generation unit 111 downmixes the 22.2ch right layer 7ch to the 6.1ch right layer 3ch (L_R, R_R,
03-05-2019
21
SR_R).
[0089]
FIG. 13 is a diagram showing a calculation formula of downmix according to the present
embodiment.
[0090]
The voice data generation unit 111 adjusts the distribution coefficient of the middle layer to
express the sense of voice spread.
a, k, v1, v2 are common coefficients.
[0091]
The coefficient a is a coefficient set to prevent data overflow during downmix processing, and a
value between 0 and 1 is arbitrarily set. The coefficient k is a surround channel gain coefficient,
which is a coefficient sent by broadcasting. The coefficients v1 and v2 are vertical direction
coefficients (vertical distribution), and values from 0 to 1 are arbitrarily set. Further, the
relationship between the vertical direction coefficients is v1 + v2 = 1. These coefficients can be
arbitrarily set in order to optimize the feeling of expansion in accordance with the position of the
upper and lower SP (sound bar installation position, television screen size).
[0092]
Expressions 38 to 40 are calculation expressions of audio data of the left layer of 6.1 ch. As
shown in Expression 38, the audio data generation unit 111 generates L_L using BtFL, BtFC, FL,
FLC, and FC. As shown in Expression 39, the audio data generation unit 111 generates R_L using
TpFL, TpFC, TpSiL, TpC, FL, FLC, and FC. As shown in Expression 40, the audio data generation
unit 111 generates SL_L using BL, BC, SiL, TpFR, TpFC, TpSiR, and TpC.
[0093]
03-05-2019
22
As shown in Expression 41, the audio data generation unit 111 generates LFE using LFE1 and
LFE2.
[0094]
Formulas 42 to 44 are formulas for calculating the sound data of the 6.1 ch right layer.
As shown in Expression 42, the audio data generation unit 111 generates L_R using TpFR, TpFC,
TpSiR, TpC, FR, FRC, and FC. As shown in Expression 43, the audio data generation unit 111
generates R_R using BtFR, BtFC, FR, FRC, and FC. As shown in Expression 44, the audio data
generation unit 111 generates SR_R using BR, BC, SiR, TpBR, TpBC, TpSiR, and TpC.
[0095]
[Effects of Fourth Embodiment] According to the above configuration, the audio processing
device 11 can easily realize a multi-channel audio reproduction environment by using the two
sound bars 12 and 13. In particular, when downmixing 22.2ch audio data to 6.1ch audio data,
downmix the 22.2ch left layer audio data to the 6.1ch left layer audio data, and Since the middle
layer audio data is merged with the 6.1ch left layer and right layer audio data, and the 22.2ch
right layer audio data is downmixed to the 6.1ch right layer audio data, the 22.2ch The audio
data of three layers (left, middle and right) can be easily reproduced by two sound bars 12, 13
(left and right).
[0096]
[Example of Implementation by Software] The control block (audio data generation unit 111) of
the audio processing device 11 may be realized by a logic circuit (hardware) formed in an
integrated circuit (IC chip) or the like, or a CPU (Central Processing Unit). It may be realized by
software using Processing Unit).
[0097]
In the latter case, the voice processing device 11 is a CPU that executes instructions of a program
that is software that realizes each function, a ROM (Read Only Memory) in which the program
03-05-2019
23
and various data are readably recorded by a computer (or CPU). Alternatively, a storage device
(these are referred to as “recording media”), a RAM (Random Access Memory) for developing
the program, and the like are provided.
The object of the present invention is achieved by the computer (or CPU) reading the program
from the recording medium and executing the program. As the recording medium, a “nontransitory tangible medium”, for example, a tape, a disk, a card, a semiconductor memory, a
programmable logic circuit or the like can be used. The program may be supplied to the
computer via any transmission medium (communication network, broadcast wave, etc.) capable
of transmitting the program. Note that one aspect of the present invention can also be realized in
the form of a data signal embedded in a carrier wave in which the program is embodied by
electronic transmission.
[0098]
[Summary] The audio processing device (11) according to aspect 1 of the present invention is an
audio processing device that generates audio data of a plurality of output channels by
downmixing audio data of a plurality of input channels. A first sound bar in which a plurality of
speaker groups outputting the audio data of the second line are arranged in the vertical direction,
and a second sound bar in which the plurality of speaker groups outputting the audio data for
each output channel are arranged in the vertical direction An audio data generation unit (111) is
provided which generates audio data for each of a plurality of output channel speakers including
a bar by downmixing audio data of the plurality of input channels.
[0099]
According to the above configuration, it is possible to easily realize a multi-channel audio
reproduction environment by using two sound bars.
[0100]
In the audio processing device according to aspect 2 of the present invention, in the aspect 1, the
audio data generation unit reduces audio data of the left layer of the plurality of input channels
to audio data of the left layer of the plurality of output channels. Mixing and merging the middle
layer audio data of the plurality of input channels into the left layer and right layer audio data of
the plurality of output channels; and outputting the plurality of audio data of the right layers of
the plurality of input channels. It may be downmixed to audio data of the right layer of the
channel.
03-05-2019
24
[0101]
According to the above configuration, the audio data of the left layer and the right layer of the
plurality of input channels are downmixed to the audio data of the left layer and the right layer of
the plurality of output channels, respectively, Is merged with audio data of left and right layers of
multiple output channels, so audio data of three layers (left, middle, right) of multiple input
channels can be easily reproduced with two sound bars (left and right) be able to.
[0102]
In the audio processing device according to aspect 3 of the present invention, in the above aspect
2, when the audio data generation unit downmixes 22.2 ch audio data to 14.2 ch audio data, The
audio data of the left layer 7 ch is downmixed to the audio data of the left layer 7 ch of the
plurality of output channels, and the audio data of the middle layer 8 ch of the plurality of input
channels is divided into the left layer 6 ch and the right layer of the plurality of output channels
The data may be merged with 6ch audio data, and the audio data of the right layer 7ch of the
plurality of input channels may be downmixed to the audio data of the right layer 7ch of the
plurality of output channels.
[0103]
According to the above configuration, in the case of downmixing audio data of 22.2 ch to audio
data of 14.2 ch, audio data of three layers (left, middle and right) can be easily by two sound bars
(left and right) It can be reproduced.
[0104]
The audio output device (1) according to aspect 4 of the present invention includes the audio
processing device (11) according to any of the above aspects 1 to 3 and a plurality of speaker
groups for outputting audio data for each output channel arranged in the longitudinal direction
And a second sound bar (13) in which a plurality of speaker groups outputting voice data for
each output channel are arranged in the longitudinal direction.
[0105]
According to the above configuration, multi-channel audio can be downmixed and output by
using two sound bars.
[0106]
A television receiver (2) according to aspect 5 of the present invention includes the audio output
device according to aspect 4 above.
03-05-2019
25
[0107]
An audio processing method according to a sixth aspect of the present invention is an audio
processing method for generating audio data of a plurality of output channels by downmixing
audio data of a plurality of input channels, the audio data of each output channel being generated
A first sound bar in which a plurality of speaker groups to be output are arranged in the vertical
direction, and a second sound bar in which a plurality of speaker groups to output audio data for
each output channel are arranged in the vertical direction An audio data generation step of
generating audio data for each of a plurality of output channel speakers by downmixing the
audio data of the plurality of input channels is included.
[0108]
The voice processing apparatus according to each aspect of the present invention may be
realized by a computer. In this case, the voice processing apparatus is used as a computer by
operating the computer as each unit (software element) included in the voice processing
apparatus. A program of an audio processing device to be realized and a computer readable
recording medium recording the same also fall within the scope of the present invention.
[0109]
The present invention is not limited to the above-described embodiments, and various
modifications can be made within the scope of the claims, and embodiments obtained by
appropriately combining the technical means disclosed in the different embodiments. Is also
included in the technical scope of the present invention.
Furthermore, new technical features can be formed by combining the technical means disclosed
in each embodiment.
[0110]
1 audio output device 11 audio processing device 12 surround bar (first sound bar) 13 surround
bar (second sound bar) 111 audio data generation unit 2 television receiver
03-05-2019
26
Документ
Категория
Без категории
Просмотров
0
Размер файла
37 Кб
Теги
jp2018121310
1/--страниц
Пожаловаться на содержимое документа