JP2017028679

Patent Translate
Powered by EPO and Google
Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate,
complete, reliable or fit for specific purposes. Critical decisions, such as commercially relevant or
financial decisions, should not be based on machine-translation output.
DESCRIPTION JP2017028679
The present invention relates to a system for providing personalized audio, comprising a memory
and a processor. A system for receiving a plurality of audio content and playing a first audio
content from a first user device for playing a first audio content of the plurality of audio content
using a plurality of speakers. Receiving a request, obtaining a first position of a first user of a first
user device relative to each speaker of the plurality of speakers, and using the plurality of
speakers and object-based audio, each speaker of the plurality of speakers It is configured to play
the first audio content of the plurality of audio content based on the first position of the first user
of the first user device with respect to. [Selected figure] Figure 1
System and method for providing personalized audio
[0001]
The provision of improved quality audio has been significantly improved with the availability of
sound bars, 5.1Ch surround sound and 7.1Ch surround sound. Systems that provide these
quality-improved audios divide the audio into multiple audio channels, and through the multiple
speakers located at different locations around the listener, the multiple audio channels. By
playing back, the quality of the audio presentation has been improved. Existing surround sound
technologies improve the perception of stereophonic sound by exploiting the listener's ability to
identify the location or source of the detected sound from the direction and distance, sound
localization.
[0002]
03-05-2019
1
The invention is personalized, substantially as shown in at least one drawing and / or described
in connection with at least one drawing, and more particularly in the claims. Address systems and
methods for providing audio.
[0003]
1 illustrates an exemplary system for providing personalized audio according to one embodiment
of the present invention.
FIG. 2 illustrates an exemplary environment using the system of FIG. 1 according to one
embodiment of the present invention. FIG. 6 illustrates another exemplary environment using the
system of FIG. 1 in accordance with one embodiment of the present invention. Fig. 6 shows an
exemplary flow chart of a method for providing personalized audio according to one embodiment
of the present invention.
[0004]
The following description contains specific information pertaining to the embodiments of the
present invention. The drawings in the present application and the accompanying detailed
description are directed to merely exemplary embodiments. Similar components or
corresponding components between the various figures may be identified by similar reference
numerals or corresponding reference numerals unless otherwise indicated. Moreover, what is
depicted in the drawings of the present application is generally not to scale and is not intended to
correspond to actual relative dimensions.
[0005]
FIG. 1 illustrates an exemplary system 100 for providing personalized audio according to one
embodiment of the present invention. As shown, system 100 includes user device 105, audio
content 107, media device 110, and a plurality of speakers 197a, 197b,..., 197n. Media device
110 includes a processor 120 and a memory 130. Processor 120 is a hardware processor, such
as a central processing unit (CPU) used in a computing device. Memory 130 is a non-transitory
storage device for storing computer code executed by processor 120 and for storing various data
and parameters.
03-05-2019
2
[0006]
The user device 105 may be a handheld personal device such as a mobile phone, a tablet
computer or the like. User device 105 may be connected to media device 110 via connection
155. In some embodiments, the user device 105 can be wirelessly utilized, and the user device
105 can be wirelessly connected to the media device 110 using wireless technology such as
Bluetooth, WiFi, etc. It can be configured. Additionally, the user device 105 can include a
software application that provides the user with a plurality of selectable audio profiles, and can
implement the user's selection of audio language and listening mode. A dialog represents an
audio of words uttered from the mouth, such as speech, opinions or narrations, and is also
considered to include interactions between two or more actors or characters.
[0007]
Audio content 107 may include media sources such as television programs, movies, audio tracks
derived from music files, or audio tracks derived from other media sources including audio
portions. In some embodiments, audio content 107 may include a single track having all the
audio from one media source, and audio content 107 may be separate from audio content 107. It
may be a plurality of tracks including a plurality of parts of For example, a movie can include
audio content for dialog, audio content for music, and audio content for effects. In some
embodiments, audio content 107 may include multiple dialog content, each dialog content
including dialogs in different languages. One user can select one language for the dialog, or
multiple users can select multiple languages for the dialog.
[0008]
Media device 110 may be configured to be connected to a plurality of speakers, such as speakers
197a, speakers 197b, ..., and speakers 197n. Media device 110 may be a computer, set top box,
DVD player or any other media device suitable for playing audio content 107 using a plurality of
speakers. In some embodiments, media device 110 may be configured to be connected to
multiple speakers via a wired or wireless connection.
[0009]
03-05-2019
3
In one embodiment, audio content 107 may be provided on multiple channels, such as 2Ch
stereo or 5.1Ch surround sound. In other embodiments, audio content 107 may be provided from
the perspective of an object, also known as object-based audio or sound. In such embodiments,
the tracks of the individual instruments of one song are not mixed, or the dialogs in the ambient
sound, sound effects and audio tracks of the movie are not mixed, but rather Can be directed
precisely to one or more of the plurality of speakers 197a-197n, and also managed at how loud
they can be played. For example, audio content 107 may be generated as metadata and
instructions regarding how and where all audio parts are to be played. The media device 110
may then use those metadata and instructions to play audio on the plurality of speakers 197a197n.
[0010]
As shown in FIG. 1, memory 130 of media device 110 includes audio application 140. Audio
application 140 is a computer algorithm for providing personalized audio and is stored in
memory 130 for execution by processor 120. In some embodiments, audio application 140 may
include position module 141 and audio profile 143. The audio application 140 may customize
the audio personalized to each listener to one or more listeners at different positions relative to
the plurality of speakers 197a, 197b, ..., 197n. The audio profile 143 can be used to supply based
on the profile.
[0011]
The audio application 140 also includes a position module 141, which is a computer code
module for obtaining the position of the user device 105 and the positions of other user devices
(not shown) in the room or theater. . In some embodiments, obtaining the location of the user
device 105 can include the transmission of a calibration signal by the media device 110. The
calibration signal may include audio signals sent from the plurality of speakers 197a, 197b, ...,
197n. In response, the user device 105 can detect calibration signals sent from each of the
plurality of speakers 197a, 197b, ..., 197n using a microphone (not shown), Triangulation
techniques can also be used to determine the position of the user device 105 based on the
location of the user device 105 relative to each of the plurality of speakers 197a, 197b,..., 197n.
In some embodiments, position module 141 may use one or more cameras (not shown) of system
100 to determine the position of user device 105. Thus, the position of each user relative to each
of the plurality of speakers 197a, 197b,..., 197n can be determined.
03-05-2019
4
[0012]
Audio application 140 also includes an audio profile 143, which includes a plurality of
predefined listening modes that may be options for different audio content. For example, the
audio profile 143 may include a listening mode having an equalizer setting that may be optional
for a movie. Equalizer settings include, for example, a reduction in bass sound, and an increase in
treble frequency that improves the quality of playback of the movie dialog for listeners with
hearing problems. The audio profile 143 may include a particular genre of program, such as a
listening mode optimized for drama and action, a custom listening mode, and a standard listening
mode that does not significantly change the audio. In some embodiments, the custom listening
mode may allow the user to enhance portions of the audio content 107, such as music, dialogs
and / or effects. Enhancing a portion of the audio content 107 may include raising or lowering
the volume of that portion of the audio content 107 relative to the other portions of the audio
content 107. Enhancing a portion of the audio content 107 can include changing the equalizer
settings to make the volume of that portion of the audio content 107 more loud. Audio profile
143 may include the language in which the user listens to the dialog. In some embodiments,
audio profile 143 may include multiple languages, and the user may select one language when
listening to the dialog.
[0013]
The plurality of speakers 197a, 197b,..., 197n may be surround sound speakers, or may be other
speakers suitable for providing audio selected from a plurality of audio contents 107. A plurality
of speakers 197a, 197b,..., 197n may be connected to the media device 110 using speaker lines,
or may be connected to the media device 110 using wireless technology. The speaker 197 may
be a mobile speaker so that the user can change the position of one or more of the plurality of
speakers 197a, 197b, ..., 197n. In some embodiments, the positions of the speakers 197a-197n
and the interference between the audio transmitted from each of the speakers 197a-197n are
used to generate a virtual speaker, the sound being virtual Speakers 197a-197n can be used to
illusion that they are derived from the speakers. In other words, the virtual speaker is a speaker
that is not physically present at a position where sound is considered to be generated.
[0014]
03-05-2019
5
FIG. 2 illustrates an exemplary environment 200 using the system 100 of FIG. 1, in accordance
with one embodiment of the present invention. The user 211 owns the user device 205a, and the
user 212 owns the user device 205b. In some embodiments, user device 205 a may be co-located
with user 211, and user device 205 b may be co-located with user 212. Thus, when media device
210 obtains the position of user device 205a relative to speakers 297a-297e, media device 210
may obtain the position of user 211 relative to speakers 297a-297e. Similarly, once media device
210 obtains the position of user device 205b relative to speakers 297a-297e, media device 210
may obtain the position of user 212 relative to speakers 297a-297e.
[0015]
The user device 205a can determine the relative position with respect to the speakers 297a to
297e by triangulation. For example, user device 205a may use the microphone of user device
205a to receive audio calibration signals from speaker 297a, speaker 297b, speaker 297d and
speaker 297e. Based on the received audio calibration signal, the user device 205a may
determine the position of the user device 205a relative to the speakers 297a-297e, for example
by triangulation. User device 205a may be connected to media device 210 via connection 255a
as shown. In some embodiments, user device 205a may transmit the determined location to
media device 210. User device 205b may receive audio calibration signals from speaker 297a,
speaker 297b, speaker 297c, and speaker 297e using the microphone of user device 205b.
Based on the received audio calibration signal, the user device 205b may determine the position
of the user device 205b relative to the speakers 297a-297e, for example by triangulation. In
some embodiments, user device 205b may be connected to media device 210 via connection
255b as shown. In some embodiments, user device 205b may transmit its location to media
device 210 via connection 255b. In another embodiment, the user device 205b receives a
calibration signal and transmits information to the media device 210 via connection 255b, for
example to determine the position of the user device 205b by triangulation. Can.
[0016]
FIG. 3 illustrates an exemplary environment 300 using the system 100 of FIG. 1 in accordance
with one embodiment of the present invention. It should be mentioned that the audio is explicitly
shown to be provided to the user 311 and the user 312. User devices 205a and 205b are not
shown in FIG. As shown in FIG. 3, the user 311 is in a first position and receives a first audio
content 356. The user 312 is in the second position and receives the second audio content 358.
03-05-2019
6
[0017]
The first audio content 356 may include a dialog of the language selected by the user 311, and
may also include other audio content such as music and effects. In some embodiments, user 311
can select a regular audio profile. Here, the normal audio profile represents the selection to
provide audio to the user 311 at a level unchanged from the audio content 107. The second
audio content 358 may include a dialog of the language selected by the user 312, and may also
include other audio content, such as music and effects. In some embodiments, the user 312 can
select a regular audio profile. Here, the normal audio profile represents the selection to provide
the audio portion to the user 312 at a level unchanged from the audio content 107.
[0018]
Each of the speakers 397a to 397e can transmit a cancellation audio 357. The cancellation audio
357 can cancel out part of the audio content transmitted from the speaker 397a, the speaker
397b, the speaker 397c, the speaker 397d, and the speaker 397e. In some embodiments, the
cancellation audio 357 can completely cancel out a portion of the first audio content 356 or a
portion of the second audio content 358. For example, if the first audio content 356 includes a
dialog in a first language and the second audio content 358 includes a dialog in a second
language, then the cancellation audio 357 is a first audio content. The first language portion of
the audio content 356 can be completely cancelled, so that the user 312 receives only the second
language dialog. In some embodiments, the cancellation audio 357 may partially cancel out a
portion of the first audio content 356 or a portion of the second audio content 358. For example,
the first audio content 356 includes the dialog of the first language at an enhanced level, and the
second audio content 358 includes the dialog of the first language at the normal level. If so, the
cancellation audio 357 may provide a dialog to the user 312 at an appropriate level, partially
canceling the dialog portion of the first audio content 356.
[0019]
FIG. 4 exemplarily shows a flowchart 400 of a method for providing personalized audio
according to one embodiment of the present invention. The method starts in step 401, where the
audio application receives audio content 107. In some embodiments, audio content 107 may
include multiple audio tracks, such as music tracks, dialog tracks, effects tracks, ambient sound
tracks, background sound tracks, and the like. In another embodiment, the audio content 107
may include all audio associated with the media played to the user in one audio track.
03-05-2019
7
[0020]
At step 402, the media device 110 receives a first play request from a first user device for
playing the first audio content of the audio content 107 using the speaker 197. In some
embodiments, the first user device is suitable for sending a smartphone, a tablet computer, or a
playback request to the media device 110 and for receiving a calibration signal sent from the
media device 110 , And other handheld devices including a microphone. The first playback
request may be a wireless signal transmitted from the first user device to the media device 110.
In some embodiments, media device 110 may send a signal to user device 105 prompting the
user to launch application software on user device 105. The application software can be used in
determining the position of the user device 105, and the user can use the application software to
select audio settings, such as language and audio profile.
[0021]
At step 403, the media device 110, in response to the first playback request, obtains a first
position of a first user of the first user device relative to each of the plurality of speakers. In some
embodiments, user device 105 can include a calibration application for use with audio
application 140. After activation of the calibration application, user device 105 may receive a
calibration signal from media device 110. The calibration signal may be an audio signal
transmitted from a plurality of speakers, eg, speaker 197, and user device 105 uses the
calibration signal to compare the user device 105 relative to each speaker of the plurality of
speakers 197. The position of can be determined. In some embodiments, user device 105
provides media device 110 with a position relative to each speaker. In another embodiment, the
user device 105 can use the microphone of the user device 105 to receive calibration signals and
transmit information to the media device 110 for processing. In some embodiments, media
device 110 may determine the position of user device 105 relative to speaker 197 based on the
information received from user device 105.
[0022]
The calibration signal sent from media device 110 may be sent using speaker 197. In some
embodiments, the calibration signal may be an audio signal that can be heard by humans, eg, an
audio signal between about 20 Hz and about 20 kHz, or an audio that can not be heard by
03-05-2019
8
humans. It may be a signal, for example an audio signal having a frequency higher than about 20
kHz. In order to determine the position of the user device 105 relative to each speaker of the
plurality of speakers 197, the speakers 197a-197n may transmit calibration signals at various
times, or the plurality of speakers 197 may be Calibration signals can be transmitted
simultaneously. In some embodiments, the calibration signal transmitted from each speaker of
the plurality of speakers 197 is a unique calibration signal that allows the user device 105 to
distinguish each calibration signal sent from each speaker 197a-197n. You may be there.
Calibration signals may be used to determine the position of user device 105 relative to each
speaker 197a-197n, and also update the position of user device 105 relative to each speaker
197a-197n. The calibration signal can be used to
[0023]
In some embodiments, the speaker 197 may be a wireless speaker or a mobile speaker that can
be repositioned by the user. Therefore, the position of each of the plurality of speakers 197a to
197n can be changed, and the distance between each of the plurality of speakers 197a to 197n
can also be changed. The calibration signal may be used to determine the relative position of
each speaker 197a-197n and / or the distance between each speaker 197a-197n. The calibration
signal can be used to update the relative position of each speaker 197a-197n and / or the
distance between each speaker 197a-197n.
[0024]
Alternatively, system 100 may use a camera to obtain, calculate and / or track the position of one
or more users. In some embodiments, system 100 can include a camera, such as a digital camera.
The system 100 can obtain the position of the user device 105 and subsequently determine the
position of the user by mapping the position of the user device 105 to the image captured by the
camera. In some embodiments, system 100 can use camera and recognition software, such as
face recognition software, to obtain the position of the user.
[0025]
Once the system 100 has obtained the position of the user, the system 100 can use the camera
to continuously track the position of the user and / or periodically update the position of the
user. Since it is also conceivable for the user to move during playback of the audio content 107, it
03-05-2019
9
may be useful to track the user's position continuously or to periodically update the user's
position. For example, a user watching a movie may have changed in position after coming back
for snacks. By tracking and / or updating the position of the user, the system 100 can continue to
provide the user with personalized audio throughout the movie playback period. In some
embodiments, system 100 is configured to detect when a user or user device has left an audio
playback environment, such as a room. In response, system 100 can suspend transmission of
personalized audio corresponding to the user until the user returns to the room. The system 100
can prompt the user to update their location if the user moves. In order to update the position of
the user, the media device 110 can transmit a calibration signal, for example, a signal with a
frequency higher than 20 kHz, to obtain the updated position of the user.
[0026]
Additionally, a calibration signal can be used to determine the audio quality of the room, for
example, the position of the wall relative to the speaker 197 and the shape of the room. The
system 100 can use the calibration signal to determine the position of the wall and can
determine how the sound echoes in the room. In some embodiments, the wall can be used as
another sound source. For example, the wall and its configuration may be considered for echo
reduction or cancellation, rather than canceling the echo or in conjunction with the echo
cancellation. The system 100 can also determine other factors that affect how sound travels in
the surrounding environment, such as the humidity in the air.
[0027]
At step 404, media device 110 receives a first audio profile from a first user device. The audio
profile may include a user selection that determines the personalized audio provided to the user.
For example, the audio profile may include language selection and / or listening mode. In some
embodiments, audio content 107 may include one dialog track of one language, or may include
multiple dialog tracks of different languages. The user of the user device 105 can select the
language in which to listen to the dialog track, and the media device 110 provides the first user
with personalized audio, including the dialog of the selected language Can. The language that the
first user listens to can include the original language of the media being played, or the language
the first user listens to is a different language than the original language of the media being
played back and Also good.
[0028]
03-05-2019
10
The listening mode may include settings designed to enhance the quality of the user's listening
experience, and may use different listening modes depending on different situations. System 100
can include an enhanced dialog listening mode, an action program, a listening mode for a drama
program, or other genre specific listening modes, a standard listening mode, and a custom
listening mode. The standard listening mode can provide audio supplied with the original media
content. The custom listening mode also allows the user to define the portions of the audio
content 107 to be enhanced, such as music, dialogs and effects.
[0029]
At step 405, the media device 110 receives a second play request from a second user device for
playing the second audio content of the plurality of audio content using the plurality of speakers.
In some embodiments, the second user device is suitable for sending a smartphone, a tablet
computer, or a playback request to the media device 110 and for receiving a calibration signal
sent from the media device 110 , And other handheld devices including a microphone. The
second playback request may be a wireless signal transmitted from the second user device to the
media device 110.
[0030]
At step 406, the media device 110, in response to the second playback request, obtains a second
position of a second user of the second user device relative to each speaker of the plurality of
speakers. In some embodiments, the second user device can include a calibration application for
use with the audio application 140. After launching the calibration application, the second user
device may receive a calibration signal from the media device 110. The calibration signal may be
an audio signal transmitted from a plurality of speakers, eg, speaker 197, and the second user
device uses the calibration signal to compare the user relative to each speaker of the plurality of
speakers 197 The position of the device 105 can be determined. In some embodiments, a second
user device can provide the media device 110 with a position relative to each speaker. In another
embodiment, the second user device may transmit information to the media device 110
regarding receipt of the calibration signal, and the media device 110 may be configured to send
the second user relative to each speaker 197 The position of can be determined.
[0031]
03-05-2019
11
At step 407, media device 110 receives a second audio profile from a second user device. The
second audio profile may include language selection and / or a second listening mode. After
receiving the second audio profile, at step 408, the media device 110 selects a first listening
mode based on the first audio profile and also selects a second listening mode based on the
second audio profile Do. In some embodiments, the first listening mode and the second listening
mode may be the same listening mode or different listening modes. Proceeding to step 409, the
media device 110 selects a first language based on the first audio profile and also selects a
second language based on the second audio profile. In some embodiments, the first language
may be the same language as the second language, and the first language and the second
language may be different languages.
[0032]
In step 410, the system 100 determines a plurality of first audio content based on the first audio
profile and the first position of the first user of the first user device relative to each of the
plurality of speakers. Play the audio content of The system 100 also determines the second audio
content of the plurality of audio content based on the second audio profile and a second position
of a second user of the second user device relative to each of the plurality of speakers.
Reproduce. In some embodiments, the first audio content of the plurality of audio content played
by the plurality of speakers may include the first dialog of the first language, and the plurality of
speakers play back the plurality of speakers. The second audio content of the plurality of audio
content may include a second dialog in a second language.
[0033]
The first audio content may include cancellation audio that cancels at least a portion of the
second audio content played by the plurality of speakers 197. In some embodiments, the
cancellation audio may partially cancel or completely cancel a portion of the second audio
content played by the plurality of speakers 197. To verify the effect of the cancellation audio, the
system 100 uses the user device 105 to see if the user hears an audio track that the user should
not hear, eg, for the user other than the selected language The user can be prompted to indicate
whether they are hearing a language dialog. In some embodiments, the user can be prompted to
provide additional subjective feedback, ie, whether the music is loud enough.
03-05-2019
12
[0034]
From the above description, it is clear that various techniques can be used to implement the
concepts disclosed herein without departing from the concepts of the present invention.
Moreover, while various concepts have been described in connection with particular
embodiments, various modifications in form and detail may be made by those skilled in the art
without departing from the concepts of the present invention. I understand that. For example, the
embodiments described above should be considered as illustrative in all aspects and not
restrictive. Furthermore, the present invention is not limited to the specific embodiments
described above, and it is understood that various reorganizations, modifications and
substitutions are possible without departing from the spirit of the present invention.
03-05-2019
13