US20100135510A1 - Apparatus for generating and playing object based audio contents - Google Patents

Apparatus for generating and playing object based audio contents Download PDF

Info

Publication number
US20100135510A1
US20100135510A1 US12/628,317 US62831709A US2010135510A1 US 20100135510 A1 US20100135510 A1 US 20100135510A1 US 62831709 A US62831709 A US 62831709A US 2010135510 A1 US2010135510 A1 US 2010135510A1
Authority
US
United States
Prior art keywords
sound source
signals
signal
audio
sound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/628,317
Other versions
US8351612B2 (en
Inventor
Jae-Hyoun Yoo
Hwan SHIM
Hyun-Joo Chung
Koen-Mo SUNG
Jeongil SEO
Kyeongok Kang
Jin-Woo Hong
Chieteuk Ahn
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Electronics and Telecommunications Research Institute ETRI
Original Assignee
Electronics and Telecommunications Research Institute ETRI
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020090020190A external-priority patent/KR20100062784A/en
Application filed by Electronics and Telecommunications Research Institute ETRI filed Critical Electronics and Telecommunications Research Institute ETRI
Assigned to ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE reassignment ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTITUTE ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AHN, CHIETEUK, HONG, JIN-WOO, KANG, KYEONGOK, CHUNG, HYUN-JOO, SEO, JEONGIL, SHIM, HWAN, SUNG, KOEN-MO, YOO, JAE-HYOUN
Publication of US20100135510A1 publication Critical patent/US20100135510A1/en
Priority to US13/709,475 priority Critical patent/US20130101122A1/en
Application granted granted Critical
Publication of US8351612B2 publication Critical patent/US8351612B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/308Electronic adaptation dependent on speaker or headphone connection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/11Positioning of individual sound objects, e.g. moving airplane, within a sound field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/15Aspects of sound capture and related signal processing for recording or reproduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/03Application of parametric coding in stereophonic audio systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/13Application of wave-field synthesis in stereophonic audio systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/301Automatic calibration of stereophonic sound system, e.g. with test microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space

Definitions

  • Example embodiments relate to an object based audio contents generating/playing apparatus, and more particularly, to an object based audio contents generating/playing apparatus that may generate/play object based audio contents regardless of a user environment of the object based audio contents.
  • MPEG-4 is an audio/video encoding standard proposed by a moving picture expert group (MPEG), the affiliated organization of an international organization for standardization/international electrotechnical commission (ISO/IEC), in 1998.
  • MPEG-4 is developed from a standard system of MPEG-1 and MPEG-2 and additionally includes a virtual reality markup language (VRML) and contents relating to an object-oriented composite file, and the like.
  • VRML virtual reality markup language
  • MPEG-4 aims at increasing an encoding rate, developing an integrated method of encoding an audio, a video, and a voice, enabling interactive audio/video to be played, and developing an error restoring technique.
  • MPEG-4 has a main feature of playing an object based audio/video. That is, MPEG-1 and MPEG-2 is limited to a general structure, a multi-transmission, and synchronization, whereas MPEG-4 additionally includes a scene description, interactivity, contents description, and a possibility of programming.
  • MPEG-4 classifies a target for encoding for each object, sets an encoding method according to an attribution of each object, describes a desired scene, and transmits the described scene in an audio binary format for scenes (AudioBIFS). Also, audiences may control information such as size of each object, a location of each object, and the like, through a terminal, when listening to the audio.
  • WFS wave field synthesis
  • a standardization project relating to the WFS scheme namely, a creating assessing and rendering in real time of high quality audio-visual environments in MPEG-4 context (CARROUSO)
  • CARROUSO a creating assessing and rendering in real time of high quality audio-visual environments in MPEG-4 context
  • Example embodiments may provide an object based audio contents generating/playing apparatus that enables the object based audio contents to be played using at least one of a wave field synthesis (WFS) scheme and a multi-channel surround scheme regardless of a reproducing environment of the audience.
  • WFS wave field synthesis
  • an apparatus of generating an object based audio contents including an object audio signal obtaining unit to obtain a plurality of object audio signals by recording a plurality of sound source signals, a recording space information obtaining unit to obtain recording space information with respect to a recording space of the plurality of sound source signals, a sound source location information obtaining unit to obtain sound location information of the plurality of sound source signals, and an encoding unit to generate object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, and the sound source location information.
  • an apparatus of reproducing object based audio contents including a decoding unit to decode a plurality of object audio signals of a plurality of sound source signals and sound source location information of the plurality of sound source signals, from the object based audio contents, a reproducing space (area) information obtaining unit to obtain reproducing space information with respect to a reproducing space of the plurality of object based audio contents, a signal synthesizing unit to synthesize a plurality of speaker signals from the decoded plurality of object audio signals based on the sound source location information and the reproducing space information, and a transmitting unit to transmit the plurality of speaker signals to a plurality of speakers respectively corresponding to the plurality of speaker signals.
  • FIG. 1 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to example embodiments
  • FIG. 2 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to other example embodiments
  • FIG. 3 is a block diagram illustrating a detailed configuration of an object based audio contents playing apparatus according to example embodiments
  • FIG. 4 is a flowchart illustrating an object based audio contents generating method according to example embodiments.
  • FIG. 5 is a flowchart illustrating an object based audio contents playing method according to example embodiments.
  • FIG. 1 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to example embodiments.
  • the object based audio contents generating apparatus 100 may include an object audio signal obtaining unit 110 , a sound source location information obtaining unit 120 , a recording space information obtaining unit 130 , and an encoding unit 140 . Also, according to example embodiments, the object based audio contents generating apparatus 100 may further include a room impulse signal emitting unit 160 and a room impulse signal receiving unit 150 .
  • a function of each element will be described in detail.
  • the object audio signal obtaining unit 110 obtains a plurality of object audio signals by recording a plurality of sound source signals.
  • a number of the plurality of sound source signals is identical to a number of object audio signals. That is, the object audio signal obtaining unit 110 may obtain a single object audio signal for a single sound source signal.
  • the object audio signal obtaining unit 110 may obtain the plurality of object audio signals using at least one of a plurality of spot microphones and a microphone array.
  • Each of the plurality of spot microphones is installed adjacent to each of plurality of sound sources, thereby obtaining an object audio signal by recording a sound source signal from each of the plurality of sound sources.
  • the microphone array is an arrangement of the plurality of microphones.
  • a plurality of object audio signals may be obtained for each sound source by classifying the plurality of sound source signals using a delay time and a sound pressure level (SPL) of a plurality of sound source signals that arrive at the microphone array.
  • SPL sound pressure level
  • the delay time of the plurality of sound source signals may include at least one of a delay time between a plurality of sound sources that arrive at a single microphone from among the plurality of microphones constituting the microphone array, and a delay time of a sound source signal that arrives at each of the plurality of microphones, when a single sound source signal arrives at each of the plurality of microphones.
  • the sound source location information obtaining unit 120 obtains sound source location information of the plurality of sound source signals.
  • the sound source location information includes information with respect to a space where a plurality of sound signals to be recorded are to be played. That is, the sound source location information may include sound image location information.
  • the sound location information namely, sound image location information, may be expressed as orthogonal coordinates, such as (x, y, z), or cylinder coordinates, such as (r, ⁇ , ⁇ ) for each of the plurality of sound source signals.
  • the sound source location information obtaining unit 120 may obtain the sound source location information using at least one of a location of the plurality of spot microphones, the delay time of the plurality of sound source signals in the microphone array, and the SPL of the plurality of sound source signals in the microphone array.
  • the sound source location information obtaining unit 120 may obtain the sound source location information by receiving a location of the plurality of sound sources inputted by a user of the object based audio contents generating apparatus 100 .
  • the recording space information obtaining unit 130 obtains recording space information with respect to a recording space of the plurality of sound source signals.
  • the recording space information is information with respect to a space where the plurality of sound sources to be recorded are to be played.
  • the object based audio contents generating apparatus 100 may further include the room impulse signal emitting unit 160 and the room impulse signal receiving unit 150 .
  • the room impulse signal emitting unit 160 emits an impulse sound source signal.
  • the impulse sound source signal is a signal used for calculating an impulse response which will be described below.
  • the room impulse signal emitting unit 160 may emit a maximum-length sequence (MLS) signal.
  • MLS maximum-length sequence
  • the room impulse signal receiving unit 150 receives the impulse sound source signal emitted from the room impulse signal emitting unit 160 , and calculates the impulse response based on the received impulse sound source signal.
  • the impulse sound source signal received in the room impulse signal receiving unit 150 includes a sound signal that directly arrives at the room impulse signal receiving unit 150 from the sound source signal emitting unit 150 and all sound signals arrive at the room impulse signal receiving unit 150 by being reflected from a surface of a wall of the recording space, an object existing in the recording space, and the like after being emitted from the room impulse signal emitting unit 160 .
  • the recording space information obtaining unit 130 may obtain the recording space information based on the calculated impulse response, and according to example embodiments, the impulse response may include a plurality of impulse signals, and the recording space information may include at least one of a incoming time difference between the plurality of impulse signals, an SPL difference between the plurality of impulse signals, a incoming azimuth difference between the plurality of signals. That is, the recording space information obtaining unit 130 may obtain the impulse response with respect to the recording space in a form of data, as well as in a form of an audio format, such as a wave file.
  • the recording space information may be expressed as an ordered pair of a time, a sound pressure, and an angle, when the recording space information includes all of the incoming time difference, the SLP difference, and the incoming azimuth difference described above.
  • the encoding unit 140 generates object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, and sound source location information.
  • each of the plurality of object audio signals may be encoded through various schemes.
  • the encoding unit 140 may encode the object audio signal by applying an audio encoding scheme optimal to the music signal, such as a transform based audio encoding scheme, and when the object audio signal is a speech signal, the encoding unit 140 may encode the object audio signal by applying an audio encoding scheme optimal to the speech signal, such as a code excited linear prediction (CELP) structural audio encoding scheme.
  • CELP code excited linear prediction
  • the encoding unit 140 may generate the object based audio contents by multiplexing an encoded object audio signal, encoded sound source location information, and encoded recording space information.
  • the object based audio contents generated in the encoding unit 140 may be transmitted via a network or may be stored in a separate recording media.
  • the object based audio contents generating apparatus 100 encodes each of the plurality of object audio signals, as opposed to mixing the plurality of the object audio signals to encode in a form of a multi-channel audio signal, generates the object based audio contents by adding additional information, such as the sound source location information, recording space information, and the like, to the encoded object audio signal, thereby enabling the user of an object based audio contents playing apparatus to generate object based audio contents appropriate for its object based audio contents playing apparatus.
  • the object based audio content playing apparatus will be described with reference to FIG. 3 .
  • FIG. 2 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to other example embodiments.
  • the object based audio contents generating apparatus 200 includes an object audio signal obtaining unit 210 , a sound source location information obtaining unit 220 , a recording space information obtaining unit 230 , a multi-channel audio mixing unit 240 , and an encoding unit 250 .
  • the object audio signal obtaining unit 210 , the sound source location information obtaining unit 220 , the recording space information obtaining unit 230 , and the encoding unit 250 of FIG. 2 respectively correspond to the object audio signal obtaining unit 110 , the sound source location information obtaining unit 120 , the recording space information obtaining unit 130 , and the encoding unit 140 of FIG. 1 . Accordingly, description of the object based audio contents generating apparatus 100 of FIG. 1 is applicable to the object based audio contents generating apparatus 200 of FIG. 2 , although the description is omitted hereinafter.
  • the object audio signal obtaining unit 210 obtains a plurality of object audio signals by recording a plurality of sound source signals.
  • the sound source location obtaining unit 220 obtains sound source location information of the plurality of sound source signals.
  • the recording space information obtaining unit 230 obtains recording space information with respect to a recording space of the plurality of sound source signals.
  • the multi-channel audio mixing unit 240 generates a multi-channel audio signal by mixing at least one of the plurality of object audio signals, the recording space information, and the sound source information.
  • the multi-channel audio mixing unit 240 may generate the multi-channel audio signal, such as a 2 channel audio signal, a 5.1 channel audio signal, a 7.1 channel audio signal, and the like, by mixing at least one object audio signal, the sound source location information, and recording space information, for backwards compatibility with an audio contents playing apparatus according to a multi-channel surround playing scheme.
  • the multi-channel audio signal such as a 2 channel audio signal, a 5.1 channel audio signal, a 7.1 channel audio signal, and the like, by mixing at least one object audio signal, the sound source location information, and recording space information, for backwards compatibility with an audio contents playing apparatus according to a multi-channel surround playing scheme.
  • the encoding unit 250 generates the object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, the sound source location information, and the multi-channel audio signal.
  • FIG. 3 is a block diagram illustrating a detailed configuration of an object based audio contents playing apparatus according to example embodiments.
  • the object based audio contents playing apparatus 300 includes an encoding unit 310 , a reproducing space information obtaining unit 320 , a signal synthesizing unit 330 , and a transmission unit 340 .
  • an encoding unit 310 includes an encoding unit 310 , a reproducing space information obtaining unit 320 , a signal synthesizing unit 330 , and a transmission unit 340 .
  • a function of each element will be described.
  • the encoding unit 310 decodes a plurality of object audio signals with respect to a plurality of sound source signals and sound source location information of the plurality of sound source signals, from the object based audio contents.
  • the object based audio contents may be transmitted from an object based audio contents generating apparatus or may be read from a separate recording medium.
  • the decoding unit 310 may generate a plurality of encoded object audio signals and encoded sound source location information by demultiplexing the object based audio contents, and may restore the plurality of object audio signals, recording space information, and sound source location information from the generated encoded plurality of object audio signals and the generated encoded sound source information.
  • the reproducing space information obtaining unit 320 obtains reproducing space information with respect to a reproducing space of the plurality of object audio signals.
  • the reproducing space information is information with respect to a reproducing space of a user where the object based audio contents is to be played, and a plurality of speakers that plays the object based audio contents may be arranged in the reproducing space.
  • the reproducing space information may include at least one of a number of the plurality of speakers arranged in the reproducing space, an interval between the plurality of speakers, an arrangement angle of the plurality of speakers, a type of speakers, location information of speakers, and size information of the reproducing space.
  • the reproducing space information obtaining unit 320 may receive the reproducing space information directly inputted from the user, and may calculate the reproducing space information using a separate microphone arranged in the reproducing space.
  • the signal synthesizing unit 330 synthesizes a plurality of speaker signals from a decoded object audio signal from among the plurality of decoded object audio signals based on the sound source location information and the reproducing space information.
  • the signal synthesizing unit 330 synthesizes the plurality of speaker signals to effectively play the object based audio contents, based on the object audio signal, the sound source location information, and the reproducing space information.
  • the plurality of speaker signals are generated by synthesizing the plurality of object audio signals according to recording space information.
  • the signal synthesizing unit 330 when the object audio signal capable of being played in a WFS scheme based on the size of the reproducing space, the number of speakers installed in the reproducing space, the type of speakers, and the location of speakers, the signal synthesizing unit 330 performs rendering of an object audio signal according to the WFS scheme, and when the object audio signal is not capable of being played in the WFS scheme based on the size of the reproducing space, the number of speakers installed in the reproducing space, the type of speakers, and the location of speakers, the signal synthesizing unit 330 synthesizes a speaker signal by rendering the object audio signal according to a multi-channel surround play scheme.
  • the signal synthesizing unit 330 may select a desired speaker to play the object audio signal.
  • the signal synthesizing unit 330 performs rendering of an object audio signal with respect to the corresponding audio object using the sound length synthesis scheme, and when the audio object exists in other angles, the signal synthesizing unit 330 performs rendering of an audio object signal with respect to the audio object existing in other angles by applying a power panning law using a satellite surround loudspeaker.
  • the transmission unit 340 respectively transmits the plurality of speaker signals to corresponding speakers.
  • a transmitted speaker signal is played via a corresponding speaker.
  • the encoding unit 310 further decodes a plurality of sound source recording space information from the object based audio contents, and the signal synthesizing unit 330 generates a direct sound with respect to the plurality of sound source signals from the object audio signal using the object audio signal, sound source information, and reproducing space information, and synthesizes the plurality of speaker signals by adding a reflected sound to the generated direct sound based on the recording space information.
  • the signal synthesizing unit 330 may generate the direct sound with respect to the plurality of sound source signals by rendering the plurality of object audio signals based on Equation 1 or Equation 2 as given below.
  • Q( ⁇ right arrow over (r) ⁇ n , ⁇ ) is a driving function of an audio signal emitted from an n th loudspeaker of the loudspeaker array
  • Q′( ⁇ right arrow over (r) ⁇ n , ⁇ ) is a driving function of an audio signal emitted from an n th loudspeaker of a tilted loudspeaker array
  • S( ⁇ ) is a virtual sound source signal
  • G n ( ⁇ n , ⁇ ) is a factor to weight a sound pressure by directional characteristics of the loudspeaker
  • Z is coordinate information of the loudspeaker
  • Z 0 is coordinate information of the sound source
  • Z 1 is coordinate information of a virtual sound source
  • k is a wave number
  • is a angle velocity
  • ⁇ n is an angle between the n th loudspeaker and the audience
  • ⁇ right arrow over (r) ⁇ n is a distance between the sound source and the audience
  • is a delivery time occurring due to a distance between the virtual sound source and the n th loudspeaker
  • cos( ⁇ n ) is a distance ratio of a virtual sound source with respect to a vertical distance and the n th loudspeaker
  • the signal synthesizing unit 330 may operate, according to a grouped reflections algorithm, the direct sound generated according to Equation 1 and Equation 2 and the recording space information expressed as an ordered combination of time, sound pressure, and angle, and may add initial reflected sound information of the recording space to the directed sound.
  • the signal synthesizing unit 330 assigns each reflected sound to the loudspeaker using angle information included in the reflected sound information, and when the loudspeaker does not exist in a corresponding angle, the signal synthesizing unit 330 synthesizes a speaker signal to enable the reflected sound to be played in a loudspeaker adjacent to the corresponding angle.
  • the signal synthesizing unit 330 may add a reverberation effect to the speaker signal using an infinite impulse response filter (IIR filter).
  • IIR filter infinite impulse response filter
  • the object audio signal may further include the multi-channel audio signal.
  • the audio signal to be played is a channel based signal and the reproducing space is set to be appropriate for the WFS scheme but the audience intends to play the audio signal according to a multi-channel surround scheme
  • the signal synthesizing unit 330 may select a loudspeaker and synthesizes a speaker signal to enable the object based audio contents to be played according to the multi-channel surround play scheme.
  • the signal synthesizing unit 330 selects a loudspeaker arranged at 0°, ⁇ 30°, and ⁇ 110° based on the front of the audience, and synthesizes the speaker signal to enable the object based audio contents to be played via the selected loudspeaker.
  • the signal synthesizing unit 330 enables the object based audio contents to be played according to the multi-channel surround scheme.
  • the object based audio contents play apparatus 300 may play the object based audio contents using at least one of the WFS scheme and the multi-channel surround scheme regardless of a reproducing environment of the audience.
  • FIG. 4 is a flowchart illustrating an object based audio contents generating method according to example embodiments. Hereinafter, a procedure performed in each operation will be described with reference to FIG. 4 .
  • a plurality of object audio signals are obtained by recording a plurality of sound source signals.
  • the plurality of object audio signals may be obtained using at least one of a plurality of spot microphones and a microphone array in operation 5410 .
  • the sound source location information may be obtained using at least one of a location of the plurality of spot microphones, a delay time of the plurality of sound source signals in the microphone array, an SPL of the plurality of sound source signals in the microphone array.
  • the sound source location information may be obtained by receiving a location of the plurality of sound sources inputted by a user.
  • the object based audio contents generating method may further include an operation (not illustrated) of emitting an impulse sound source signal and receiving the emitted impulse sound source signal, and an operation (not illustrated) of calculating an impulse response based on the received impulse sound source signal.
  • the recording space information may be obtained based on the calculated impulse response in operation S 430 .
  • the impulse response includes a plurality of impulse signals
  • the recording space information includes at least one of a incoming time difference between the plurality of impulse signals, an SPL difference between the plurality of impulse signals, and a incoming azimuth difference between the plurality of impulse signals.
  • object based audio contents are generated by encoding at least one of the plurality of object audio signals, the recording space information, and the sound source location information.
  • the object based audio contents generating method may further include an operation of generating a multi-channel audio signal by mixing at least one of the plurality of object audio signals, the recording space information, and the sound source location information.
  • the object based audio contents may be generated by encoding at least one of the plurality of object audio signals, the recording space information, the sound source location information, and the multi-channel audio signal in operation S 440 .
  • FIG. 5 is a flowchart illustrating an object based audio contents playing method according to example embodiments. Hereinafter, a procedure performed in each operation will be described with reference to FIG. 5 .
  • a plurality of object audio signals with respect to a plurality of sound sources and sound source location information with respect to a plurality of sound source signals are decoded from the object based audio contents.
  • the reproducing space information may include at least one of a number of a plurality of speakers arranged in the reproducing space, an interval between the plurality of speakers, an arrangement angle of the plurality of speakers, a type of speakers, location information of the speakers, and size information of the reproducing space.
  • the reproducing space information may be directly received from the user or may be calculated using a separate microphone arranged in the reproducing space in operation S 520 .
  • a plurality of speaker signals is synthesized from decoded object audio signal based on the sound source location information and reproducing space information.
  • a reverberation effect may be added to the plurality of speaker signals using an IIR filter in operation 5530 .
  • the plurality of speaker signals are respectively transmitted to corresponding speakers.
  • a transmitted speaker signal may be played via a corresponding speaker.
  • FIG. 1 through FIG. 3 A few example embodiments of the object based audio contents generating/playing method have been shown and described, and the object based audio contents generating/playing apparatus described in FIG. 1 through FIG. 3 is applicable to the present example embodiment. Accordingly, detailed descriptions thereof will be omitted.
  • the object based audio contents generating/playing method may be recorded in computer-readable media including program instructions to implement various operations embodied by a computer.
  • the media may also include, alone or in combination with the program instructions, data files, data structures, and the like.
  • Examples of computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD ROM disks and DVDs; magneto-optical media such as optical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory, and the like.
  • Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter.
  • the described hardware devices may be configured to act as one or more software modules in order to perform the operations of the above-described example embodiments, or vice versa.

Abstract

Disclosed is an object based audio contents generating/playing apparatus. The object based audio contents generating/playing apparatus may include an object audio signal obtaining unit to obtain a plurality of object audio signals by recording a plurality of sound source signals, a recording space information obtaining unit to obtain recording space information with respect to a recording space of the plurality of sound source signals, a sound source location information obtaining unit to obtain sound location information of the plurality of sound source signals, and an encoding unit to generate object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, and the sound source location information, thereby enabling the object based audio contents to be played using at least one of a WFS scheme and a multi-channel surround scheme regardless of a reproducing environment of the audience.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims the benefit of Korean Patent Application No. 10-2008-0121112, filed on Dec. 2, 2008, and Korean Patent Application No. 10-2009-0020190, filed on Mar. 10, 2009, in the Korean Intellectual Property Office, the disclosures of which are incorporated herein by reference.
  • BACKGROUND
  • 1. Field
  • Example embodiments relate to an object based audio contents generating/playing apparatus, and more particularly, to an object based audio contents generating/playing apparatus that may generate/play object based audio contents regardless of a user environment of the object based audio contents.
  • 2. Description of the Related Art
  • MPEG-4 is an audio/video encoding standard proposed by a moving picture expert group (MPEG), the affiliated organization of an international organization for standardization/international electrotechnical commission (ISO/IEC), in 1998. MPEG-4 is developed from a standard system of MPEG-1 and MPEG-2 and additionally includes a virtual reality markup language (VRML) and contents relating to an object-oriented composite file, and the like. MPEG-4 aims at increasing an encoding rate, developing an integrated method of encoding an audio, a video, and a voice, enabling interactive audio/video to be played, and developing an error restoring technique.
  • MPEG-4 has a main feature of playing an object based audio/video. That is, MPEG-1 and MPEG-2 is limited to a general structure, a multi-transmission, and synchronization, whereas MPEG-4 additionally includes a scene description, interactivity, contents description, and a possibility of programming. MPEG-4 classifies a target for encoding for each object, sets an encoding method according to an attribution of each object, describes a desired scene, and transmits the described scene in an audio binary format for scenes (AudioBIFS). Also, audiences may control information such as size of each object, a location of each object, and the like, through a terminal, when listening to the audio.
  • As a representative object based audio contents playing method, there is wave field synthesis (WFS) scheme. The WFS scheme generates a wavefront identical to a first wavefront in a space classified as a loudspeaker array by synthesizing sounds played through a plurality of loudspeakers from the first wavefront generated from a first sound source.
  • A standardization project relating to the WFS scheme, namely, a creating assessing and rendering in real time of high quality audio-visual environments in MPEG-4 context (CARROUSO), has conducted research to transmit a sound source in a form of an object through MPEG-4 having a feature of object-oriented and commutativity, and to play using the WFS scheme.
  • SUMMARY
  • Example embodiments may provide an object based audio contents generating/playing apparatus that enables the object based audio contents to be played using at least one of a wave field synthesis (WFS) scheme and a multi-channel surround scheme regardless of a reproducing environment of the audience.
  • According to example embodiments, there may be provided an apparatus of generating an object based audio contents, the apparatus including an object audio signal obtaining unit to obtain a plurality of object audio signals by recording a plurality of sound source signals, a recording space information obtaining unit to obtain recording space information with respect to a recording space of the plurality of sound source signals, a sound source location information obtaining unit to obtain sound location information of the plurality of sound source signals, and an encoding unit to generate object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, and the sound source location information.
  • According to example embodiments, there may be provided an apparatus of reproducing object based audio contents, the apparatus including a decoding unit to decode a plurality of object audio signals of a plurality of sound source signals and sound source location information of the plurality of sound source signals, from the object based audio contents, a reproducing space (area) information obtaining unit to obtain reproducing space information with respect to a reproducing space of the plurality of object based audio contents, a signal synthesizing unit to synthesize a plurality of speaker signals from the decoded plurality of object audio signals based on the sound source location information and the reproducing space information, and a transmitting unit to transmit the plurality of speaker signals to a plurality of speakers respectively corresponding to the plurality of speaker signals.
  • Additional aspects and/or advantages will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the embodiments.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and/or other aspects and advantages will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
  • FIG. 1 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to example embodiments;
  • FIG. 2 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to other example embodiments;
  • FIG. 3 is a block diagram illustrating a detailed configuration of an object based audio contents playing apparatus according to example embodiments;
  • FIG. 4 is a flowchart illustrating an object based audio contents generating method according to example embodiments; and
  • FIG. 5 is a flowchart illustrating an object based audio contents playing method according to example embodiments.
  • DETAILED DESCRIPTION
  • Reference will now be made in detail to example embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. Example embodiments are described below to explain the present disclosure by referring to the figures.
  • FIG. 1 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to example embodiments.
  • According to example embodiments, the object based audio contents generating apparatus 100 may include an object audio signal obtaining unit 110, a sound source location information obtaining unit 120, a recording space information obtaining unit 130, and an encoding unit 140. Also, according to example embodiments, the object based audio contents generating apparatus 100 may further include a room impulse signal emitting unit 160 and a room impulse signal receiving unit 150. Hereinafter, a function of each element will be described in detail.
  • The object audio signal obtaining unit 110 obtains a plurality of object audio signals by recording a plurality of sound source signals.
  • In this instance, a number of the plurality of sound source signals is identical to a number of object audio signals. That is, the object audio signal obtaining unit 110 may obtain a single object audio signal for a single sound source signal.
  • According to example embodiments, the object audio signal obtaining unit 110 may obtain the plurality of object audio signals using at least one of a plurality of spot microphones and a microphone array.
  • Each of the plurality of spot microphones is installed adjacent to each of plurality of sound sources, thereby obtaining an object audio signal by recording a sound source signal from each of the plurality of sound sources.
  • The microphone array is an arrangement of the plurality of microphones. When the microphone array is used, a plurality of object audio signals may be obtained for each sound source by classifying the plurality of sound source signals using a delay time and a sound pressure level (SPL) of a plurality of sound source signals that arrive at the microphone array.
  • Here, the delay time of the plurality of sound source signals may include at least one of a delay time between a plurality of sound sources that arrive at a single microphone from among the plurality of microphones constituting the microphone array, and a delay time of a sound source signal that arrives at each of the plurality of microphones, when a single sound source signal arrives at each of the plurality of microphones.
  • The sound source location information obtaining unit 120 obtains sound source location information of the plurality of sound source signals.
  • Here, the sound source location information includes information with respect to a space where a plurality of sound signals to be recorded are to be played. That is, the sound source location information may include sound image location information. The sound location information, namely, sound image location information, may be expressed as orthogonal coordinates, such as (x, y, z), or cylinder coordinates, such as (r, θ, φ) for each of the plurality of sound source signals.
  • According to example embodiments, the sound source location information obtaining unit 120 may obtain the sound source location information using at least one of a location of the plurality of spot microphones, the delay time of the plurality of sound source signals in the microphone array, and the SPL of the plurality of sound source signals in the microphone array.
  • Also, according to other example embodiments, the sound source location information obtaining unit 120 may obtain the sound source location information by receiving a location of the plurality of sound sources inputted by a user of the object based audio contents generating apparatus 100.
  • The recording space information obtaining unit 130 obtains recording space information with respect to a recording space of the plurality of sound source signals.
  • Here, the recording space information is information with respect to a space where the plurality of sound sources to be recorded are to be played.
  • As described above, according to example embodiments, the object based audio contents generating apparatus 100 may further include the room impulse signal emitting unit 160 and the room impulse signal receiving unit 150.
  • The room impulse signal emitting unit 160 emits an impulse sound source signal.
  • The impulse sound source signal is a signal used for calculating an impulse response which will be described below.
  • As an example, the room impulse signal emitting unit 160 may emit a maximum-length sequence (MLS) signal.
  • The room impulse signal receiving unit 150 receives the impulse sound source signal emitted from the room impulse signal emitting unit 160, and calculates the impulse response based on the received impulse sound source signal.
  • The impulse sound source signal received in the room impulse signal receiving unit 150 includes a sound signal that directly arrives at the room impulse signal receiving unit 150 from the sound source signal emitting unit 150 and all sound signals arrive at the room impulse signal receiving unit 150 by being reflected from a surface of a wall of the recording space, an object existing in the recording space, and the like after being emitted from the room impulse signal emitting unit 160.
  • In this instance, the recording space information obtaining unit 130 may obtain the recording space information based on the calculated impulse response, and according to example embodiments, the impulse response may include a plurality of impulse signals, and the recording space information may include at least one of a incoming time difference between the plurality of impulse signals, an SPL difference between the plurality of impulse signals, a incoming azimuth difference between the plurality of signals. That is, the recording space information obtaining unit 130 may obtain the impulse response with respect to the recording space in a form of data, as well as in a form of an audio format, such as a wave file. The recording space information may be expressed as an ordered pair of a time, a sound pressure, and an angle, when the recording space information includes all of the incoming time difference, the SLP difference, and the incoming azimuth difference described above.
  • The encoding unit 140 generates object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, and sound source location information.
  • In this instance, each of the plurality of object audio signals may be encoded through various schemes. As an example, when an object audio signal is a music signal, the encoding unit 140 may encode the object audio signal by applying an audio encoding scheme optimal to the music signal, such as a transform based audio encoding scheme, and when the object audio signal is a speech signal, the encoding unit 140 may encode the object audio signal by applying an audio encoding scheme optimal to the speech signal, such as a code excited linear prediction (CELP) structural audio encoding scheme.
  • In this instance, the encoding unit 140 may generate the object based audio contents by multiplexing an encoded object audio signal, encoded sound source location information, and encoded recording space information.
  • The object based audio contents generated in the encoding unit 140 may be transmitted via a network or may be stored in a separate recording media.
  • As described above, the object based audio contents generating apparatus 100 according to example embodiments encodes each of the plurality of object audio signals, as opposed to mixing the plurality of the object audio signals to encode in a form of a multi-channel audio signal, generates the object based audio contents by adding additional information, such as the sound source location information, recording space information, and the like, to the encoded object audio signal, thereby enabling the user of an object based audio contents playing apparatus to generate object based audio contents appropriate for its object based audio contents playing apparatus. The object based audio content playing apparatus will be described with reference to FIG. 3.
  • FIG. 2 is a block diagram illustrating a detailed configuration of an object based audio contents generating apparatus according to other example embodiments.
  • According to other example embodiments, the object based audio contents generating apparatus 200 includes an object audio signal obtaining unit 210, a sound source location information obtaining unit 220, a recording space information obtaining unit 230, a multi-channel audio mixing unit 240, and an encoding unit 250.
  • The object audio signal obtaining unit 210, the sound source location information obtaining unit 220, the recording space information obtaining unit 230, and the encoding unit 250 of FIG. 2 respectively correspond to the object audio signal obtaining unit 110, the sound source location information obtaining unit 120, the recording space information obtaining unit 130, and the encoding unit 140 of FIG. 1. Accordingly, description of the object based audio contents generating apparatus 100 of FIG. 1 is applicable to the object based audio contents generating apparatus 200 of FIG. 2, although the description is omitted hereinafter.
  • The object audio signal obtaining unit 210 obtains a plurality of object audio signals by recording a plurality of sound source signals.
  • The sound source location obtaining unit 220 obtains sound source location information of the plurality of sound source signals.
  • The recording space information obtaining unit 230 obtains recording space information with respect to a recording space of the plurality of sound source signals.
  • The multi-channel audio mixing unit 240 generates a multi-channel audio signal by mixing at least one of the plurality of object audio signals, the recording space information, and the sound source information.
  • That is, the multi-channel audio mixing unit 240 may generate the multi-channel audio signal, such as a 2 channel audio signal, a 5.1 channel audio signal, a 7.1 channel audio signal, and the like, by mixing at least one object audio signal, the sound source location information, and recording space information, for backwards compatibility with an audio contents playing apparatus according to a multi-channel surround playing scheme.
  • The encoding unit 250 generates the object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, the sound source location information, and the multi-channel audio signal.
  • FIG. 3 is a block diagram illustrating a detailed configuration of an object based audio contents playing apparatus according to example embodiments.
  • The object based audio contents playing apparatus 300 according to example embodiments includes an encoding unit 310, a reproducing space information obtaining unit 320, a signal synthesizing unit 330, and a transmission unit 340. Hereinafter, a function of each element will be described.
  • The encoding unit 310 decodes a plurality of object audio signals with respect to a plurality of sound source signals and sound source location information of the plurality of sound source signals, from the object based audio contents.
  • The object based audio contents may be transmitted from an object based audio contents generating apparatus or may be read from a separate recording medium.
  • The decoding unit 310 may generate a plurality of encoded object audio signals and encoded sound source location information by demultiplexing the object based audio contents, and may restore the plurality of object audio signals, recording space information, and sound source location information from the generated encoded plurality of object audio signals and the generated encoded sound source information.
  • The reproducing space information obtaining unit 320 obtains reproducing space information with respect to a reproducing space of the plurality of object audio signals.
  • The reproducing space information is information with respect to a reproducing space of a user where the object based audio contents is to be played, and a plurality of speakers that plays the object based audio contents may be arranged in the reproducing space.
  • Accordingly, according to example embodiments, the reproducing space information may include at least one of a number of the plurality of speakers arranged in the reproducing space, an interval between the plurality of speakers, an arrangement angle of the plurality of speakers, a type of speakers, location information of speakers, and size information of the reproducing space.
  • Also, according to example embodiments, the reproducing space information obtaining unit 320 may receive the reproducing space information directly inputted from the user, and may calculate the reproducing space information using a separate microphone arranged in the reproducing space.
  • The signal synthesizing unit 330 synthesizes a plurality of speaker signals from a decoded object audio signal from among the plurality of decoded object audio signals based on the sound source location information and the reproducing space information.
  • That is, the signal synthesizing unit 330 synthesizes the plurality of speaker signals to effectively play the object based audio contents, based on the object audio signal, the sound source location information, and the reproducing space information. In this instance, the plurality of speaker signals are generated by synthesizing the plurality of object audio signals according to recording space information.
  • According to example embodiments, when the object audio signal capable of being played in a WFS scheme based on the size of the reproducing space, the number of speakers installed in the reproducing space, the type of speakers, and the location of speakers, the signal synthesizing unit 330 performs rendering of an object audio signal according to the WFS scheme, and when the object audio signal is not capable of being played in the WFS scheme based on the size of the reproducing space, the number of speakers installed in the reproducing space, the type of speakers, and the location of speakers, the signal synthesizing unit 330 synthesizes a speaker signal by rendering the object audio signal according to a multi-channel surround play scheme. When the object audio signal is rendered in an environment where a speaker array is installed, according to the multi-channel surround play scheme, the signal synthesizing unit 330 may select a desired speaker to play the object audio signal.
  • As an example, in a case that a loudspeaker array is arranged in front of the reproducing space based on an audience, and a 2 channel surround speaker is installed behind the reproducing space, when the audio object, that is, the sound source, exists in an angle between both ends of the loudspeaker array based on the audience, the signal synthesizing unit 330 performs rendering of an object audio signal with respect to the corresponding audio object using the sound length synthesis scheme, and when the audio object exists in other angles, the signal synthesizing unit 330 performs rendering of an audio object signal with respect to the audio object existing in other angles by applying a power panning law using a satellite surround loudspeaker.
  • The transmission unit 340 respectively transmits the plurality of speaker signals to corresponding speakers. A transmitted speaker signal is played via a corresponding speaker.
  • According to example embodiments, the encoding unit 310 further decodes a plurality of sound source recording space information from the object based audio contents, and the signal synthesizing unit 330 generates a direct sound with respect to the plurality of sound source signals from the object audio signal using the object audio signal, sound source information, and reproducing space information, and synthesizes the plurality of speaker signals by adding a reflected sound to the generated direct sound based on the recording space information.
  • As an example, in a case that the loudspeaker array is arranged in front of the reproducing space and the plurality of object audio signals is intended to be played via the loudspeaker array using the WFS scheme, the signal synthesizing unit 330 may generate the direct sound with respect to the plurality of sound source signals by rendering the plurality of object audio signals based on Equation 1 or Equation 2 as given below.
  • Q ( r n , ω ) = S ( ω ) z - z 1 z - z 0 cos ( θ n ) G n ( θ n , ω ) jk 2 π - jk r n - r m r n - r m [ Equation 1 ] Q ( r n , ω ) = N n · S ( ω ) jk 2 π cos ( θ n ) G n ( θ n - α n , ω ) z - z 1 z - z 0 - jk r n - r m r n - r m [ Equation 2 ]
  • Here, Q({right arrow over (r)}n, ω) is a driving function of an audio signal emitted from an nth loudspeaker of the loudspeaker array, Q′({right arrow over (r)}n, ω) is a driving function of an audio signal emitted from an nth loudspeaker of a tilted loudspeaker array, S(ω) is a virtual sound source signal, Gnn, ω) is a factor to weight a sound pressure by directional characteristics of the loudspeaker, Z is coordinate information of the loudspeaker, Z0 is coordinate information of the sound source, Z1 is coordinate information of a virtual sound source, k is a wave number, ω is a angle velocity, θn is an angle between the nth loudspeaker and the audience, {right arrow over (r)}n is a distance between the sound source and the audience, {right arrow over (r)}m is a distance between the loudspeaker and the audience, Nn is a normalization parameter, and αn is an angle between the tilted loudspeaker and the audience.
  • Also, in Equation 1 and Equation 2,
  • z - z 1 z - z 0
  • is a weight with respect to a size of the virtual sound source signal,
  • jk 2 π
  • is a high frequency amplifying equalizing coefficient, e−jk|{right arrow over (r)} n −/{right arrow over (r)} m | is a delivery time occurring due to a distance between the virtual sound source and the nth loudspeaker, cos(θn) is a distance ratio of a virtual sound source with respect to a vertical distance and the nth loudspeaker, and
  • 1 r n - r m
  • is a single cylindrical wave.
  • Subsequently, the signal synthesizing unit 330 may operate, according to a grouped reflections algorithm, the direct sound generated according to Equation 1 and Equation 2 and the recording space information expressed as an ordered combination of time, sound pressure, and angle, and may add initial reflected sound information of the recording space to the directed sound. In this instance, the signal synthesizing unit 330 assigns each reflected sound to the loudspeaker using angle information included in the reflected sound information, and when the loudspeaker does not exist in a corresponding angle, the signal synthesizing unit 330 synthesizes a speaker signal to enable the reflected sound to be played in a loudspeaker adjacent to the corresponding angle.
  • Also, according to example embodiments, the signal synthesizing unit 330 may add a reverberation effect to the speaker signal using an infinite impulse response filter (IIR filter).
  • As described above with reference to FIG. 2, according to example embodiments, the object audio signal may further include the multi-channel audio signal. In a case that the audio signal to be played is a channel based signal and the reproducing space is set to be appropriate for the WFS scheme but the audience intends to play the audio signal according to a multi-channel surround scheme, the signal synthesizing unit 330 may select a loudspeaker and synthesizes a speaker signal to enable the object based audio contents to be played according to the multi-channel surround play scheme. As an example, in a case that the multi-channel audio signal is a 5.1 channel audio signal, the loudspeaker array is in front of the reproducing space, and 2 channel surround speaker is behind the reproducing space, the signal synthesizing unit 330 selects a loudspeaker arranged at 0°, ±30°, and ±110° based on the front of the audience, and synthesizes the speaker signal to enable the object based audio contents to be played via the selected loudspeaker.
  • Also, when the audio signal to be played is the multi-channel audio signal, and the reproducing space is set to be appropriate for the multi-channel surround scheme, the signal synthesizing unit 330 enables the object based audio contents to be played according to the multi-channel surround scheme.
  • As described above, the object based audio contents play apparatus 300 according to example embodiments may play the object based audio contents using at least one of the WFS scheme and the multi-channel surround scheme regardless of a reproducing environment of the audience.
  • FIG. 4 is a flowchart illustrating an object based audio contents generating method according to example embodiments. Hereinafter, a procedure performed in each operation will be described with reference to FIG. 4.
  • In operation S410, a plurality of object audio signals are obtained by recording a plurality of sound source signals.
  • According to example embodiments, the plurality of object audio signals may be obtained using at least one of a plurality of spot microphones and a microphone array in operation 5410.
  • In operation S420, sound source location information of the plurality of sound source signals is obtained.
  • According to example embodiments, the sound source location information may be obtained using at least one of a location of the plurality of spot microphones, a delay time of the plurality of sound source signals in the microphone array, an SPL of the plurality of sound source signals in the microphone array.
  • Also, according to other example embodiments, in operation S420, the sound source location information may be obtained by receiving a location of the plurality of sound sources inputted by a user.
  • In operation S430, recording space information with respect to the plurality of sound source signals is obtained.
  • According to example embodiments, the object based audio contents generating method may further include an operation (not illustrated) of emitting an impulse sound source signal and receiving the emitted impulse sound source signal, and an operation (not illustrated) of calculating an impulse response based on the received impulse sound source signal. In this instance, the recording space information may be obtained based on the calculated impulse response in operation S430. Also, in this instance, according to example embodiments, the impulse response includes a plurality of impulse signals, and the recording space information includes at least one of a incoming time difference between the plurality of impulse signals, an SPL difference between the plurality of impulse signals, and a incoming azimuth difference between the plurality of impulse signals.
  • In operation 5440, object based audio contents are generated by encoding at least one of the plurality of object audio signals, the recording space information, and the sound source location information.
  • Also, according to example embodiments, the object based audio contents generating method may further include an operation of generating a multi-channel audio signal by mixing at least one of the plurality of object audio signals, the recording space information, and the sound source location information. In this instance, the object based audio contents may be generated by encoding at least one of the plurality of object audio signals, the recording space information, the sound source location information, and the multi-channel audio signal in operation S440.
  • FIG. 5 is a flowchart illustrating an object based audio contents playing method according to example embodiments. Hereinafter, a procedure performed in each operation will be described with reference to FIG. 5.
  • In operation S510, a plurality of object audio signals with respect to a plurality of sound sources and sound source location information with respect to a plurality of sound source signals are decoded from the object based audio contents.
  • In operation S520, reproducing space information with respect to a reproducing space of the plurality of object audio signals is obtained.
  • According to example embodiments, the reproducing space information may include at least one of a number of a plurality of speakers arranged in the reproducing space, an interval between the plurality of speakers, an arrangement angle of the plurality of speakers, a type of speakers, location information of the speakers, and size information of the reproducing space.
  • Also, according to example embodiments, the reproducing space information may be directly received from the user or may be calculated using a separate microphone arranged in the reproducing space in operation S520.
  • In operation S530, a plurality of speaker signals is synthesized from decoded object audio signal based on the sound source location information and reproducing space information.
  • According to example embodiments, a reverberation effect may be added to the plurality of speaker signals using an IIR filter in operation 5530.
  • In operation S540, the plurality of speaker signals are respectively transmitted to corresponding speakers. A transmitted speaker signal may be played via a corresponding speaker.
  • A few example embodiments of the object based audio contents generating/playing method have been shown and described, and the object based audio contents generating/playing apparatus described in FIG. 1 through FIG. 3 is applicable to the present example embodiment. Accordingly, detailed descriptions thereof will be omitted.
  • The object based audio contents generating/playing method according to the above-described example embodiments may be recorded in computer-readable media including program instructions to implement various operations embodied by a computer. The media may also include, alone or in combination with the program instructions, data files, data structures, and the like. Examples of computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD ROM disks and DVDs; magneto-optical media such as optical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory, and the like. Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter. The described hardware devices may be configured to act as one or more software modules in order to perform the operations of the above-described example embodiments, or vice versa.
  • Although a few example embodiments have been shown and described, it would be appreciated by those skilled in the art that changes may be made in these example embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.

Claims (10)

1. An apparatus of generating an object based audio contents, the apparatus comprising:
an object audio signal obtaining unit to obtain a plurality of object audio signals by recording a plurality of sound source signals;
a recording space information obtaining unit to obtain recording space information with respect to a recording space of the plurality of sound source signals;
a sound source location information obtaining unit to obtain sound location information of the plurality of sound source signals; and
an encoding unit to generate object based audio contents by encoding at least one of the plurality of object audio signals, the recording space information, and the sound source location information.
2. The apparatus of claim 1, wherein the object audio signal obtaining unit obtains the plurality of object audio signals using at least one of a plurality of spot microphones and a microphone array.
3. The apparatus of claim 2, wherein the sound source location information obtaining unit obtains the sound source location information using at least one of locations of the plurality of spot microphones, a delay time of the plurality of sound source signals in the microphone array, a sound pressure level of the plurality of sound source signals in the microphone array.
4. The apparatus of claim 1, further comprising:
an impulse sound source signal emitting unit to emit an impulse sound source signal; and
an impulse sound signal receiving unit to receive the impulse sound source signal and to calculate an impulse response based on the received impulse sound source signal,
wherein the recording space information obtaining unit obtains the recording space information based on the generated impulse response.
5. The apparatus of claim 4, wherein the impulse response includes a plurality of impulse signals, and the recording space information includes at least one of a incoming time difference between the plurality of impulse signals, a sound pressure level difference between the plurality of impulse signals, and a incoming azimuth difference between the plurality of impulse signals.
6. The apparatus of claim 1, further comprising:
a multi-channel audio mixing unit to generate a multi-channel audio signal by mixing at least one of the plurality of object audio signals, the recording space information, and the sound source location information,
wherein the encoding unit further encodes the multi-channel audio signal.
7. An apparatus of reproducing object based audio contents, the apparatus comprising:
a decoding unit to decode a plurality of object audio signals of a plurality of sound source signals and sound source location information of the plurality of sound source signals, from the object based audio contents;
a reproducing space information obtaining unit to obtain reproducing space information with respect to a reproducing space of the plurality of object based audio contents;
a signal synthesizing unit to synthesize a plurality of speaker signals from the decoded plurality of object audio signals based on the sound source location information and the reproducing space information; and
a transmitting unit to transmit the plurality of speaker signals to a plurality of speakers respectively corresponding to the plurality of speaker signals.
8. The apparatus of claim 7, wherein the reproducing space information includes at least one of the plurality of speakers, an interval between the plurality of speakers, an arrangement angle of the plurality of speakers, a type of the plurality of speakers, location information of the speaker, and size information of the reproducing space.
9. The apparatus of claim 7, wherein the decoding unit further decodes recording space information of the plurality of sound source signals from the object based audio contents, and
the signal synthesizing unit directly generates a direct sound with respect to the plurality of sound source signals from the object based audio signal using the sound source location information and the reproducing space information, and synthesizes the plurality of speaker signals by adding a reflection sound to the direct sound based on the direct sound and the recording space information.
10. The apparatus of claim 7, wherein the signal synthesizing unit adds a reverberation effect to the speaker signal using an infinite impulse response filter.
US12/628,317 2008-12-02 2009-12-01 Apparatus for generating and playing object based audio contents Expired - Fee Related US8351612B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/709,475 US20130101122A1 (en) 2008-12-02 2012-12-10 Apparatus for generating and playing object based audio contents

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR20080121112 2008-12-02
KR10-2008-0121112 2008-12-02
KR10-2009-0020190 2009-03-10
KR1020090020190A KR20100062784A (en) 2008-12-02 2009-03-10 Apparatus for generating and playing object based audio contents

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/709,475 Continuation US20130101122A1 (en) 2008-12-02 2012-12-10 Apparatus for generating and playing object based audio contents

Publications (2)

Publication Number Publication Date
US20100135510A1 true US20100135510A1 (en) 2010-06-03
US8351612B2 US8351612B2 (en) 2013-01-08

Family

ID=41621914

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/628,317 Expired - Fee Related US8351612B2 (en) 2008-12-02 2009-12-01 Apparatus for generating and playing object based audio contents
US13/709,475 Abandoned US20130101122A1 (en) 2008-12-02 2012-12-10 Apparatus for generating and playing object based audio contents

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/709,475 Abandoned US20130101122A1 (en) 2008-12-02 2012-12-10 Apparatus for generating and playing object based audio contents

Country Status (2)

Country Link
US (2) US8351612B2 (en)
EP (1) EP2194527A3 (en)

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070061026A1 (en) * 2005-09-13 2007-03-15 Wen Wang Systems and methods for audio processing
US20100226500A1 (en) * 2006-04-03 2010-09-09 Srs Labs, Inc. Audio signal processing
US20110040397A1 (en) * 2009-08-14 2011-02-17 Srs Labs, Inc. System for creating audio objects for streaming
WO2013006338A3 (en) * 2011-07-01 2013-10-10 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
WO2014003513A1 (en) * 2012-06-29 2014-01-03 인텔렉추얼디스커버리 주식회사 Apparatus and method for evaluating a source of sound from user
WO2014028890A1 (en) * 2012-08-16 2014-02-20 Parametric Sound Corporation Multi-dimensional parametric audio system and method
WO2014035902A3 (en) * 2012-08-31 2014-04-17 Dolby Laboratories Licensing Corporation Reflected and direct rendering of upmixed content to individually addressable drivers
US20140140515A1 (en) * 2012-11-20 2014-05-22 Electronics And Telecommunications Research Institute Apparatus and method for generating multimedia data, and apparatus and method for playing multimedia data
US20140153752A1 (en) * 2012-12-05 2014-06-05 Samsung Electronics Co., Ltd Audio apparatus, method of processing audio signal, and a computer-readable recording medium storing program for performing the method
US9026450B2 (en) 2011-03-09 2015-05-05 Dts Llc System for dynamically creating and rendering audio objects
US20150356975A1 (en) * 2013-01-15 2015-12-10 Electronics And Telecommunications Research Institute Apparatus for processing audio signal for sound bar and method therefor
US20150382127A1 (en) * 2013-02-22 2015-12-31 Dolby Laboratories Licensing Corporation Audio spatial rendering apparatus and method
US20160029138A1 (en) * 2013-04-03 2016-01-28 Dolby Laboratories Licensing Corporation Methods and Systems for Interactive Rendering of Object Based Audio
CN105578380A (en) * 2011-07-01 2016-05-11 杜比实验室特许公司 System and Method for Adaptive Audio Signal Generation, Coding and Rendering
US20160133267A1 (en) * 2013-07-22 2016-05-12 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Concept for audio encoding and decoding for audio channels and audio objects
US9558785B2 (en) 2013-04-05 2017-01-31 Dts, Inc. Layered audio coding and transmission
US20170086005A1 (en) * 2014-03-25 2017-03-23 Intellectual Discovery Co., Ltd. System and method for processing audio signal
CN107454511A (en) * 2012-08-31 2017-12-08 杜比实验室特许公司 For making sound from viewing screen or the loudspeaker of display surface reflection
US10277998B2 (en) 2013-07-22 2019-04-30 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for low delay object metadata coding
US10282160B2 (en) * 2012-10-11 2019-05-07 Electronics And Telecommunications Research Institute Apparatus and method for generating audio data, and apparatus and method for playing audio data
WO2019127759A1 (en) * 2017-12-28 2019-07-04 武汉华星光电半导体显示技术有限公司 Voice image acquisition coding method and device
US10701504B2 (en) 2013-07-22 2020-06-30 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for realizing a SAOC downmix of 3D audio content
US20210385597A1 (en) * 2020-06-03 2021-12-09 Yamaha Corporation Sound signal processing method, sound signal processing device, and storage medium that stores sound signal processing program
US11223923B2 (en) * 2016-12-30 2022-01-11 Zte Corporation Data processing method and apparatus, acquisition device, and storage medium
CN114731459A (en) * 2019-11-20 2022-07-08 杜比国际公司 Method and apparatus for personalizing audio content

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100942142B1 (en) * 2007-10-11 2010-02-16 한국전자통신연구원 Method and apparatus for transmitting and receiving of the object based audio contents
KR101547809B1 (en) 2011-07-01 2015-08-27 돌비 레버러토리즈 라이쎈싱 코오포레이션 Synchronization and switchover methods and systems for an adaptive audio system
WO2014184706A1 (en) * 2013-05-16 2014-11-20 Koninklijke Philips N.V. An audio apparatus and method therefor
WO2014184353A1 (en) 2013-05-16 2014-11-20 Koninklijke Philips N.V. An audio processing apparatus and method therefor
CN104240711B (en) 2013-06-18 2019-10-11 杜比实验室特许公司 For generating the mthods, systems and devices of adaptive audio content
WO2015017037A1 (en) 2013-07-30 2015-02-05 Dolby International Ab Panning of audio objects to arbitrary speaker layouts
JP6291035B2 (en) * 2014-01-02 2018-03-14 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Audio apparatus and method therefor
KR102170398B1 (en) 2014-03-12 2020-10-27 삼성전자 주식회사 Method and apparatus for performing multi speaker using positional information
KR20160020377A (en) 2014-08-13 2016-02-23 삼성전자주식회사 Method and apparatus for generating and reproducing audio signal
EP3286930B1 (en) 2015-04-21 2020-05-20 Dolby Laboratories Licensing Corporation Spatial audio signal manipulation
CN105070304B (en) * 2015-08-11 2018-09-04 小米科技有限责任公司 Realize method and device, the electronic equipment of multi-object audio recording
CN105761721A (en) * 2016-03-16 2016-07-13 广东佳禾声学科技有限公司 Voice coding method carrying position information
KR20190083863A (en) * 2018-01-05 2019-07-15 가우디오랩 주식회사 A method and an apparatus for processing an audio signal
US11393483B2 (en) 2018-01-26 2022-07-19 Lg Electronics Inc. Method for transmitting and receiving audio data and apparatus therefor
CN113039815B (en) * 2018-11-09 2022-11-11 候本株式会社 Sound generating method and device for executing the same
WO2023212879A1 (en) * 2022-05-05 2023-11-09 北京小米移动软件有限公司 Object audio data generation method and apparatus, electronic device, and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040111171A1 (en) * 2002-10-28 2004-06-10 Dae-Young Jang Object-based three-dimensional audio system and method of controlling the same
US20050141723A1 (en) * 2003-12-29 2005-06-30 Tae-Jin Lee 3D audio signal processing system using rigid sphere and method thereof
US20060115100A1 (en) * 2004-11-30 2006-06-01 Christof Faller Parametric coding of spatial audio with cues based on transmitted channels

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100739798B1 (en) 2005-12-22 2007-07-13 삼성전자주식회사 Method and apparatus for reproducing a virtual sound of two channels based on the position of listener
CN101356573B (en) * 2006-01-09 2012-01-25 诺基亚公司 Control for decoding of binaural audio signal

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040111171A1 (en) * 2002-10-28 2004-06-10 Dae-Young Jang Object-based three-dimensional audio system and method of controlling the same
US20050141723A1 (en) * 2003-12-29 2005-06-30 Tae-Jin Lee 3D audio signal processing system using rigid sphere and method thereof
US20060115100A1 (en) * 2004-11-30 2006-06-01 Christof Faller Parametric coding of spatial audio with cues based on transmitted channels

Cited By (91)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8027477B2 (en) 2005-09-13 2011-09-27 Srs Labs, Inc. Systems and methods for audio processing
US20070061026A1 (en) * 2005-09-13 2007-03-15 Wen Wang Systems and methods for audio processing
US9232319B2 (en) 2005-09-13 2016-01-05 Dts Llc Systems and methods for audio processing
US20100226500A1 (en) * 2006-04-03 2010-09-09 Srs Labs, Inc. Audio signal processing
US8831254B2 (en) 2006-04-03 2014-09-09 Dts Llc Audio signal processing
US8396575B2 (en) 2009-08-14 2013-03-12 Dts Llc Object-oriented audio streaming system
US20110040396A1 (en) * 2009-08-14 2011-02-17 Srs Labs, Inc. System for adaptively streaming audio objects
US8396576B2 (en) 2009-08-14 2013-03-12 Dts Llc System for adaptively streaming audio objects
US8396577B2 (en) 2009-08-14 2013-03-12 Dts Llc System for creating audio objects for streaming
US20110040395A1 (en) * 2009-08-14 2011-02-17 Srs Labs, Inc. Object-oriented audio streaming system
US20110040397A1 (en) * 2009-08-14 2011-02-17 Srs Labs, Inc. System for creating audio objects for streaming
US9167346B2 (en) 2009-08-14 2015-10-20 Dts Llc Object-oriented audio streaming system
US9721575B2 (en) 2011-03-09 2017-08-01 Dts Llc System for dynamically creating and rendering audio objects
US9165558B2 (en) 2011-03-09 2015-10-20 Dts Llc System for dynamically creating and rendering audio objects
US9026450B2 (en) 2011-03-09 2015-05-05 Dts Llc System for dynamically creating and rendering audio objects
KR101845226B1 (en) 2011-07-01 2018-05-18 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
RU2617553C2 (en) * 2011-07-01 2017-04-25 Долби Лабораторис Лайсэнзин Корпорейшн System and method for generating, coding and presenting adaptive sound signal data
KR102003191B1 (en) 2011-07-01 2019-07-24 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
KR20190086785A (en) * 2011-07-01 2019-07-23 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
CN103650539A (en) * 2011-07-01 2014-03-19 杜比实验室特许公司 System and method for adaptive audio signal generation, coding and rendering
KR20200058593A (en) * 2011-07-01 2020-05-27 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
US9179236B2 (en) 2011-07-01 2015-11-03 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR20140017682A (en) * 2011-07-01 2014-02-11 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
US11962997B2 (en) 2011-07-01 2024-04-16 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
US10327092B2 (en) 2011-07-01 2019-06-18 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
AU2012279357B2 (en) * 2011-07-01 2016-01-14 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR102608968B1 (en) 2011-07-01 2023-12-05 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
CN105578380A (en) * 2011-07-01 2016-05-11 杜比实验室特许公司 System and Method for Adaptive Audio Signal Generation, Coding and Rendering
US11412342B2 (en) 2011-07-01 2022-08-09 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR102115723B1 (en) 2011-07-01 2020-05-28 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
CN105792086A (en) * 2011-07-01 2016-07-20 杜比实验室特许公司 System and method for adaptive audio signal generation, coding and rendering
US9467791B2 (en) 2011-07-01 2016-10-11 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR101685447B1 (en) 2011-07-01 2016-12-12 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
RU2731025C2 (en) * 2011-07-01 2020-08-28 Долби Лабораторис Лайсэнзин Корпорейшн System and method for generating, encoding and presenting adaptive audio signal data
KR20220081385A (en) * 2011-07-01 2022-06-15 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
KR102406776B1 (en) 2011-07-01 2022-06-10 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
KR102185941B1 (en) 2011-07-01 2020-12-03 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
US9622009B2 (en) 2011-07-01 2017-04-11 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR101946795B1 (en) 2011-07-01 2019-02-13 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
WO2013006338A3 (en) * 2011-07-01 2013-10-10 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
US9800991B2 (en) 2011-07-01 2017-10-24 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
EP3893521A1 (en) * 2011-07-01 2021-10-13 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR20190014601A (en) * 2011-07-01 2019-02-12 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
US10904692B2 (en) 2011-07-01 2021-01-26 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
US9942688B2 (en) 2011-07-01 2018-04-10 Dolby Laboraties Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
US10477339B2 (en) 2011-07-01 2019-11-12 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
KR20200137034A (en) * 2011-07-01 2020-12-08 돌비 레버러토리즈 라이쎈싱 코오포레이션 System and method for adaptive audio signal generation, coding and rendering
US10057708B2 (en) 2011-07-01 2018-08-21 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
US10165387B2 (en) 2011-07-01 2018-12-25 Dolby Laboratories Licensing Corporation System and method for adaptive audio signal generation, coding and rendering
WO2014003513A1 (en) * 2012-06-29 2014-01-03 인텔렉추얼디스커버리 주식회사 Apparatus and method for evaluating a source of sound from user
WO2014028890A1 (en) * 2012-08-16 2014-02-20 Parametric Sound Corporation Multi-dimensional parametric audio system and method
CN107454511A (en) * 2012-08-31 2017-12-08 杜比实验室特许公司 For making sound from viewing screen or the loudspeaker of display surface reflection
US11277703B2 (en) 2012-08-31 2022-03-15 Dolby Laboratories Licensing Corporation Speaker for reflecting sound off viewing screen or display surface
US9532158B2 (en) 2012-08-31 2016-12-27 Dolby Laboratories Licensing Corporation Reflected and direct rendering of upmixed content to individually addressable drivers
WO2014035902A3 (en) * 2012-08-31 2014-04-17 Dolby Laboratories Licensing Corporation Reflected and direct rendering of upmixed content to individually addressable drivers
US10282160B2 (en) * 2012-10-11 2019-05-07 Electronics And Telecommunications Research Institute Apparatus and method for generating audio data, and apparatus and method for playing audio data
US9357325B2 (en) * 2012-11-20 2016-05-31 Electronics And Telecommunications Research Institute Apparatus and method for generating multimedia data, and apparatus and method for playing multimedia data
US20140140515A1 (en) * 2012-11-20 2014-05-22 Electronics And Telecommunications Research Institute Apparatus and method for generating multimedia data, and apparatus and method for playing multimedia data
US20140153752A1 (en) * 2012-12-05 2014-06-05 Samsung Electronics Co., Ltd Audio apparatus, method of processing audio signal, and a computer-readable recording medium storing program for performing the method
US10462596B2 (en) * 2012-12-05 2019-10-29 Samsung Electronics Co., Ltd. Audio apparatus, method of processing audio signal, and a computer-readable recording medium storing program for performing the method
US20150356975A1 (en) * 2013-01-15 2015-12-10 Electronics And Telecommunications Research Institute Apparatus for processing audio signal for sound bar and method therefor
US20150382127A1 (en) * 2013-02-22 2015-12-31 Dolby Laboratories Licensing Corporation Audio spatial rendering apparatus and method
US9854378B2 (en) * 2013-02-22 2017-12-26 Dolby Laboratories Licensing Corporation Audio spatial rendering apparatus and method
US20160029138A1 (en) * 2013-04-03 2016-01-28 Dolby Laboratories Licensing Corporation Methods and Systems for Interactive Rendering of Object Based Audio
US11727945B2 (en) * 2013-04-03 2023-08-15 Dolby Laboratories Licensing Corporation Methods and systems for interactive rendering of object based audio
US20220059103A1 (en) * 2013-04-03 2022-02-24 Dolby International Ab Methods and systems for interactive rendering of object based audio
US10515644B2 (en) 2013-04-03 2019-12-24 Dolby Laboratories Licensing Corporation Methods and systems for interactive rendering of object based audio
US9997164B2 (en) * 2013-04-03 2018-06-12 Dolby Laboratories Licensing Corporation Methods and systems for interactive rendering of object based audio
US11081118B2 (en) 2013-04-03 2021-08-03 Dolby Laboratories Licensing Corporation Methods and systems for interactive rendering of object based audio
US9837123B2 (en) 2013-04-05 2017-12-05 Dts, Inc. Layered audio reconstruction system
US9558785B2 (en) 2013-04-05 2017-01-31 Dts, Inc. Layered audio coding and transmission
US9613660B2 (en) 2013-04-05 2017-04-04 Dts, Inc. Layered audio reconstruction system
US10249311B2 (en) * 2013-07-22 2019-04-02 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Concept for audio encoding and decoding for audio channels and audio objects
US11463831B2 (en) 2013-07-22 2022-10-04 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for efficient object metadata coding
US10659900B2 (en) 2013-07-22 2020-05-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for low delay object metadata coding
US11227616B2 (en) * 2013-07-22 2022-01-18 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Concept for audio encoding and decoding for audio channels and audio objects
US10277998B2 (en) 2013-07-22 2019-04-30 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for low delay object metadata coding
US10715943B2 (en) 2013-07-22 2020-07-14 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for efficient object metadata coding
US20220101867A1 (en) * 2013-07-22 2022-03-31 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Concept for audio encoding and decoding for audio channels and audio objects
US11330386B2 (en) 2013-07-22 2022-05-10 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for realizing a SAOC downmix of 3D audio content
US11337019B2 (en) 2013-07-22 2022-05-17 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for low delay object metadata coding
US11910176B2 (en) 2013-07-22 2024-02-20 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for low delay object metadata coding
US10701504B2 (en) 2013-07-22 2020-06-30 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Apparatus and method for realizing a SAOC downmix of 3D audio content
US20190180764A1 (en) * 2013-07-22 2019-06-13 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Concept for audio encoding and decoding for audio channels and audio objects
US20160133267A1 (en) * 2013-07-22 2016-05-12 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Concept for audio encoding and decoding for audio channels and audio objects
US20170086005A1 (en) * 2014-03-25 2017-03-23 Intellectual Discovery Co., Ltd. System and method for processing audio signal
US11223923B2 (en) * 2016-12-30 2022-01-11 Zte Corporation Data processing method and apparatus, acquisition device, and storage medium
WO2019127759A1 (en) * 2017-12-28 2019-07-04 武汉华星光电半导体显示技术有限公司 Voice image acquisition coding method and device
CN114731459A (en) * 2019-11-20 2022-07-08 杜比国际公司 Method and apparatus for personalizing audio content
US20210385597A1 (en) * 2020-06-03 2021-12-09 Yamaha Corporation Sound signal processing method, sound signal processing device, and storage medium that stores sound signal processing program
US11659344B2 (en) * 2020-06-03 2023-05-23 Yamaha Corporation Sound signal processing method, sound signal processing device, and storage medium that stores sound signal processing program

Also Published As

Publication number Publication date
EP2194527A2 (en) 2010-06-09
US20130101122A1 (en) 2013-04-25
EP2194527A3 (en) 2013-09-25
US8351612B2 (en) 2013-01-08

Similar Documents

Publication Publication Date Title
US8351612B2 (en) Apparatus for generating and playing object based audio contents
RU2741738C1 (en) System, method and permanent machine-readable data medium for generation, coding and presentation of adaptive audio signal data
US20220159400A1 (en) Reproduction apparatus, reproduction method, information processing apparatus, information processing method, and program
JP5337941B2 (en) Apparatus and method for multi-channel parameter conversion
US10356545B2 (en) Method and device for processing audio signal by using metadata
JP5688030B2 (en) Method and apparatus for encoding and optimal reproduction of a three-dimensional sound field
KR20100062784A (en) Apparatus for generating and playing object based audio contents
CN104822036B (en) The technology of audio is perceived for localization
US20160165375A1 (en) Method and apparatus for generating side information bitstream of multi-object audio signal
KR102213895B1 (en) Encoding/decoding apparatus and method for controlling multichannel signals
US20210377691A1 (en) Signal processing device, method, and program
US10271156B2 (en) Audio signal processing method
JP2022515998A (en) A device and method for playing a spatially expanded sound source, or a device and method for generating a bitstream from a spatially expanded sound source.
CN104756524A (en) Apparatus and method for creating proximity sound effects in audio systems
US20140310010A1 (en) Apparatus for encoding and apparatus for decoding supporting scalable multichannel audio signal, and method for apparatuses performing same
KR20040005025A (en) Implementing method of multi channel sound and apparatus thereof
CN112823534B (en) Signal processing device and method, and program
CN114915874B (en) Audio processing method, device, equipment and medium
KR101114431B1 (en) Apparatus for generationg and reproducing audio data for real time audio stream and the method thereof

Legal Events

Date Code Title Description
AS Assignment

Owner name: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTIT

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YOO, JAE-HYOUN;SHIM, HWAN;CHUNG, HYUN-JOO;AND OTHERS;SIGNING DATES FROM 20090515 TO 20090525;REEL/FRAME:023584/0100

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20210108