[go: up one dir, main page]

WO2009001292A1 - Procédé de fusion d'au moins deux trains de paramètres audio orientés objet d'entrée en un train de paramètres audio orientés objet de sortie - Google Patents

Procédé de fusion d'au moins deux trains de paramètres audio orientés objet d'entrée en un train de paramètres audio orientés objet de sortie Download PDF

Info

Publication number
WO2009001292A1
WO2009001292A1 PCT/IB2008/052502 IB2008052502W WO2009001292A1 WO 2009001292 A1 WO2009001292 A1 WO 2009001292A1 IB 2008052502 W IB2008052502 W IB 2008052502W WO 2009001292 A1 WO2009001292 A1 WO 2009001292A1
Authority
WO
WIPO (PCT)
Prior art keywords
oriented
audio parameter
stream
oriented audio
parameter values
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/IB2008/052502
Other languages
English (en)
Inventor
Dirk J. Breebaart
Erik G. P. Schuijers
Arnoldus W. J. Oomen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Koninklijke Philips NV
Original Assignee
Koninklijke Philips Electronics NV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics NV filed Critical Koninklijke Philips Electronics NV
Publication of WO2009001292A1 publication Critical patent/WO2009001292A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • G10L19/20Vocoders using multiple modes using sound class specific coding, hybrid encoders or object based coding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/167Audio streaming, i.e. formatting and decoding of an encoded audio signal representation into a data stream for transmission or storage purposes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/03Application of parametric coding in stereophonic audio systems

Definitions

  • a method of merging at least two input object-oriented audio parameter streams into an output object-oriented audio parameter stream is a method of merging at least two input object-oriented audio parameter streams into an output object-oriented audio parameter stream
  • the invention relates to a method of merging at least two input object-oriented audio parameter streams into an output object-oriented audio parameter stream, each object- oriented audio parameter stream comprising object-oriented audio parameter values, said object-oriented audio parameter values representing statistical properties of audio objects as a function of time.
  • MPEG framework a workgroup has been started on object-based spatial audio coding.
  • the aim of this workgroup is to "explore new technology and reuse of current MPEG Surround components and technologies for the bit rate efficient coding of multiple sound sources or objects into a number of down-mix channels and corresponding spatial parameters".
  • the aim is to encode multiple audio objects in a limited set of down-mix channels with corresponding parameters.
  • users may interact with the content for example by repositioning the individual audio objects.
  • Such interaction with the content is easily realized in object-oriented decoders. It is then realized by including a rendering step that follows the decoding process. Said rendering is combined with the decoding as a single processing step to prevent the need of determining individual objects.
  • Said rendering is combined with the decoding as a single processing step to prevent the need of determining individual objects.
  • For loudspeaker playback such combination is described in Faller, C, "Parametric joint-coding of audio sources", Proc. 120 th AES Convention, Paris, France, May 2006.
  • headphone playback an efficient combination of decoding and head- related transfer function processing is described in Breebaart, J., Herre, J., Villemoes, L., Jin, C, Kj ⁇ rling, K., Plogsties, J., Koppens, J. (2006), "Multi-channel goes mobile: MPEG Surround binaural rendering", Proc. 29th AES conference, Seoul, Korea.
  • object-oriented audio parameters as provided by object-oriented coding form a stream that reflects statistical properties of the audio objects as a function of time. Therefore, these object-oriented audio parameters are valid for one certain frame or even for a portion of a frame.
  • Each object-oriented audio parameter stream comprises object-oriented audio parameter values.
  • the object-oriented audio parameter values represent statistical properties of audio objects as a function of time.
  • the method comprises the following steps. First, calculating a synchronized stream for each said input object-oriented audio parameter stream takes place. Said synchronized stream has object- oriented parameter values at predetermined temporal positions. Said predetermined temporal positions are the same for all synchronized streams. For each synchronized stream said object-oriented parameter values at the predetermined temporal positions are calculated by means of interpolating of the object-oriented parameter values of the corresponding input object-oriented audio parameter stream. Second, creating of the output object-oriented audio parameter stream is performed. Said output object-oriented audio parameter stream has object-oriented parameter values at the predetermined temporal position obtained by combining the object-oriented audio parameter values of the synchronized streams at said same predetermined temporal position.
  • the advantage of the method of merging at least two input object-oriented audio parameter streams according to the invention is that no delaying of the object-oriented audio parameter streams is required in order to merge said streams. Instead a very simple processing is performed in order to obtain the object-oriented audio parameter values of the synchronized streams.
  • the additional benefit of the proposed method is that the problem of different object-oriented audio parameter positions within frames across the object-oriented audio parameter streams is overcome.
  • filtering is applied to the object-oriented audio parameter values of the synchronized stream. Applying, e.g. a simple piecewise linear interpolation has an effect that the resulting interpolated object-oriented audio parameter values are low-pass filtered.
  • linear interpolation e.g.
  • high-pass filtering can be employed to reduce the effect of low-pass filtering due to interpolation.
  • the advantage of applying filtering to the object-oriented audio parameter values of the synchronized stream is that it ensures a similar dynamic behavior as the corresponding input object-oriented audio parameter stream. In other words it improves the quality of synchronization, as it helps the object-oriented audio parameter values of the synchronized stream to resemble the original behavior of said parameters.
  • Said synchronization process provides the synchronized stream for the corresponding input object-oriented audio parameter stream.
  • the applied filtering is adaptive in order to match statistical properties of the corresponding input object-oriented audio parameter values.
  • the object-oriented audio parameters reflect statistical properties of the audio objects as a function of time, it is desired that the synchronization process takes these fluctuations over time into account.
  • a Linear Predictive Coding analysis is used to determine an envelope of the object-oriented audio parameter values of the input object-oriented audio parameter stream and subsequently said envelope is imposed on the object-oriented audio parameter values of the corresponding synchronized stream during the filtering.
  • This way of post-processing/filtering of the object-oriented audio parameter values of the synchronized stream ensures that the resulting synchronized parameters possess similar characteristics as the original input parameters.
  • the mutually different frequency resolutions are matched by means of up-sampling the object-oriented audio parameter values to a higher frequency resolution. It can occur that the input object-oriented audio parameter streams may have mutually different frequency resolutions. In this case, the frequency resolutions must be matched.
  • the up-sampling of object-oriented audio parameter values to a higher frequency resolution has the advantage that it is simple, and does not require an extensive computational effort. Since in most systems, frequency resolutions have common edge/threshold frequencies for parameter bands, the up-sampling can be achieved by simply copying appropriate parameter values.
  • the invention further provides device claims, a computer program product enabling a programmable device to perform the method according to the invention, as well as a teleconferencing system comprising a device according to the invention.
  • Fig. 1 schematically shows a flow diagram of a method of merging at least two input object-oriented audio parameter streams into an output object-oriented audio parameter stream according to the invention
  • Fig. 2 schematically shows an example architecture in which two input object- oriented audio parameter streams are merged into an output object-oriented audio parameter stream;
  • Fig. 3 shows object-oriented audio parameter values of the synchronized stream that are obtained by means of interpolation between object-oriented audio parameter values of the corresponding input object-oriented audio parameter stream at the predetermined temporal positions;
  • Fig. 4 shows schematically an architecture in which filtering is applied to the object-oriented audio parameter values of the synchronized stream
  • Fig. 5 illustrates a use of a Linear Predictive Coding analysis to determine a spectral envelope of the object-oriented audio parameter values of the input object-oriented audio parameter stream and subsequently imposing said envelope on the object-oriented audio parameter values of the corresponding synchronized stream during the filtering;
  • Fig. 6 illustrates matching of the mutually different frequency resolutions by means of up-sampling the object-oriented audio parameter values to a higher frequency resolution.
  • Fig. 1 schematically shows a flow diagram of a method of merging at least two input object-oriented audio parameter streams into an output object-oriented audio parameter stream according to the invention.
  • Each object-oriented audio parameter stream comprises object-oriented audio parameter values.
  • Said object-oriented audio parameter values represent statistical properties of audio objects as a function of time.
  • the method comprises the following steps.
  • the step 110 comprises calculating a synchronized stream for each said input object-oriented audio parameter stream.
  • Said synchronized stream has object-oriented parameter values at predetermined temporal positions. Said predetermined temporal positions are the same for all synchronized streams.
  • said object-oriented parameter values at the predetermined temporal positions are calculated by means of interpolating of the object-oriented parameter values of the corresponding input object- oriented audio parameter stream.
  • the step 120 comprises creating of the output object- oriented audio parameter stream.
  • Said output object-oriented audio parameter stream has object-oriented parameter values at the predetermined temporal position obtained by combining the object-oriented audio parameter values of the synchronized streams at said same predetermined temporal position.
  • Said combining is realized by means of concatenating of object-oriented audio parameter values corresponding to the input object-oriented audio parameter streams.
  • the object-oriented audio parameters as comprised in the object-oriented audio parameter streams are arranged according to time/frequency tiles at each temporal position.
  • Each object-oriented audio parameter is associated with an audio object, and each audio object is assigned to one of the time/frequency tiles in turn. Therefore, the concatenation of object-oriented audio parameter values corresponding to the input object- oriented audio parameter streams as used in the step 120 is performed for each of the time/frequency tile separately.
  • the sequence of concatenated parameter values for a specific time/frequency tile is arbitrary.
  • Fig. 2 schematically shows an example architecture in which two input object- oriented audio parameter streams are merged into an output object-oriented audio parameter stream.
  • the object-oriented audio parameter will be used interchangeably with the object parameters throughout the present application.
  • Object-oriented encoders 210 and 220 at the two participant sides encode various audio object sets 201 and 202 at these two sides, respectively.
  • Each of the decoders generates a down-mix stream 211 and 212, respectively, and the corresponding object parameter streams 221 and 222, respectively.
  • These two object-oriented audio streams are fed into the unit 230, which performs merging of the streams.
  • This unit 230 generates the merged down- mix 213 and the merged object parameter stream.
  • the merged object-oriented audio stream can be decoded at the receiving participant side in the object-oriented decoder 240 based on the user data 224.
  • Said user data 224 regards e.g. the object positioning in the three-dimensional space.
  • a rendered output 214 is generated that is presented to the user.
  • the object parameters reflect statistical properties of the audio objects as a function of time. Hence, these parameters are time varying and are valid only for specific time intervals. These time intervals are typically called frames. Said object parameters are valid for a certain frame or even a portion of frame. In a teleconferencing application, it is very unlikely that frames of multiple object-oriented audio encoders are perfectly aligned in time. Moreover, different audio objects may give rise to different framing and to different object parameter positions, as object parameter positions are preferably content dependent. Without the synchronization processing of the proposed invention, merging of the object- oriented audio streams would require to delay at least one of the down-mixes and the corresponding object parameters in order to align frame boundaries of the streams to be merged.
  • the advantage of the method of merging at least two object-oriented audio parameter streams according to the invention is that no delaying of the object-oriented audio parameter streams is required in order to merge said streams. Instead a very simple processing is performed in order to obtain the object-oriented audio parameter values of the synchronized stream.
  • the additional benefit of the proposed method is that the problem of different object-oriented audio parameter positions within frames across the object-oriented audio parameter streams is overcome.
  • Fig. 3 shows object-oriented audio parameter values 331, 332, and 333 of the synchronized stream that are obtained by means of interpolation between object-oriented audio parameter values 321, 322, 323, and 324, of the corresponding input object-oriented audio parameter stream at the predetermined temporal positions 351, 352, and 353.
  • Fig. 3 depicts object parameter values for three consecutive frames 0, 1, and 2, for input object parameter streams 310 and 320.
  • the input object parameter stream 310 is a reference stream that remains intact, while the second input object parameter stream 320 is going to be synchronized so that the new positions of the object parameters align with the object parameters of the input object parameter stream 310.
  • the framing and the positioning of the object parameters of the two input streams 310 and 320 are not aligned.
  • the framing and parameter positions of the object parameter stream 310 are copied to the synchronized stream 330.
  • the object parameter values of the object parameter stream 320 are interpolated between the object parameter values of the input object parameter stream 320, as indicated by the dashed lines.
  • the interpolated object parameter values at the temporal positions 311, 312, and 313, are copied to the synchronized stream 330.
  • the predetermined temporal positions correspond to temporal positions of the object-oriented audio parameter values in one of the input object- oriented audio parameter streams.
  • both streams could be synchronized at the temporal positions missing in the other object parameter stream.
  • Yet another option is to select the temporal positions depending on the density of synchronization positions and/or computational complexity.
  • two input streams are merged also larger numbers of the input streams can be merged. It is very much application dependent. For the teleconferencing application the number of streams to be merged will take the number of participants a single participant is communicating with.
  • the interpolation is piecewise linear.
  • interpolation of object parameters can be implemented in a different domain, for example:
  • log ⁇ (n,b,p)) w p _ ⁇ log ⁇ (n,b,p - ⁇ ))+ w p+ ⁇ log ⁇ (n,b,p + ⁇ )) .
  • Fig. 4 shows schematically an architecture in which filtering is applied to the object-oriented audio parameter values of the synchronized stream.
  • the applied filtering is adaptive to statistical properties of the corresponding input object-oriented audio parameter values.
  • the elements 401 and 402 represent the input object parameter streams, whereby the stream 401 is the reference stream that provides the temporal positions for synchronizing of the second object parameter stream 402. Said temporal positions for synchronizing are comprised in the control parameters denoted by 411.
  • the object parameters with the corresponding temporal positions 412 of the object parameter stream 402 are fed into synchronization unit 410 and statistics processing unit 420.
  • the unit 410 synchronizes the object parameter stream 402 using an interpolation. Applying, e.g.
  • the resulting synchronized stream 431 is fed into the statistics processing unit 420 and to a filter 440.
  • the filter 440 is preferably a Io w- order high-pass filter.
  • filter coefficients 432 are determined by the unit 420 and provided to the filter 440.
  • Said filter 440 generates at its output synchronized object parameter values 441 that exhibit a similar dynamic behavior as in the original input object-oriented audio parameter stream.
  • Fig. 5 illustrates a use of a Linear Predictive Coding analysis to determine a spectral envelope of the object-oriented audio parameter values of the input object-oriented audio parameter stream and subsequently imposing said envelope on the object-oriented audio parameter values of the corresponding synchronized stream during the filtering.
  • temporal positions 511 are derived and fed into the synchronization unit 510.
  • object parameters and the corresponding temporal positions 512 are derived and fed into the synchronization unit 510.
  • the synchronization unit 510 synchronizes the object parameters of the stream 502 at the positions 511.
  • an LPC analysis is conducted on both the original input object parameters as well as the synchronized object parameters in the units 520 and 540, respectively.
  • the LPC analysis performed on the synchronized object parameters 531 in the unit 550 results in the so-called spectral whitening of the values provided at its input.
  • the filter unit 550 acts as a spectral whitening filter.
  • the second filter unit 570 imposes the spectral envelope of the original object parameters of the stream 502 onto the whitened object parameters 562.
  • the two filter stages 550 and 570 can be combined into a single filter unit.
  • the LPC analysis is preferably conducted on an auto-correlation.
  • the auto- correlation estimates are performed using a sliding window over time.
  • the input object-oriented audio parameter streams have mutually different frequency resolutions.
  • the input object-oriented audio streams to be merged are encoded by different object-oriented encoders, which might use mutually different frequency resolutions.
  • the mutually different frequency resolutions are matched by means of averaging the object-oriented audio parameter values for a higher frequency resolution. It is a very simple way of dealing with this issue that does not require excessive computational effort.
  • the object-oriented audio parameters comprise at least (relative) level information defined for separate time/frequency tiles of various audio objects. Said level information is related to an amount of energy in the time/frequency tile of the audio object that the parameter value refers to.
  • Said frequency used for time/frequency tiles can have different resolutions, or in other words can have different splitting into frequency bands. The number of frequency bands as well as their width can vary for different resolutions.
  • Fig. 6 illustrates matching of the mutually different frequency resolutions by means of up-sampling the object-oriented audio parameter values to a higher frequency resolution. Since in most systems frequency resolutions have common edge frequencies for parameter bands up-sampling can be achieved by copying of appropriate object-oriented audio parameter values.
  • the reference input object parameter stream has 9 frequency bands depicted by 610 in Fig. 6.
  • the second input object parameter stream has a lower frequency resolution 630 with 6 bands. Given the common edge frequencies of both object parameter streams are aligned, the up-sampled frequency resolution of the second stream is obtained by copying parameter values, depicted by dashed arrows, to overlapping frequency bands as depicted in 620 of the figure.
  • the band bl in 630 is an equivalent of the two bands bl and b2 in 610. Therefore the left object parameter value from bl of 630 is copied to bl band of 620, while the right object parameter value from bl of 630 is copied to b2 band of 620.
  • the proposed method of merging at least two input object-oriented audio parameter streams into an output object-oriented audio parameter stream can be realized in a device that comprises: synchronizing means and combining means.
  • the processing means calculate the synchronized stream for each input object parameter stream.
  • Said synchronized stream has object-oriented parameter values at the predetermined temporal positions.
  • Said predetermined temporal positions are the same for all synchronized streams.
  • For each synchronized stream said object-oriented parameter values at the predetermined temporal positions are calculated by means of interpolating of the object-oriented parameter values of the corresponding input object parameter stream.
  • the combining means create the output object parameter stream.
  • the output object parameter has object-oriented parameter values at the predetermined temporal position obtained by combining the object-oriented audio parameter values of the synchronized streams at said same predetermined temporal position. Said combining is realized by means of concatenating of object-oriented audio parameter values corresponding to the input object-oriented audio parameter streams.
  • a computer program product executes the method according to the invention.
  • a teleconferencing system comprises a device according to the invention.
  • any reference signs placed between parentheses shall not be construed as limiting the claim.
  • the word “comprising” does not exclude the presence of elements or steps other than those listed in a claim.
  • the word “a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
  • the invention can be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Stereophonic System (AREA)

Abstract

Il est décrit un procédé de fusion d'au moins deux trains de paramètres audio orientés objet d'entrée en un train de paramètres audio orientés objet de sortie. Chaque train de paramètres audio orientés objet comprend des valeurs de paramètres audio orientés objet. Lesdites valeurs de paramètres audio orientés objet représentent des propriétés statistiques d'objets audio en fonction du temps. Ledit procédé comprend les étapes suivantes. Premièrement, le calcul d'un train synchronisé pour chacun desdits trains de paramètres audio orientés objet d'entrée est effectué. Ledit train synchronisé a des valeurs de paramètres orientés objet à des positions temporelles prédéterminées. Lesdites positions temporelles prédéterminées sont les mêmes pour tous les trains synchronisés. Pour chaque train synchronisé, lesdites valeurs de paramètres orientés objet à des positions temporelles prédéterminées sont calculées à l'aide d'une interpolation des valeurs de paramètres orientés objet du train de paramètres audio orientés objet d'entrée correspondant. Deuxièmement, la création du train de paramètres audio orientés objet de sortie est effectuée. Ledit train de paramètres audio orientés objet de sortie a des valeurs de paramètres orientés objet à la position temporelle prédéterminée obtenues par combinaison des valeurs de paramètres audio orientés objet des trains synchronisés à ladite position temporelle prédéterminée identique.
PCT/IB2008/052502 2007-06-27 2008-06-24 Procédé de fusion d'au moins deux trains de paramètres audio orientés objet d'entrée en un train de paramètres audio orientés objet de sortie Ceased WO2009001292A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP07111149 2007-06-27
EP07111149.6 2007-06-27

Publications (1)

Publication Number Publication Date
WO2009001292A1 true WO2009001292A1 (fr) 2008-12-31

Family

ID=39768243

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2008/052502 Ceased WO2009001292A1 (fr) 2007-06-27 2008-06-24 Procédé de fusion d'au moins deux trains de paramètres audio orientés objet d'entrée en un train de paramètres audio orientés objet de sortie

Country Status (2)

Country Link
TW (1) TW200921643A (fr)
WO (1) WO2009001292A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011020067A1 (fr) * 2009-08-14 2011-02-17 Srs Labs, Inc. Système pour diffuser en continu de manière adaptative des objets audio
WO2014184618A1 (fr) * 2013-05-17 2014-11-20 Nokia Corporation Appareil audio orienté objet spatial
US9026450B2 (en) 2011-03-09 2015-05-05 Dts Llc System for dynamically creating and rendering audio objects
US9558785B2 (en) 2013-04-05 2017-01-31 Dts, Inc. Layered audio coding and transmission
TWI607654B (zh) * 2011-07-01 2017-12-01 杜比實驗室特許公司 用於增強3d音頻編輯與呈現之設備、方法及非暫態媒體

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6463414B1 (en) * 1999-04-12 2002-10-08 Conexant Systems, Inc. Conference bridge processing of speech in a packet network environment
WO2004039096A1 (fr) * 2002-10-25 2004-05-06 Dilithium Networks Pty Limited Procede et appareil pour une detection dtmf et un melange vocal dans le domaine de parametres celp
US20050102137A1 (en) * 2001-04-02 2005-05-12 Zinser Richard L. Compressed domain conference bridge
WO2005078707A1 (fr) * 2004-02-16 2005-08-25 Koninklijke Philips Electronics N.V. Transcodeur et procede de transcodage a cet effet

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6463414B1 (en) * 1999-04-12 2002-10-08 Conexant Systems, Inc. Conference bridge processing of speech in a packet network environment
US20050102137A1 (en) * 2001-04-02 2005-05-12 Zinser Richard L. Compressed domain conference bridge
WO2004039096A1 (fr) * 2002-10-25 2004-05-06 Dilithium Networks Pty Limited Procede et appareil pour une detection dtmf et un melange vocal dans le domaine de parametres celp
WO2005078707A1 (fr) * 2004-02-16 2005-08-25 Koninklijke Philips Electronics N.V. Transcodeur et procede de transcodage a cet effet

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8396575B2 (en) 2009-08-14 2013-03-12 Dts Llc Object-oriented audio streaming system
US8396577B2 (en) 2009-08-14 2013-03-12 Dts Llc System for creating audio objects for streaming
US8396576B2 (en) 2009-08-14 2013-03-12 Dts Llc System for adaptively streaming audio objects
WO2011020067A1 (fr) * 2009-08-14 2011-02-17 Srs Labs, Inc. Système pour diffuser en continu de manière adaptative des objets audio
US9167346B2 (en) 2009-08-14 2015-10-20 Dts Llc Object-oriented audio streaming system
US9721575B2 (en) 2011-03-09 2017-08-01 Dts Llc System for dynamically creating and rendering audio objects
US9026450B2 (en) 2011-03-09 2015-05-05 Dts Llc System for dynamically creating and rendering audio objects
US9165558B2 (en) 2011-03-09 2015-10-20 Dts Llc System for dynamically creating and rendering audio objects
US10244343B2 (en) 2011-07-01 2019-03-26 Dolby Laboratories Licensing Corporation System and tools for enhanced 3D audio authoring and rendering
TWI607654B (zh) * 2011-07-01 2017-12-01 杜比實驗室特許公司 用於增強3d音頻編輯與呈現之設備、方法及非暫態媒體
US9838826B2 (en) 2011-07-01 2017-12-05 Dolby Laboratories Licensing Corporation System and tools for enhanced 3D audio authoring and rendering
US10609506B2 (en) 2011-07-01 2020-03-31 Dolby Laboratories Licensing Corporation System and tools for enhanced 3D audio authoring and rendering
US11057731B2 (en) 2011-07-01 2021-07-06 Dolby Laboratories Licensing Corporation System and tools for enhanced 3D audio authoring and rendering
US11641562B2 (en) 2011-07-01 2023-05-02 Dolby Laboratories Licensing Corporation System and tools for enhanced 3D audio authoring and rendering
US12047768B2 (en) 2011-07-01 2024-07-23 Dolby Laboratories Licensing Corporation System and tools for enhanced 3D audio authoring and rendering
US9613660B2 (en) 2013-04-05 2017-04-04 Dts, Inc. Layered audio reconstruction system
US9558785B2 (en) 2013-04-05 2017-01-31 Dts, Inc. Layered audio coding and transmission
US9837123B2 (en) 2013-04-05 2017-12-05 Dts, Inc. Layered audio reconstruction system
US9706324B2 (en) 2013-05-17 2017-07-11 Nokia Technologies Oy Spatial object oriented audio apparatus
WO2014184618A1 (fr) * 2013-05-17 2014-11-20 Nokia Corporation Appareil audio orienté objet spatial

Also Published As

Publication number Publication date
TW200921643A (en) 2009-05-16

Similar Documents

Publication Publication Date Title
JP5934922B2 (ja) 復号装置
JP6538128B2 (ja) オーディオ・オブジェクトを含むオーディオ・シーンの効率的な符号化
TWI629681B (zh) 使用頻譜域重新取樣來編碼或解碼多通道信號之裝置、方法及相關電腦程式
AU2005280041B2 (en) Multichannel decorrelation in spatial audio coding
KR101424752B1 (ko) 공간적 출력 다채널 오디오 신호를 결정하기 위한 장치
JP5645951B2 (ja) ダウンミックス信号表現に基づくアップミックス信号を提供する装置、マルチチャネルオーディオ信号を表しているビットストリームを提供する装置、方法、コンピュータプログラム、および線形結合パラメータを使用してマルチチャネルオーディオ信号を表しているビットストリーム
EP1851997B1 (fr) Systeme de codage/decodage multicanal transparent ou presque transparent
AU2006340728B2 (en) Enhanced method for signal shaping in multi-channel audio reconstruction
US8817992B2 (en) Multichannel audio coder and decoder
KR101104578B1 (ko) 비상관기, 그 비상관기를 이용하는 오디오 디코더, 오디오 입력신호에 근거한 출력신호의 발생방법 및 기록매체
JP6134867B2 (ja) レンダラ制御式空間アップミックス
TW200926147A (en) Audio coding using downmix
CN101401455A (zh) 使用子带滤波器的立体声呈现技术
EP1971979A1 (fr) Décodage de signaux audio binauraux
DK2171712T3 (en) A method and device for improving spatial audio signals
WO2009001292A1 (fr) Procédé de fusion d'au moins deux trains de paramètres audio orientés objet d'entrée en un train de paramètres audio orientés objet de sortie
Purnhagen et al. Immersive audio delivery using joint object coding
JP4976304B2 (ja) 音響信号処理装置、音響信号処理方法およびプログラム
WO2007080225A1 (fr) Décodage de signaux audio binauraux
Valero et al. A new parametric stereo and multichannel extension for mpeg-4 enhanced low delay aac (aac-eld)
EP4621772A2 (fr) Traitement d'audio à codage paramétrique
Yu et al. Low-complexity binaural decoding using time/frequency domain HRTF equalization
MX2008009565A (en) Apparatus and method for encoding/decoding signal

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08776462

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 08776462

Country of ref document: EP

Kind code of ref document: A1