This page provides resources and complementary results for the research article:

"The ERBlet Transform: An Auditory-Based Time-Frequency Representation with Perfect Reconstruction"

T. Necciari, P. Balazs, N. Holighaus, and P. Søndergaard

presented at the 38th International Conference on Acoustics, Speech, and Signal Processing (ICASSP2013). A PDF version of the article is available here for download.

Abstract: This paper describes a method for obtaining a perceptually motivated and perfectly invertible time-frequency representation of a sound signal. Based on frame theory and the recent non-stationary Gabor transform, a linear representation with resolution evolving across frequency is formulated and implemented as a non-uniform filterbank. To match the human auditory time-frequency resolution, the transform uses Gaussian windows equidistantly spaced on the psychoacoustic "ERB" frequency scale. Additionally, the transform features adaptable resolution and redundancy. Simulations showed that perfect reconstruction can be achieved using fast iterative methods and preconditioning even using one filter per ERB and a very low redundancy (1.08). Comparison with a linear gammatone filterbank showed that the ERBlet approximates well the auditory time-frequency resolution.

  • Complementary results:
    • ERBlet windows representation: K = 35 ERBlet filters computed for the frequency range 0-8 kHz using V = 1 filter/ERB.
ERBlet analysis windows  ERBlet dual windows
    • Comparison between ERBlet and other representations. Simulations were performed on a 5-sec musical excerpt from the band Manowar (song "Heart of Steel", studio version) in mono format, sampled at 44.1~kHz, 16~bits/sample. All analyses considered the frequency band 0-22.05~kHz.
ERBlet transform Standard Gabor transform
 redundancy = 12, relative reconstruction error < 10-15.  redundancy = 11.80, relative reconstruction error < 10-15.
 ERBlet transform  Constant-Q transform
 redundancy = 12, relative reconstruction error < 10-15.

redundancy = 12, relative reconstruction error < 10-15.
Implementation in [1].

 ERBlet transform  Linear gammatone filterbank
redundancy = 12, relative reconstruction error < 10-15.

redundancy = 128, relative reconstruction error = 1.4 for a delay of 4 ms and no post-processing correction of the filterbank delay. Accounting for the filterbank delay at the output of the re-synthesizer module led to relative reconstruction errors of 4.11 x 10-1, 1.01 x 10-1 and 2.86 x 10-3 for delays of 4, 8 and 16 ms, respectively. Implementation in [2].

  • Algorithms (pseudo-code):
Iterative reconstruction algorithm
 NSGT analysis algorithm
 NSGT synthesis algorithm
  • Matlab/Octave scripts available for download. Archive content:
    • Scripts for computing the ERBlet transform and its inverse. Includes the iterative reconstruction using the conjugate gradients method (Algorithm 1 above).
    • Scripts for generating the figures 1 and 2 presented in the manuscript.

IMPORTANT NOTE: The Matlab/Octave toolboxes Linear Time-Frequency Analysis (LTFAT, version 1.2.0 and above) [3] and Auditory Modeling (AM) must be installed to run the ERBlet codes. These toolboxes are freely available at Sourceforge.

  • References:
    • [1] G. A. Velasco, N. Holighaus, M. Dörfler, and T. Grill,"Constructing an invertible constant-Q transform with nonstationary Gabor frames", in Proceedings of the14th International Conference on Digital Audio Effects (DAFx-11), Paris, France, September 19-23 2011, pp.93–99.
    • [2] V. Hohmann, "Frequency analysis and synthesis using a gammatone filterbank", Acta Acust. united Ac., vol. 88, no. 3, pp. 433–442, 2002.
    • [3] P. L. Søndergaard, B. Torrésani, and P. Balazs, "The linear time-frequency analysis toolbox", Int. J. Wavelets. Multi., vol. 10, no. 4, pp. 1250032, July 2012.

Upcoming Events

ARI-Guest-Talk: Imitation of novel conspecific and human speech sounds in the killer whale (Orcinus orca)

28. Jänner 2019

Imitation of novel conspecific and human speech sounds in the killer whale (Orcinus orca) - José Francisco Zamorano Abramson

14.00 o'clock,

Seminar Room, Wohllebengasse 12-14 / Ground Floor

Read more ...

ARI-Guest-Talk: Non-verbal vocalisations in conversational speech

29. Jänner 2019

Non-verbal vocalisations in conversational speech - Jürgen Trouvain

14.30 o'clock,

Seminar Room, Wohllebengasse 12-14 / Ground Floor

Read more ...

Applied Harmonic Analysis and Friends


June 1st - 5th 2020



Read more ...