I've used the FFT data from the Analyser node using the getByteFrequencyData method in the Web Audio API to create a spectrum visualizer as shown below:

Spectrum Visualizer

In this instance I have 256 bins of data. What exactly do the numbers in this correspond to? Is it the decibel level of each frequency component. If so how do I know what the value of the frequency of each bin corresponds to?

I would like to know so I can experiment in building a graphic eq and so would like to know at which points to indicate the filter bands. Ideally I'd like to represent frequencies from 20Hz to 20kHz and plot intervals between those accordingly.

Thanks for any help.

up vote 29 down vote accepted

yes,getByteFrequencyData results in a normalized array of values between 0 and 255. (it copies the data to the array it gets passed-in).

the frequency bands are split equally, so each element N of your array corresponds to:

N * samplerate/fftSize

so, the first bin is 0.
and, assuming a samplerate of 44100 and a <analyzerNode>.fftSize of 512 the second would be: 86.13 Hz, and so on...

you will find these two questions and answers useful, on dsp and on SO:

Note that the length of your sampledata is half the <analyzerNode>.fftSize, effectively limiting the frequency-range to half the samplerate.

  • getByteFrequencyData doesn't return anything; rather, it copies the frequency data values (0-255) into the unsigned array (Uint8Array() in javascript) that you pass into the method as a parameter. W3C Draft – Web Audio API – getByteFrequencyData. Just in case anyone gets caught trying to use some form of returned value from this method call... – Danny Bullis Mar 28 '13 at 18:31
  • @fettereddingoskidney, thanks, updated. – kr1 Mar 28 '13 at 18:47
  • 15
    BTW default sample rate may be different. E.g. for me it is 48000. One should check audioContext.sampleRate for certain value. – kirilloid Aug 25 '13 at 3:07
  • 3
    For anyone looking for more information, half of the sample rate is called the Nyquist Frequency. – Ian Hunter Jul 8 '14 at 14:05
  • 1
    0Hz is, indeed, the DC offset - so it's a number that if you subtract it from your original signal would make that signal zero-mean. It's not amplitude in the usual sense for for audio. – Arunas May 16 '17 at 22:01

With 256 bins, each one will be ~86 Hz apart (44100 kHz sample rate / fftSize, where fftSize is twice the number of bins). So you start at zero and go up in 86 Hz increments from there.

The actual values in the bins are just a representation of how much of each frequency is present in the signal (i.e. how "loud" the frequency is).

  • the web audio api returns a number of bins which is half the fftSize. so, IMHO you should divide by bins.length * 2 or better using fftSize directly, see also my answer. – kr1 Feb 9 '13 at 16:17

Your Answer

 

By clicking "Post Your Answer", you acknowledge that you have read our updated terms of service, privacy policy and cookie policy, and that your continued use of the website is subject to these policies.

Not the answer you're looking for? Browse other questions tagged or ask your own question.