What does the FFT data in the Web Audio API correspond to?

11,574

Solution 1

yes,getByteFrequencyData results in a normalized array of values between 0 and 255. (it copies the data to the array it gets passed-in).

the frequency bands are split equally, so each element N of your array corresponds to:

N * samplerate/fftSize

so, the first bin is 0.
and, assuming a samplerate of 44100 and a <analyzerNode>.fftSize of 512 the second would be: 86.13 Hz, and so on...

you will find these two questions and answers useful, on dsp and on SO:

Note that the length of your sampledata is half the <analyzerNode>.fftSize, effectively limiting the frequency-range to half the samplerate.

Solution 2

With 256 bins, each one will be ~86 Hz apart (44100 kHz sample rate / fftSize, where fftSize is twice the number of bins). So you start at zero and go up in 86 Hz increments from there.

The actual values in the bins are just a representation of how much of each frequency is present in the signal (i.e. how "loud" the frequency is).

Solution 3

TL;DR I suppose it producing dBu


if anyone, just as me not being an audio engineer, struggling trying to figure out a proper scale (what it actually means) for a results from getByteFrequencyData/getFloatFrequencyData here is my findings:

  • first check out this web audio spec there we can find an equation to go from a frequency data to dB value itself

Y[k]=20log10X^[k]

and see that dBu computation looks quite similar dBu

  • previously I've watched this video to understand the basics

  • so build a scale to display dB as dBu with zero lying at -4dBu (I hope this make sense) dB scale

  • and now attempt to validate this scale with a proper stream. so I could produce -24 dBFS (-18dBFS = +4dBu this is suppose to be a standard?) signal which is should be at -18 + -24 + 4 - -4 = -34 at scale above. and this is what actually I'm getting in the end enter image description here

Share:
11,574
RobotEyes
Author by

RobotEyes

Updated on July 21, 2022

Comments

  • RobotEyes
    RobotEyes almost 2 years

    I've used the FFT data from the Analyser node using the getByteFrequencyData method in the Web Audio API to create a spectrum visualizer as shown below:

    Spectrum Visualizer

    In this instance I have 256 bins of data. What exactly do the numbers in this correspond to? Is it the decibel level of each frequency component. If so how do I know what the value of the frequency of each bin corresponds to?

    I would like to know so I can experiment in building a graphic eq and so would like to know at which points to indicate the filter bands. Ideally I'd like to represent frequencies from 20Hz to 20kHz and plot intervals between those accordingly.

    Thanks for any help.