I am trying to understand how FFmpeg stores data in AVFrameafter the sound has been decoded.
Basically, if I print data in an array AVFrame->data[], I get several unsigned 8-bit integers that are audio in raw format.
From what I can understand from FFmpeg doxygen, the data format is expressed in enum AVSampleFormatand there are two main categories: alternating and flat. In an alternating type, data is stored in the first line of an array AVFrame->datawith a size AVFrame->linesize[0], while in a planar type, each channel of an audio file is stored in a separate line in the array AVFrame->data, and the arrays are sized AVFrame->linesize[0].
Is there a manual / tutorial that explains what the numbers in the array mean for each of the formats?
source
share