Signal Processing Concepts and Engineering Insights. 


Explore signal processing concepts, algorithm comparisons, and practical engineering insights.
Topics include FFT vs STFT, FRF analysis, filtering techniques, and other signal processing methods used in real engineering workflows. 

Understanding Audio File Parameters: From Sample Rate to Codecs and Beyond

Introduction

Digital audio files are often described using a wide range of technical parameters: sample rate, bit depth, bitrate, codec, and many others. While these terms are commonly listed together, they represent fundamentally different aspects of how audio is captured, stored, compressed, and reproduced.

Understanding these parameters is not just about memorizing definitions. It is about recognizing how they interact to determine:

  • Signal fidelity
  • Storage efficiency
  • Perceptual quality
  • System compatibility

This article provides a structured and in-depth explanation of the key parameters that define digital audio.


Core Signal Representation Parameters

These define how the original analog signal is digitized.

Sample Rate (Sampling Frequency)

The sample rate defines how many samples per second are taken from a continuous-time signal.

Nyquist criterion

(Nyquist criterion)


Why it matters

  • Determines maximum representable frequency
  • Controls time resolution


Typical values

Use Case
Sample Rate
CD Audio
44.1 kHz
Video / Pro Audio
48 kHz
High-resolution audio
96 kHz / 192 kHz


Bit Depth

Bit depth defines the number of discrete amplitude levels.

Bit Depth


Why it matters

  • Controls quantization noise
  • Determines dynamic range


Where does 6.02 come from?

Dynamic range is defined as the ratio between the maximum signal level and the quantization noise floor.

For an N-bit system, the number of quantization levels is 2N.


The dynamic range in decibels is:

20 log10(2N) = 20 × N × log10(2)


Since log10(2) ≈ 0.301

20 × 0.301 ≈ 6.02

Thus

Dynamic Range ≈ 6.02N dB


Typical values

Use Case
Bit Depth
Dynamic Range
Consumer audio
16-bit
16 × 6.02 ≈ 96 dB
Studio recording
24-bit
24 × 6.02 ≈ 144 dB
Internal DSP
32-bit float
32 × 6.02 ≈ 193 dB


More precise

When quantization noise is modeled as a uniform random signal, the actual dynamic range becomes:

Dynamic Range ≈ 6.02N + 1.76 dB

The additional 1.76 dB accounts for the statistical distribution of quantization noise.


Channels

Channels represent independent audio streams.


Common configurations

Type
Description
Mono
Single channel
Stereo
Left + Right
5.1 / 7.1
Surround sound


Why it matters

  • Spatial perception
  • Directionality
  • Immersion


Duration

The total length of the audio signal.


Why it matters

  • Directly affects file size
  • Defines time-domain coverage


Data Rate and Storage Parameters

Digital audio storage behaves very differently depending on whether the signal is stored as raw PCM data or as a compressed format.

For this reason, it is essential to distinguish between parameters that apply to:

  • Uncompressed audio (e.g., WAV)
  • Compressed audio (e.g., MP3, AAC)


Bitrate (Compressed Audio Context)

Bitrate refers to the amount of data used per second:

Bitrate


Important Clarification

Bitrate (e.g., 128 kbps, 320 kbps) is a quantity, not a mode.


Bitrate Control Modes

CBR, VBR, and ABR are not different types of bitrate—they are encoding strategies:

Mode
Description
CBR (Constant Bitrate)
Fixed data rate throughout
VBR (Variable Bitrate)
Allocates bits dynamically based on complexity
ABR (Average Bitrate)
Targets an average bitrate over time


Key Insight

These modes are primarily used in lossy compression codecs such as MP3, AAC, and Opus.


PCM Data Rate (Uncompressed Audio)

For uncompressed PCM audio (e.g., WAV), data rate is determined directly by signal parameters:

Data Rate

where:

  • fs : sample rate
  • N : bit depth
  • C : number of channels


Example

44.1 kHz, 16-bit, stereo:

Data Rate example

(≈ 1.4 Mbps)


Byte Rate (PCM-Specific)

Byte rate is the number of bytes per second:

Byte Rate


Where It Applies

  • Primarily used in WAV / PCM formats
  • Defined explicitly in file headers


Why It Matters

  • Playback timing
  • Buffering
  • Streaming raw PCM data


Block Align (PCM-Specific)

Block align defines the number of bytes per sample frame:

Block Align


Interpretation

  • One “frame” = all channel samples at a given time instant
  • Ensures correct decoding alignment


Key Point

Block align is meaningful only for interleaved PCM data


Why These Concepts Do Not Apply to Compressed Audio

In compressed formats:

  • Data is not stored as fixed-size samples
  • Frames vary in size
  • Bit allocation is dynamic


Therefore:

Parameter
PCM
Compressed
Sample Rate


Bit Depth

❌ (implicit)
Byte Rate


Block Align


Bitrate



Compression and Encoding Parameters

Codec

A codec defines how audio is compressed and reconstructed.


Two Major Categories

Type
Examples
Lossless
FLAC, ALAC
Lossy
MP3, AAC, Opus


Key Role

  • Determines compression efficiency
  • Defines perceptual trade-offs


Container Format

A container defines how encoded data is stored.


Important Distinction

A container does not define how audio is compressed.


Examples

Container
Codec Inside
WAV
PCM
MP4
AAC
MKV
Multiple


Compression Ratio

Compression Ratio


Typical Values

Format
Ratio
FLAC
~2:1
MP3
~10:1


Encoder

The encoder is the implementation of a codec.


Why It Matters

Even with the same codec:

  • Different encoders produce different results
  • Psychoacoustic models vary


Example

  • LAME MP3 encoder vs older encoders


Perceptual and Signal Quality Parameters

Audio Quality

Not a single measurable value.


Depends on:

  • Bitrate
  • Codec
  • Psychoacoustic model
  • Listening environment


Bandwidth

The frequency range represented.


Example

System
Bandwidth
Telephone
~300 - 3400 Hz
CD Audio
~20 Hz - 22 kHz (Audible Frequency)


Dynamic Range

Difference between minimum and maximum signal levels.


Why it matters

  • Determines expressive detail
  • Affects perceived loudness


Data Representation Details

Endianness

Defines byte order.


Type

Type
Description
Little-endian
Least significant byte first
Big-endian
Most significant byte first


Why it matters

  • Cross-platform compatibility
  • Binary interpretation


PCM vs Floating Point
Type
Description
PCM
Integer representation
Float
Higher precision, avoids clipping


Additional Advanced Parameters

Quantization Type
  • Uniform
  • Non-uniform (μ-law, A-law)


Dithering
  • Randomizes quantization error
  • Reduces distortion


Noise Shaping
  • Moves noise to less audible frequencies


Interleaving
  • Multi-channel data arrangement


Frame Size
  • Defines chunk size in compressed formats


Latency
  • Delay introduced by processing


Jitter
  • Timing variation in sampling


Psychoacoustic Model
  • Used in lossy codecs
  • Removes inaudible components


How Everything Connects


Analog Signal
↓ Sampling (Sample Rate)
↓ Quantization (Bit Depth)
↓ Encoding (Codec → compression)
↓ Packaging (Container → file format)
↓ Storage / Transmission (Bitrate involved) 

Audio file


Key relationship

Parameter
Role
Sample Rate
Maximum representable frequency range
Bit Depth
Amplitude resolution
Bitrate
Data rate controlling compression quality
Codec
Compression algorithm (lossy or lossless)
Container
File format encapsulating encoded audio streams and metadata


Practical Engineering Insight

Audio Production

  • High sample rate → improved editing flexibility (e.g., time-stretching, pitch shifting)
  • High bit depth → increased headroom and reduced quantization noise


Streaming

  • Low bitrate → reduced bandwidth usage and faster transmission
  • High compression → potential quality degradation (depending on codec efficiency)


Real-time Systems

  • Low latency is critical
  • Block size affects latency and computational load


Key Insight

Digital audio is not defined by a single parameter.
It is the result of multiple interacting design choices across sampling, quantization, compression, and representation.

Trade-offs are unavoidable: improving one parameter often impacts others.


Suggested Further Reading

You may also find these topics helpful:


Infographic explaining digital audio file parameters including sample rate, bit depth, bitrate, codec, compression, and signal processing workflow from analog to encoded audio