COPPER

A PS Audio Publication

Issue 22 • Free Online Magazine

Issue 22 QUIBBLES AND BITS

Purer, More Perfect Sound

I have spent a lot of time focusing on the Red Book standard of CD Audio, which calls for 16-bit data at a 44.1kHz sample rate.  In doing so, I imagine I  have given implicit support to the commonly-held notion that this standard is perfectly adequate for use in high-end audio applications.  After all, this is what Sony and Phillips promised us back in 1982: “Pure, Perfect Sound – Forever”.  While the ‘forever’ part referred to the actual CD disk itself, the ‘pure, perfect sound’ part was clearly attributed to their new digital audio standard.

But if this really is ‘pure and perfect’ as the promise grandly claims, why, then, do we have all these Hi-Rez audio formats floating around?  Why do we need 24-bit formats if 16-bit is pure and perfect? Why all these high sample rates if what’s-his-name’s theory says 44.1kHz is plenty high enough?  Why bother with DSD and its phenomenally complicated underpinnings?  These are good questions, and need to be addressed.

The argument for the superiority of 24-bit over 16-bit is perhaps the least contentious and best understood aspect.  Simply put, every extra bit of bit depth allows us to push the noise floor down by an additional 6dB.  So the 8 extra bits offered by a 24-bit format translates to a whopping 48dB reduction in the noise floor.  Very roughly speaking, the very best analog electronics, microphones, and magnetic tape, all have a noise floor which is of the order of 20dB above that of 24-bit audio, and 20dB below that of 16-bit audio.  In simple terms, this means that 16-bit audio cannot come close to capturing the full dynamic range of the best analog signals you might ask it to encode, but 24-bit audio has it handily covered with plenty of margin in hand.

So, in areas of noise floor and dynamic range, going from 16-bit to 24-bit buys us a great deal of potential benefit at a cost of only 50% in additional file size.  I say ‘potential’ benefit, because you do require seriously high quality source material in order to take advantage of those benefits.  But if you’re reading this, the chances are you have more than just a passing interest in ‘seriously high quality source material’!

Moving on to the sample rate.  There are two immediately obvious benefits to increasing the sample rate.  First of all, there is actually a further reduction in the noise floor to be had.  Without going into all the ifs and buts, this amounts to an additional 3dB of noise floor reduction for every doubling of the sample rate.  However, since moving to 24-bit sampling has bought us all the noise floor reduction we can handle, this aspect of increased sample rate is not really very exciting.  The second obvious benefit is more nuanced – the fact that by increasing the sampling rate we can also increase the frequency range that we can capture.  This bears examination in more detail.

Suppose we double the sample rate from 44.1kHz to 88.2kHz, and thereby double the Nyquist frequency from 22.05kHz to 44.1kHz.  This means we can now capture a range of audio frequencies that extend quite a way above the nominal upper limit of audibility.  This is where things start to get contentious.  The main issue is that the vast majority of scientific study tends to support the general notion that as human beings we don’t hear any frequencies above 20kHz.  Furthermore, that only applies to what Vinny Gambini memorably referred to as “yoots”.  Once you get to my age, you’ll be lucky if you can still make it to 15kHz.  However, the topic of frequency audibility does throw up some occasional odd results.

For example, in a scientific paper from 2000, Tsutomu Oohashi et al reported that by measuring a number of subjects’ EEG and PET scan responses, they showed that when played Gamelan music (known to be rich in ultrasonic harmonics) their brains did in fact respond to frequencies above 22kHz.  The problem is that this paper’s results – referred to as the ‘Hypersonic Effect’ – have not been satisfactorily replicated, and in addition some valid technical criticisms have been made of its methodology.  Even the authors themselves showed no apparent interest in following up on their work.

Overall, this, like other arguments of its ilk, doesn’t appear to provide any plausible basis upon which to build a case in favor of high sample rate PCM formats.

Various other studies have looked at ways of assessing whether or not the human ear/brain is capable of resolving sonic effects due to relative timing factors – such as the discernability of time misalignment between signals from spatially displaced speakers (Milind Kunchur, 2007) – and a reasonably consistent picture emerges that we are sensitive to time alignments in the range 1-10μs.  This has been argued to imply that our brains are able to process audio signals with a bandwidth approaching 100kHz.  This is interesting because it suggests (in a roundabout kind of way) that we may be able to perceive certain effects of hypersonic harmonics of audible frequencies, without necessarily being able to detect those harmonics in isolation (i.e. Oohashi’s idea redux).

Any treatment of timing-related issues opens the door to the discussion of phase, since phase and timing are one and the same thing.  A phase error applied to a particular frequency simply means that that frequency is delayed (or advanced) in time.  For example, a time delay of 10μs is exactly the same as a phase retard of 0.628 radians for a 10kHz frequency.  Interesting things happen when you start to apply different time delays (i.e. different phase errors) to the different individual frequencies within a waveform.  What happens is that the frequency response doesn’t change per se, but the shape of the waveform does.  So the question arises – are phase errors which change the shape of a waveform, but not its frequency content, audible?  It has taken a while, but a consensus seems to be emerging that yes, they are.

The connection between the audibility of phase and the improved fidelity of high sample-rate PCM recording is a slightly tortuous one, but I believe it is very important.  Recall that Nyquist/Shannon sampling theory requires the very simple assumption that the signal is band-limited to the Nyquist Frequency.  Therefore, in any practical implementation, a signal has to be passed through a low-pass filter before it can be digitally sampled.  I’m running short of room in this column, but basically, the lower the sampling rate, the more aggressive this low-pass ‘anti-aliasing’ filter needs to be, because the Nyquist Frequency gets closer and closer to the upper limit of the audio band.  With 44.1kHz sampling, the two are very close indeed.

As a pretty good rule of thumb, the more aggressive the filter, the greater the phase (and other) errors it introduces.  These errors are at their worst at the frequencies above the top of the audio band where the filter is doing most of its work, but the residual phase errors do leak down into the audio band itself.  My contention is that the ‘sound’ of PCM audio – to the extent that it adds a distinct sound signature of its own – is the ‘sound’ of the band-limiting filters that the signal must go through before being digitally encoded.  And I attribute a significant element of that ‘sound signature’ – if not the whole ball of wax – to phase errors.  By progressively increasing the sample rate, it allows you to use progressively more gentle anti-aliasing filters, whose phase errors may be less and less sonically intrusive.

Interestingly, with DSD, because the sample rate is so high, we can consider it to have effectively a linear phase response across the entire audio band and into the hypersonic range.  Indeed, this may well be a significant reason why its adherents prefer its sound over high sample-rate PCM.

I believe that end-to-end (i.e. Mic-to-Speaker) phase linearity is going to be the next major advance in digital audio technology, one which we will see flower during the next decade.  In fact, it may be that this is at least one of the precepts behind the nascent MQA technology.

More from Issue 22

View All Articles in Issue 22

Search Copper Magazine

#227 Seth Lewis Gets in the Groove With Take a Look Around: a Tribute to the Meters by Frank Doris Feb 02, 2026 #227 Passport to Sound: May Anwar’s Audio Learning Experience for Young People by Frank Doris Feb 02, 2026 #227 Conjectures on Cosmic Consciousness by B. Jan Montana Feb 02, 2026 #227 The Big Takeover Turns 45 by Wayne Robins Feb 02, 2026 #227 Music and Chocolate: On the Sensory Connection by Joe Caplan Feb 02, 2026 #227 Singer/Songwriter Chris Berardo: Getting Wilder All the Time by Ray Chelstowski Feb 02, 2026 #227 The Earliest Stars of Country Music, Part One by Jeff Weiner Feb 02, 2026 #227 The Vinyl Beat Goes Down to Tijuana (By Way of Los Angeles), Part Two by Rudy Radelic Feb 02, 2026 #227 How to Play in a Rock Band, 20: On the Road With Blood, Sweat & Tears’ Guitarist Gabe Cummins by Frank Doris Feb 02, 2026 #227 From The Audiophile’s Guide: Audio Specs and Measuring by Paul McGowan Feb 02, 2026 #227 Our Brain is Always Listening by Peter Trübner Feb 02, 2026 #227 PS Audio in the News by PS Audio Staff Feb 02, 2026 #227 The Listening Chair: Sleek Style and Sound From the Luxman L3 by Howard Kneller Feb 02, 2026 #227 The Los Angeles and Orange County Audio Society Celebrates Its 32nd Anniversary, Honoring David and Sheryl Lee Wilson and Bernie Grundman by Harris Fogel Feb 02, 2026 #227 Back to My Reel-to-Reel Roots, Part 26: Half Full – Not Half Empty, Redux by Ken Kessler Feb 02, 2026 #227 That's What Puzzles Us... by Frank Doris Feb 02, 2026 #227 Record-Breaking by Peter Xeni Feb 02, 2026 #227 The Long and Winding Road by B. Jan Montana Feb 02, 2026 #226 JJ Murphy’s Sleep Paralysis is a Genre-Bending Musical Journey Through Jazz, Fusion and More by Frank Doris Jan 05, 2026 #226 Stewardship by Consent by B. Jan Montana Jan 05, 2026 #226 Food, Music, and Sensory Experience: An Interview With Professor Jonathan Zearfoss of the Culinary Institute of America by Joe Caplan Jan 05, 2026 #226 Studio Confidential: A Who’s Who of Recording Engineers Tell Their Stories by Frank Doris Jan 05, 2026 #226 Pilot Radio is Reborn, 50 Years Later: Talking With CEO Barak Epstein by Frank Doris Jan 05, 2026 #226 The Vinyl Beat Goes Down to Tijuana (By Way of Los Angeles), Part One by Rudy Radelic Jan 05, 2026 #226 Capital Audiofest 2025: Must-See Stereo, Part Two by Frank Doris Jan 05, 2026 #226 My Morning Jacket’s Carl Broemel and Tyler Ramsey Collaborate on Their Acoustic Guitar Album, Celestun by Ray Chelstowski Jan 05, 2026 #226 The People Who Make Audio Happen: CanJam SoCal 2025, Part Two by Harris Fogel Jan 05, 2026 #226 How to Play in a Rock Band, 19: Touring Can Make You Crazy, Part One by Frank Doris Jan 05, 2026 #226 Linda Ronstadt Goes Bigger by Wayne Robins Jan 05, 2026 #226 From The Audiophile’s Guide: Active Room Correction and Digital Signal Processing by Paul McGowan Jan 05, 2026 #226 PS Audio in the News by Frank Doris Jan 05, 2026 #226 Back to My Reel-to-Reel Roots, Part 25: Half-Full, Not Empty by Ken Kessler Jan 05, 2026 #226 Happy New Year! by Frank Doris Jan 05, 2026 #226 Turn It Down! by Peter Xeni Jan 05, 2026 #226 Ghost Riders by James Schrimpf Jan 05, 2026 #226 A Factory Tour of Audio Manufacturer German Physiks by Markus "Marsu" Manthey Jan 04, 2026 #225 Capital Audiofest 2025: Must-See Stereo, Part One by Frank Doris Dec 01, 2025 #225 Otis Taylor and the Electrics Delivers a Powerful Set of Hypnotic Modern Blues by Frank Doris Dec 01, 2025 #225 A Christmas Miracle by B. Jan Montana Dec 01, 2025 #225 T.H.E. Show New York 2025, Part Two: Plenty to See, Hear, and Enjoy by Frank Doris Dec 01, 2025 #225 Underappreciated Artists, Part One: Martin Briley by Rich Isaacs Dec 01, 2025 #225 Rock and Roll is Here to Stay by Wayne Robins Dec 01, 2025 #225 A Lifetime of Holiday Record (and CD) Listening by Rudy Radelic Dec 01, 2025 #225 Little Feat: Not Saying Goodbye, Not Yet by Ray Chelstowski Dec 01, 2025 #225 How to Play in a Rock Band, Part 18: Dealing With Burnout by Frank Doris Dec 01, 2025 #225 The People Who Make Audio Happen: CanJam SoCal 2025 by Harris Fogel Dec 01, 2025 #225 Chicago’s Sonic Sanctuaries: Four Hi‑Fi Listening Bars Channeling the Jazz‑Kissa Spirit by Olivier Meunier-Plante Dec 01, 2025

Purer, More Perfect Sound

I have spent a lot of time focusing on the Red Book standard of CD Audio, which calls for 16-bit data at a 44.1kHz sample rate.  In doing so, I imagine I  have given implicit support to the commonly-held notion that this standard is perfectly adequate for use in high-end audio applications.  After all, this is what Sony and Phillips promised us back in 1982: “Pure, Perfect Sound – Forever”.  While the ‘forever’ part referred to the actual CD disk itself, the ‘pure, perfect sound’ part was clearly attributed to their new digital audio standard.

But if this really is ‘pure and perfect’ as the promise grandly claims, why, then, do we have all these Hi-Rez audio formats floating around?  Why do we need 24-bit formats if 16-bit is pure and perfect? Why all these high sample rates if what’s-his-name’s theory says 44.1kHz is plenty high enough?  Why bother with DSD and its phenomenally complicated underpinnings?  These are good questions, and need to be addressed.

The argument for the superiority of 24-bit over 16-bit is perhaps the least contentious and best understood aspect.  Simply put, every extra bit of bit depth allows us to push the noise floor down by an additional 6dB.  So the 8 extra bits offered by a 24-bit format translates to a whopping 48dB reduction in the noise floor.  Very roughly speaking, the very best analog electronics, microphones, and magnetic tape, all have a noise floor which is of the order of 20dB above that of 24-bit audio, and 20dB below that of 16-bit audio.  In simple terms, this means that 16-bit audio cannot come close to capturing the full dynamic range of the best analog signals you might ask it to encode, but 24-bit audio has it handily covered with plenty of margin in hand.

So, in areas of noise floor and dynamic range, going from 16-bit to 24-bit buys us a great deal of potential benefit at a cost of only 50% in additional file size.  I say ‘potential’ benefit, because you do require seriously high quality source material in order to take advantage of those benefits.  But if you’re reading this, the chances are you have more than just a passing interest in ‘seriously high quality source material’!

Moving on to the sample rate.  There are two immediately obvious benefits to increasing the sample rate.  First of all, there is actually a further reduction in the noise floor to be had.  Without going into all the ifs and buts, this amounts to an additional 3dB of noise floor reduction for every doubling of the sample rate.  However, since moving to 24-bit sampling has bought us all the noise floor reduction we can handle, this aspect of increased sample rate is not really very exciting.  The second obvious benefit is more nuanced – the fact that by increasing the sampling rate we can also increase the frequency range that we can capture.  This bears examination in more detail.

Suppose we double the sample rate from 44.1kHz to 88.2kHz, and thereby double the Nyquist frequency from 22.05kHz to 44.1kHz.  This means we can now capture a range of audio frequencies that extend quite a way above the nominal upper limit of audibility.  This is where things start to get contentious.  The main issue is that the vast majority of scientific study tends to support the general notion that as human beings we don’t hear any frequencies above 20kHz.  Furthermore, that only applies to what Vinny Gambini memorably referred to as “yoots”.  Once you get to my age, you’ll be lucky if you can still make it to 15kHz.  However, the topic of frequency audibility does throw up some occasional odd results.

For example, in a scientific paper from 2000, Tsutomu Oohashi et al reported that by measuring a number of subjects’ EEG and PET scan responses, they showed that when played Gamelan music (known to be rich in ultrasonic harmonics) their brains did in fact respond to frequencies above 22kHz.  The problem is that this paper’s results – referred to as the ‘Hypersonic Effect’ – have not been satisfactorily replicated, and in addition some valid technical criticisms have been made of its methodology.  Even the authors themselves showed no apparent interest in following up on their work.

Overall, this, like other arguments of its ilk, doesn’t appear to provide any plausible basis upon which to build a case in favor of high sample rate PCM formats.

Various other studies have looked at ways of assessing whether or not the human ear/brain is capable of resolving sonic effects due to relative timing factors – such as the discernability of time misalignment between signals from spatially displaced speakers (Milind Kunchur, 2007) – and a reasonably consistent picture emerges that we are sensitive to time alignments in the range 1-10μs.  This has been argued to imply that our brains are able to process audio signals with a bandwidth approaching 100kHz.  This is interesting because it suggests (in a roundabout kind of way) that we may be able to perceive certain effects of hypersonic harmonics of audible frequencies, without necessarily being able to detect those harmonics in isolation (i.e. Oohashi’s idea redux).

Any treatment of timing-related issues opens the door to the discussion of phase, since phase and timing are one and the same thing.  A phase error applied to a particular frequency simply means that that frequency is delayed (or advanced) in time.  For example, a time delay of 10μs is exactly the same as a phase retard of 0.628 radians for a 10kHz frequency.  Interesting things happen when you start to apply different time delays (i.e. different phase errors) to the different individual frequencies within a waveform.  What happens is that the frequency response doesn’t change per se, but the shape of the waveform does.  So the question arises – are phase errors which change the shape of a waveform, but not its frequency content, audible?  It has taken a while, but a consensus seems to be emerging that yes, they are.

The connection between the audibility of phase and the improved fidelity of high sample-rate PCM recording is a slightly tortuous one, but I believe it is very important.  Recall that Nyquist/Shannon sampling theory requires the very simple assumption that the signal is band-limited to the Nyquist Frequency.  Therefore, in any practical implementation, a signal has to be passed through a low-pass filter before it can be digitally sampled.  I’m running short of room in this column, but basically, the lower the sampling rate, the more aggressive this low-pass ‘anti-aliasing’ filter needs to be, because the Nyquist Frequency gets closer and closer to the upper limit of the audio band.  With 44.1kHz sampling, the two are very close indeed.

As a pretty good rule of thumb, the more aggressive the filter, the greater the phase (and other) errors it introduces.  These errors are at their worst at the frequencies above the top of the audio band where the filter is doing most of its work, but the residual phase errors do leak down into the audio band itself.  My contention is that the ‘sound’ of PCM audio – to the extent that it adds a distinct sound signature of its own – is the ‘sound’ of the band-limiting filters that the signal must go through before being digitally encoded.  And I attribute a significant element of that ‘sound signature’ – if not the whole ball of wax – to phase errors.  By progressively increasing the sample rate, it allows you to use progressively more gentle anti-aliasing filters, whose phase errors may be less and less sonically intrusive.

Interestingly, with DSD, because the sample rate is so high, we can consider it to have effectively a linear phase response across the entire audio band and into the hypersonic range.  Indeed, this may well be a significant reason why its adherents prefer its sound over high sample-rate PCM.

I believe that end-to-end (i.e. Mic-to-Speaker) phase linearity is going to be the next major advance in digital audio technology, one which we will see flower during the next decade.  In fact, it may be that this is at least one of the precepts behind the nascent MQA technology.

0 comments

Leave a comment

0 Comments

Your avatar

Loading comments...

🗑️ Delete Comment

Enter moderator password to delete this comment:

✏️ Edit Comment

Enter your email to verify ownership: