What kind of evidence do you consider as sufficient?

Status
Not open for further replies.
To follow the example, one can not rule out that on some machines there exists an audible difference when bitwise identical files are replayed/routed from different hardware devices

Sometimes it is better to be more precise, if one really wants to sort it out.

You made my point for me, precise stating of all conditions is important. Different hardware can mean anything, the claim stated (in fact here just this week) is that the same bits played through exactly the same hardware chain have an audible difference. If you play the files through different DAC's or clock recovery process you are no longer testing a single hypothesis. You should understand by now the claims are made in many cases deliberately to say our basic knowledge of the physical world at the most fundamental level is wrong (at least for hi end audio).
 
Last edited:
This is all off the track, we first need a mutually agreed upon method to determine actual audibility and then we can test any claims. I mention the extraordinary claims only because using them IMO is a distraction and waste of time. Let's stick with various DAC implementations or something similar.
 
But that does not tell you, if the observed results exists due to the conditions of the test or because _these_ listener really can´t (means never) detect the difference.
Test condition causes (aurally) debilitating problem?

As stated before, we already know that the same test subjects get significantly _less_ correct trials when probing the same sensory difference in an ABX test than in an (for example) A/B test.
Stated but not proven, right?
 
Please explain how two bitwise identical files, played through exactly the same hardware, can sound different.

What kind of evidence do you consider as sufficient?

That´s the link to my post, for simplification i copy your assertion and my response:

Waly wrote:
From the same author, for whatever reason ("hidden variable") bitwise identical digital files can sound different, depending on the support they are stored on (e.g. thumb drive vs. hard drive).

my response:
one can not rule out that on some machines there exists an audible difference when bitwise identical files are replayed/routed from different hardware devices, but one can rule out that the reason is some kind of a "ghost in the machine" .

Obviously neither in your nor in my sentences there is "exactly the same hardware" mentioned, to the contrary we both wrote explicitely about _different_ hardware.


You are either trolling, or a Nobel prize is around the corner for you.

Or there is a third option that you´ve missed...... :cool:
 
You made my point for me, precise stating of all conditions is important.

So we do agree at this point.

Different hardware can mean anything, the claim stated (in fact here just this week) is that the same bits played through exactly the same hardware chain have an audible difference.

Seems that i´ve missed that statement so can´t comment.....

If you play the files through different DAC's or clock recovery process you are no longer testing a single hypothesis.

Even the same DAC but data delivery by different processes might be sufficient to evoke differences, that´s the reason why i/we often point to the distinction between claims of audibility and offered explanation for this claimed audibility.

You should understand by now the claims are made in many cases deliberately to say our basic knowledge of the physical world at the most fundamental level is wrong (at least for hi end audio).

I understand that, but to use always/mainly the most outrageous examples is a familiar usage of the eristics playbook (consciously or nonconsciously) and we have seen in this thread and the other recent threads about controlled listening tests, that even providing scientific evidence for some pitfalls of the commonly used tests (specific and in general) is blamed as "discrediting blind tests" .

So, as you´ve already suggested, it is much better to avoid these (often mainly bias driven) characterizations and to work out acceptable conditions for tests.
 
Your that directly above my that. I'm sorry you found that disturbing

Ah, no "that" is not all i´ve learned about....and yes, you could have saved me a lot of time if you had told me that 30 years ago (when we engineering students all developed a deep understanding of statistics without realizing it :) ) and even today it would help. (Jakob´s rule, a fact has to be repeated in audio forums at least 148 times before it slowly begins to sink in)

Further, never mind, there are a lot more mysterious "thats" disturbing me. :)
 
Please explain how two bitwise identical files, played through exactly the same hardware, can sound different.
Different packeting of USB data, for example. Of course the hardware must be "bad enough" for easy to measure effects of modulation of the clocks, supplies/references, etc, aka Logic Induced Distortion.

With synchronuous averaging and differential testing a lot of things can be isolated. Attached is a screenshot of an extreme example of this distortion. DUT was a lowly TASCAM US122-Mk2 bus-powered DAC/ADC which is low-fi enough to produce lots of this distortion. The setup was analog loopback by necessity (sample sync'd record while playback) so it's not clear if the DAC section is mainly responsible but actually it doesn't matter for the topic where at the DAC-->ADC conversion the disturbance occurs.

Stimulus was a LogSweep with a short pause with a 2^N length, repeated forever. Two recordings with 10000 averages were made which differed only in start offset of the sample buffers transmitted. Then condensed data blocks were subtracted after sample offset adjustment, with a level fine trim to get the lowest residual in the midrange frequencies (there is gain difference from slow reference drift as well less perfect nulling at the frequency extremes from slow clock drifts which cause small phase changes from all the analog lowpasses and high passes involved which spoils the nulling). Of course all calculation is done in 64bit-floating point, results were double checked and a lot of other experiments and null tests have been made to make sure the difference is real.

If there were no explicit distortion mechanism while the small gain and clock drifts would still be present the residual would not null perfectly as well, but the residual would be linear, only showing signs of random noise. As we can see, there is gross systematic distortion showing in the difference.

With the same technique I could isolate the microphonic/vibrational crosstalk into the device in a clarity that is breathtaking.

In the meantime I've refined the process so as to get rid of the influence of the gain and clock drift (which is always present even in state-of-the-art products) by interleaving the source data. Processing has also been optimized.
My best DAC/ADC, a RME ADI2-Pro FS, a much better device than the lowly TASCAM, now allows me to isolate effects to degree that keeps amazing me, both of the device itself as well as external DUT -- the loopback cable being the simplest DUT, isolating its microphony for example. Currently I'm in the process of designing and building an intelligent µC-controlled synchronous "switching box" so that I can actually switch something in hardware.
 

Attachments

  • USB_LogicInducedDistortion.gif
    USB_LogicInducedDistortion.gif
    24 KB · Views: 104
This seems to be going in circles. Is the answer to the OP not simply - as long as all the variables are controlled and understood, and there's statistical significance in the result, we'd all be ok with it?

Although the abstraction level seems appropriate to address the question, it was one aim to learn if beside the scientific requirements there are additional demands like "no financial interests involved" or "supervising" play a role.

Additionally the (still quite new especially in the audio field) demand for a preregistration - to avoid testing til finally some positive result appears which is then reported without mentioning the other attempts - or the qualification as "extraordinary" claim.

Usually we can be sure that it is impossible to control all variables, some can be controlled, the impact of others can often only be randomized and it depends on the objective if total control would be necessary or even desirable.
So on this level of abstraction i´d say the methodology has to follow best practice considering the objective of the experiment.

The question of statistical significance is always debatable as highly subjective, the usual 0.05 criterion should be sufficient as we would like to see replications anyway.
 
Different packeting of USB data, for example. Of course the hardware must be "bad enough" for easy to measure effects of modulation of the clocks, supplies/references, etc, aka Logic Induced Distortion.

With synchronuous averaging and differential testing a lot of things can be isolated. Attached is a screenshot of an extreme example of this distortion. DUT was a lowly TASCAM US122-Mk2 bus-powered DAC/ADC which is low-fi enough to produce lots of this distortion. The setup was analog loopback by necessity (sample sync'd record while playback) so it's not clear if the DAC section is mainly responsible but actually it doesn't matter for the topic where at the DAC-->ADC conversion the disturbance occurs.

Stimulus was a LogSweep with a short pause with a 2^N length, repeated forever. Two recordings with 10000 averages were made which differed only in start offset of the sample buffers transmitted. Then condensed data blocks were subtracted after sample offset adjustment, with a level fine trim to get the lowest residual in the midrange frequencies (there is gain difference from slow reference drift as well less perfect nulling at the frequency extremes from slow clock drifts which cause small phase changes from all the analog lowpasses and high passes involved which spoils the nulling). Of course all calculation is done in 64bit-floating point, results were double checked and a lot of other experiments and null tests have been made to make sure the difference is real.

If there were no explicit distortion mechanism while the small gain and clock drifts would still be present the residual would not null perfectly as well, but the residual would be linear, only showing signs of random noise. As we can see, there is gross systematic distortion showing in the difference.

With the same technique I could isolate the microphonic/vibrational crosstalk into the device in a clarity that is breathtaking.

In the meantime I've refined the process so as to get rid of the influence of the gain and clock drift (which is always present even in state-of-the-art products) by interleaving the source data. Processing has also been optimized.
My best DAC/ADC, a RME ADI2-Pro FS, a much better device than the lowly TASCAM, now allows me to isolate effects to degree that keeps amazing me, both of the device itself as well as external DUT -- the loopback cable being the simplest DUT, isolating its microphony for example. Currently I'm in the process of designing and building an intelligent µC-controlled synchronous "switching box" so that I can actually switch something in hardware.

Can’t follow, but USB, even in different versions, won’t packetize data different, leading to different outputs. Let me know next time you copy a piece of software through USB of whatever version and the file(s) gets corrupted in the process.
 
Although the abstraction level seems appropriate to address the question, it was one aim to learn if beside the scientific requirements there are additional demands like "no financial interests involved" or "supervising" play a role.

Additionally the (still quite new especially in the audio field) demand for a preregistration - to avoid testing til finally some positive result appears which is then reported without mentioning the other attempts - or the qualification as "extraordinary" claim.

Usually we can be sure that it is impossible to control all variables, some can be controlled, the impact of others can often only be randomized and it depends on the objective if total control would be necessary or even desirable.
So on this level of abstraction i´d say the methodology has to follow best practice considering the objective of the experiment.

The question of statistical significance is always debatable as highly subjective, the usual 0.05 criterion should be sufficient as we would like to see replications anyway.

As the person who made the pre-registration suggestion, I have no problems with a corporate sponsor although I'd be surprised to see any forthcoming (just to be clear, not saying you suggested I did). The p <0.05 wouldn't be my first suggestion as many fields are moving away from specific endpoints as "valid" since it creates an effect of chasing validity rather than pursuing knowledge. AIP holds onto its 6 sigma for discoveries but the efforts involved to get that are, well, CERN like in nature. :)

We can argue significance until the cows come home after the data is out :D
 
Status
Not open for further replies.