|
Post by Bob Olhsson on Mar 6, 2019 16:08:48 GMT -6
Jitter was a serious consideration back in the '90s because of incompetent chip design but it eventually got straightened out. The same people attempted to design digital TV studio gear and couldn't hide behind ABX tests when a director pointed at a glitch on the screen and wanted to know "what the f!@k is THAT?"
|
|
|
Post by cyrano on Mar 7, 2019 17:18:33 GMT -6
PS - in retrospect, thinking about it, perhaps the term "inter sample latency" is misleading. (I didn't coin the phrase) Obviously there IS no "between" samples in the DAW - everything is gridded to a discrete sample. What happens in real world hybrid mixing, is that once re-digitized, the analog affected incoming audio does not always (does not usually) "line up" with the leading edge of a transient that has been analog processed in a round the loop DA/AD fashion matching the leading edge of the unprocessed ITB transient. Hope that makes better sense. Thanks, drBill. That's what I suspected. When going back into the analog world and resampling, issues can occur. BUT, signals not lining up is a phase issue. Nothing to do with samples, or sampling. However, I now can see how someone would come up with the phrase "inter sample latency". At the very least, it should be "inter sampling latency". Or, better, "Resampling latency". Still, latency usually being in the millisecond range and phase issues being in the microsecond range, it seems very unfortunate wording. Besides, a Google search turns up exactly one manufacturer using it, combined with some other mumbo-jumbo. And the phrase does have a scientific use, in CPU performance testing.
|
|