On Wed, Oct 04, 2006 at 10:51:08PM -0500, Andres Cabrera wrote:
I've written a paper analyzing the characteristics
of some software
dynamic range compressors. The interesting part concerning linux, is
that all my results show jaggedness on the gain reduction curves. I did
the tests using Ardour, Audacity and Rezound with the same results,
which points to something strange in some part of the process.
Or in your measurement methods which are ill-defined, making it all but
impossible to correctly interpret some of the results.
- How do you measure gain ? By comparing single input/output samples ?
It seems so, otherwise how do you obtain a gain vs. time curve at
50 Hz with sub-millisecond resolution (Fig. 3) ?
This will be invalid in many cases, for example if the plugin includes
audio delay to achieve 'zero latency', as you suggest some of them do.
- This delay is the first thing that should be measured. Without
this information it is impossible to evaluate the results.
- How on earth do can you define the level of a white noise signal
by a peak value ?
- What is a square wave at 0dB FS ? Positive and negative samples
at the maximum amplitude ? That does no correspond to a analog
square wave signal.
- How do you expect to measure distortion using square waves ?
--
FA
Lascia la spina, cogli la rosa.