Hi,

I have a requirement to do a re-sampling based on drift detected on jitter buffer. 
Suppose I have set jack process callback to 128 samples, and I detected that I have to do
an resampling by 10%, ie n_in = 1.1 * nout then I need 12.8 samples more, which is not available and
has to wait for next callback. This will delay my processing. If I make my buffer size
more than 128, I may have to do a decimation based on drift in which direction.

How to handle these cases of having   varying buffer size requirements ?

-ben