Bert, I missed it because I do not have one or worked with one, which is not to say it is bad or anything, I just took the examples I recall because it is familiar to me. There is more of these for sure. I hope it was good enough to illustrate the points with some real-life examples.
Cheers, Magnus On 2020-02-21 13:45, ew via time-nuts wrote: > You missed my favorite HP5345A only direct 500 MHz and internal 500 MHz, > only recently replaced it with 53132A still use it for 40 GHz work > Bert Kehren > > > > In a message dated 2/21/2020 7:27:32 AM Eastern Standard Time, > mag...@rubidium.se writes: > > Hi Taka, > > On 2020-02-21 04:45, Taka Kamiya via time-nuts wrote: >> I was in electronics in big ways in 70s. Then had a long break and came >> back to it in last few years. Back then, if I wanted 1s resolution, the >> gate time had to be 1s. So measuring ns and ps was pretty much impossible. >> As I understand it, HP53132A (my main counter) takes thousands of samples (I >> assume t samples) to arrive at most likely real frequency. That was >> something I had hard time wrapping my head around. > It actually does two things. > > First, it interpolates the occurrence of a rising edge (for start and > stop channel), so if this does not happen in perfect alignment with a > rising edge of the reference/coarse clock. Often the OCXO/Rubidium is > for 10 MHz, but then a 90-500 MHz oscillator is locked to the > reference, and this higher clock is then used instead of the 10 MHz for > coarse-counting. Coarse-counting is counting of cycles just back in the > good old days of counters. The resolution is increased further not by > raising the counting frequency, but by measuring the time-error of the > trigger channel event in relation to the coarse-counter clock-edge. > Thus, measuring 0.000-0.999 of a coarse-counting cycle. In practice it > becomes hard to design for that, as the shorter end has problem is gate > delay times to be well decided, so one add one or two coarse cycles to > do 1.000-1.999 or 2.000-2.999 cycles, but these extra cycles is only for > the interpolator design, so once the fractional cycle is known the other > can be ignored. > > Just to give you an idea of what different counters do, here is from the > top of my head some numbers: > > HP5370A: Ref 10 MHz, Coarse 200 MHz, Interpolation gain 256, time > resolution < 20 ps > HP5328A: Ref 10 MHz, Coarse 10 MHz, Interpolation gain 1, time > resolution 100 ns > HP5328A with Option 040-042 and HP5328B: Ref 10 MHz, Coarse 100 MHZ, > Interpolation gain 1 (TI-average has other interpolation means), time > resolution 10 ns or for TI-avg 10 ps (claimed) > HP5335A: Ref 10 MHz, Coarse 10 MHz, Interpolation gain 200, time > resolution 1 ns > HP5372A: Ref 10 MHz, Coarse 500 MHz, Interpolation gain 10, time > resolution 200 ps > HP53132A: Ref 10 MHz, Coarse 100 MHz, Interpolation gain 1000, time > resolution 100 ps > SR620: Ref 10 MHz, Coarse 90 MHz, Interpolation gain 512?, time > resolution < 25 ps (don't recall details) > PM6863: Ref 10 MHz, Coarse 500 MHz, Interpolation gain 1, time > resolution 2 ns > CNT-90: Ref 10 MHz, Coarse 100 MHz, Interpolation gain 512, time > resolution 100 ps (claimed) > CNT-91: Ref 10 MHz, Coarse 100 MHz, Interpolation gain 512, time > resolutions 50 ps (claimed) > SIA3000: Ref 100 MHz, Coarse, 100 MHz, Interpolation gain 50000, time > resolution 200 fs > > As I write claimed above, the actual performance can be better, but the > spec on the sheet did not overstate it more. While all the numbers may > not be 100% correct, I think they help to illustrate the relationships > very well. As you calculate the length of the coarse counter period from > it's frequency, and then divide with the interpolation gain, which is by > how many steps the period is interpolated, the raw time resolutions pops > out. > > Interpolation methods differs, but typically first an error signal is > generated and then it is stored into a capacitor which is then measured > with some slower technique. The 5335A use a very simple technique where > the discharge of the capacitor is done with a much lower current than > the charging, so now the discharge time can be measured using the coarse > clock. This is called pulse-stretching. Today the far most common > technique is to use an ADC to digitize the voltage. > > The 5328 counters have a unique interpolation technique by > phase-modulating the reference clock with noise, effectively shifting > the reference transitions around and that way interpolate over time a > higher resolution. It works better than claimed. > > Remember that this single-shot resolution is reduced by the trigger > jitter as well as unstability of reference oscillator. In practice the > trigger jitter or resolution dominates as a 1/tau limit as you look at > the Allan Deviation, to fix that you need to buy a better counter or > signal condition for better trigger. > > The second trick used in 53132 for measuring frequency is averaging. It > uses an average technique originally from optical frequency measures to > accumulate data into blocks and then subtract the time-stamps of two > subsequent blocks. This is the same as average the output of a number of > overlapping frequency estimations. > > This has advantages as white noise is supressed with a steeper slope, > and the associated deviation is the modified Allan Deviation MDEV. > >> >> >> I understand most of what you said, but I've never taken statistics, so I am >> guessing on some part. I can see how adev goes down as tau gets longer. >> Basically, averaging is taking place. But I am still not sure why at some >> point, it goes back up. I understand noise will start to take effect, but >> the same noise has been there all along while adev was going down. Then, >> why is this inflection point where sign of slope suddenly changes? > OK, so the trouble is that rather than only white noise as classical > statistics deal with, we have at least 4 noise types, with different > frequency slopes. As we try to analyze this with standard deviation, the > standard deviation estimator (RMS estimator) does not converge, is > simply keeps producing noise even if we add more values. To put that in > another way, we do not gain more knowledge by doing more measurements. > The classical white noise is what is called white phase modulation > noise, we then have flicker phase noise, white frequency noise and > flicker frequency noise. All these noise-types is to be expected > according to the David Leeson model, and it is due to those that we need > to use more advanced statistics as introduced by David Allan. > > The White Phase Modulation has a flat frequency response in phase noise > spectrum, 1/tau in ADEV. > The Flicker Phase Modulation has 1/sqrt(f) respone in phase noise > spectrum, 1/tau in ADEV. > The White Frequency Modulation has 1/f respone in phase noise spectrum, > 1/sqrt(tau) in ADEV. > The Flicker Frequency Modulation has 1/sqrt(f^3) response in phase noise > spectrum, flat in ADEV. > > In addition to this, linear frequency drift creates a slope that scales > with drift and tau, so that is an upper limit. Thermal sensitivity tends > to lay ontop as well, so does other disturbances. > > Depending on details of oscillators and their sensitivity to thermal > noise, their effective minimum shifts around. > >> >> >> Also, to reach adev(tau=10), it takes longer than 10 seconds. Manual for >> TimeLab basically says more samples are taken than just 10, but does not >> elaborate further. Say it takes 50 seconds to get there, and say that's the >> lowest point of adev, does that mean it is the best to set gate time to 10 >> second or 50 second? (or even, take whatever gate time and repeat the >> measurement until accumulated gate time equals tau? > The Allan Deviation takes a number of estimates to produce values, but > remember these are stability values for a certain observationtime for > frequency, not the frequency measure itself. > > Cheers, > Magnus > >> --------------------------------------- >> (Mr.) Taka Kamiya >> KB4EMF / ex JF2DKG >> >> >> On Thursday, February 20, 2020, 7:54:22 PM EST, Magnus Danielson >> <mag...@rubidium.se> wrote: >> >> Hi Taka, >> >> On 2020-02-20 19:40, Taka Kamiya via time-nuts wrote: >>> I have a question concerning frequency standard and their Allen deviation. >>> (to measure Allen Dev in frequency mode using TimeLab) >>> >>> It is commonly said that for shorter tau measurement, I'd need OCXO because >>> it's short tau jitter is superior to just about anything else. Also, it is >>> said that for longer tau measurement, I'd need something like Rb or Cs >>> which has superior stability over longer term. >> Seems reasonably correct. >>> Here's the question part. A frequency counter that measures DUT basically >>> puts out a reading every second during the measurement. When TimeLab is >>> well into 1000s or so, it is still reading every second; it does not change >>> the gate time to say, 1000s. >>> That being the case, why this consensus of what time source to use for what >>> tau? >>> I recall reading on TICC, in time interval mode, anything that's reasonably >>> good is good enough. I'm aware TI mode and Freq mode is entirely >>> different, but it is the same in fact that measurement is made for very >>> short time span AT A TIME. >>> I'm still trying to wrap my small head around this. >> OK. >> >> I can understand that this is confusing. You are not alone being >> confused about it, so don't worry. >> >> As you measure frequency, you "count" a number of cycles over some time, >> hence the name frequency counter. The number of periods (sometimes >> called events) over the observation time (also known as time-base or >> tau) can be used to estimate frequency like this: >> >> f = events / time >> >> while it is practical that average period time becomes >> >> t = time / events >> >> In modern counters (that is starting from early 70thies) we can >> interpolate time to achieve better time-resolution for the integer >> number of events. >> >> This is all nice and dandy, but now consider that the start and stop >> events is rather represented by time-stamps in some clock x, such that >> for the measurements we have >> >> time = x_stop - x_start >> >> This does not really change anything for the measurements, but it helps >> to bridge over to the measurement of Allan deviation for multiple tau. >> It turns out that trying to build a standard deviation for the estimated >> frequency becomes hard, so that is why a more indirect method had to be >> applied, but the Allan deviation fills the role of the standard >> deviation for the frequency estimation of two phase-samples being the >> time-base time tau inbetween. As we now combine the counters noise-floor >> with that of the reference, the Allan deviation plots provide a slopes >> of different directions due to different noises. At the lowest point on >> the curve, is where the least deviation of frequency measurement occurs. >> Due to the characteristics of a crystal oscillator to that of the >> rubidium, cesium or hydrogen maser, the lowest point occurs at different >> taus, and provide different values. Lowest value is better, so there is >> where I should select the time-base for my frequency measurement. So, >> this may be at 10 s, 100 s or 1000 s, which means that the frequency >> measurement should be using start and stop measurements with that >> distance. OK, fine. So what about TimeLab in all this. Well, as we >> measure with a TIC we collect a bunch of phase-samples at some base >> rate, such as 10 Hz or whatever. TimeLab and other tools can then use >> this to calculate Allan Deviation for a number of different taus simply >> by using three samples, these being tau in between and algoritmically do >> that for different taus. One then collects a number of such measurements >> to form an average, the more, the better confidence interval we can but >> on the Allan Deviation estimation, but it does not improve our frequency >> estimation, just our estimation of uncertainty for that frequency >> estimation for that tau. Once you have that Allan Deviation plot, you >> can establish the lowest point and then only need two phase samples to >> estimate frequency. >> >> So, the measurement per second thing is more collection of data rather >> than frequency estimation in itself. >> >> Cheers, >> Magnus >> >> >> _______________________________________________ >> time-nuts mailing list -- time-nuts@lists.febo.com >> To unsubscribe, go to >> http://lists.febo.com/mailman/listinfo/time-nuts_lists.febo.com >> and follow the instructions there. >> >> _______________________________________________ >> time-nuts mailing list -- time-nuts@lists.febo.com >> To unsubscribe, go to >> http://lists.febo.com/mailman/listinfo/time-nuts_lists.febo.com >> and follow the instructions there. > > _______________________________________________ > time-nuts mailing list -- time-nuts@lists.febo.com > To unsubscribe, go to > http://lists.febo.com/mailman/listinfo/time-nuts_lists.febo.com > and follow the instructions there. > _______________________________________________ > time-nuts mailing list -- time-nuts@lists.febo.com > To unsubscribe, go to > http://lists.febo.com/mailman/listinfo/time-nuts_lists.febo.com > and follow the instructions there. _______________________________________________ time-nuts mailing list -- time-nuts@lists.febo.com To unsubscribe, go to http://lists.febo.com/mailman/listinfo/time-nuts_lists.febo.com and follow the instructions there.