[time-nuts] timestamps on downconverted data streams

Jim Lux jimlux at earthlink.net
Sun Dec 20 09:40:34 EST 2015


Here's an interesting problem.

You have a fast sampler that is collecting samples off-the-air (e.g. the 
end of LORAN) with a fairly wide bandwidth: say 10 Megasamples per second.

Those samples get post processed in a digital downconverter (not 
necessarily in real time) to a narrower band representation at a lower 
sample rate.

You know when the input samples were acquired: e.g. you've got a good 
oscillator, and a reliable sync pulse.  For instance, your handy GPSDO 
(or the ensemble of H2 masers in your garage) might give you a 1pps tick 
good to, say, 20 ns, so you know when your 10 MHz samples were taken (to 
20ns)

Is there a consistent (and standardized) way to calculate and report the 
time of the output samples.

Each output sample is composed of information from multiple input samples.

One could test the system by digitizing a signal with known timing (e.g. 
a 1 MHz sine wave, where the zero crossing is "on the second") and then 
look for the zero crossing in the downconverted output.  Depending on 
the filtering in the downconverter, there's some time vs frequency 
characteristic that could be used to back out any deltas for other 
frequencies.

So you could report the time of the low rate output samples in terms of 
the time of the input sample, at least for the 'center frequency' of the 
downconverter.



More information about the time-nuts mailing list