In digital signal processing, upsampling can refer to the entire process of increasing the sampling rate of a signal, or it can refer to just one step of the process, the other step being interpolation. Complementary to decimation, which decreases sampling rate, it is a specific case of sample rate conversion in a multi-rate digital signal processing system. When upsampling is performed on a sequence of samples of a signal or other continuous function, it produces an approximation of the sequence that would have been obtained by sampling the signal at a higher rate (or density, as in the case of a photograph). For example, if compact disc audio at 44,100 samples/second is upsampled by a factor of 5/4, the resulting sample-rate is 55,125.
Rate increase by an integer factor L can be explained as a 2-step process, with an equivalent implementation that is more efficient:
In this application, the filter is called an interpolation filter, and its design is discussed below. When the interpolation filter is an FIR type, its efficiency can be improved, because the zeros contribute nothing to its dot product calculations. It is an easy matter to omit them from both the data stream and the calculations. The calculation performed by an efficient interpolating FIR filter for each output sample is a dot product:
where the h[o] sequence is the impulse response, and K is the largest value of k for which h[j + kL] is non-zero. In the case L = 2, h[o] can be designed as a half-band filter, where almost half of the coefficients are zero and need not be included in the dot products. Impulse response coefficients taken at intervals of L form a subsequence, and there are L such subsequences (called phases) multiplexed together. Each of L phases of the impulse response is filtering the same sequential values of the x[o] data stream and producing one of L sequential output values. In some multi-processor architectures, these dot products are performed simultaneously, in which case it is called a polyphase filter.
For completeness, we now mention that a possible, but unlikely, implementation of each phase is to replace the coefficients of the other phases with zeros in a copy of the h[o] array, and process the sequence at L times faster than the original input rate. Then L-1 of every L outputs are zero. The desired y[o] sequence is the sum of the phases, where L-1 terms of the each sum are identically zero. Computing L-1 zeros between the useful outputs of a phase and adding them to a sum is effectively decimation. It's the same result as not computing them at all. That equivalence is known as the second Noble identity.
Let X(f) be the Fourier transform of any function, x(t), whose samples at some interval, T, equal the x[n] sequence. Then the discrete-time Fourier transform (DTFT) of the x[n] sequence is the Fourier series representation of a periodic summation of X(f):
When T has units of seconds, has units of hertz. Sampling L times faster (at interval T/L) increases the periodicity by a factor of L:
which is also the desired result of interpolation. An example of both these distributions is depicted in the top two graphs of Fig.1.
When the additional samples are inserted zeros, they increase the data rate, but they have no effect on the frequency distribution until the zeros are replaced by the interpolation filter. Many filter design programs use frequency units of cycles/sample, which is achieved by normalizing the frequency axis, based on the new data rate (L/T). The result is shown in the third graph of Fig.1. Also shown is the passband of the interpolation filter needed to make the third graph resemble the second one. Its cutoff frequency is [note 1] In terms of actual frequency, the cutoff is Hz, which is the Nyquist frequency of the original x[n] sequence.
The same result can be obtained from Z-transforms, constrained to values of complex-variable, z, of the form Then the transform is the same Fourier series with different frequency normalization. By comparison with Eq.1, we deduce:
which is depicted by the fourth graph in Fig.1. When the zeros are inserted, the transform becomes:
depicted by the bottom graph. In these normalizations, the effective data-rate is always represented by the constant 2? (radians/sample) instead of 1. In those units, the interpolation filter bandwidth is ?/L, as show on the bottom graph. The corresponding physical frequency is Hz, the original Nyquist frequency.
Let L/M denote the upsampling factor, where L > M.
Upsampling requires a lowpass filter after increasing the data rate, and decimation requires a lowpass filter before downsampling. Therefore, both operations can be accomplished by a single filter with the lower of the two cutoff frequencies. For the L > M case, the interpolation filter cutoff, cycles per intermediate sample, is the lower frequency.
Manage research, learning and skills at defaultLogic. Create an account using LinkedIn or facebook to manage and organize your Digital Marketing and Technology knowledge. defaultLogic works like a shopping cart for information -- helping you to save, discuss and share.Visit defaultLogic's partner sites below: