Lecture - Fundamental Concepts
Website: | WueCampus |
Kurs: | vhb : Radio-Astronomical Interferometry (DEMO) |
Buch: | Lecture - Fundamental Concepts |
Gedruckt von: | Gast |
Datum: | Sonntag, 24. November 2024, 13:34 |
1. Fourier optics
Learning Objectives:
- Basic concepts applied to telescopes
- Fourier and convolution mathematics
- Application of telescopes as spatial filters
2. Interferometry
Learning Objectives:
- Interferometry basics and concepts
- Application in radio astronomy
- Introduction to important values like visibility function and bandwidth
- Influences of the instrument set up to the measured values
- Orientation of pointing in the scope of coordinates
4. Receiver Response
Learning Objectives:
- Process and advantages of heterodyne frequency conversion
- Sensitivity of a radio-interferometrical array
- Effects of sampling, weighting and gridding
- Effect of finite bandwidth, bandwidth smearing
- Calibration of radio-interferometrical observations
4.1. Heterodyne frequency conversion
4.2. Interferometer sensitivity
In order to estimate whether a source can be detected by a radio telescope or array, the so-called signal-to-noise ratio (SNR) has to be calculated. For a single telescope the signal is given by the measured antenna temperature and the noise is given by the radiometer equation
in which is a dimensionless constant depending on the receiver system used, is the system temperature, is the bandwidth of the receiver equipment and is the integration time.
In radio interferometry, the signal is given by the brightness distribution of a source that must be calculated by the Fourier integral of the measured visibility :
Therefore, to calculate the SNR in radio interferometry, one has to calculate the uncertainty of the brightness distribution from the uncertainty of the complex visibility . As in every measurement, the measured data need to be sampled, meaning that the visibility is only measured at discrete locations along the tracks in the -plane. Let: be the integration time of the individual samples in the -plane, be the total integration time and be the total number of antennas. Then the total number of measured visibilities is given by
in which
is the number of antenna pairs or rather the number of individual two-element interferometers. Since this brightness distribution suffers from incomplete -coverage, the resulting image is called a "dirty image", which is also indicated by the superscript "D" of the brightness distribution . The visibilities are measured at discrete locations , so the brightness distribution is given by the discrete Fourier transform
in which is a normalization constant and is a sampling function which is given by
This sampling function is only non-zero where the visibility is measured in the -plane.
To suppress sidelobes, the beam shape of the observing array can be controlled by applying appropriate weights to the measured visibilities. Furthermore, if the observing array consists of telescopes that have different collecting areas , and different receivers with different system temperatures , frequency bandwidths and integration times , one can also apply weights to control these differences. The weights can be written as
in which accounts for different telescope properties, is a taper which controls the beam shape and weights the density of the measured visibilities (more detailed information on the weights, especially on the -factor, will be given in the next section 2.4.3). With these weights, the brightness distribution can then be written as
Because there is no measurement in the center of the -plane, the terms of the dirty image vanish, leading to
in which
and
Note that the two additional terms and vanish under the integral of and can therefore also be neglected for the sum of .
To calculate the sensitivity in the image plane, a point source at the phase and image center is considered. For such a point source, the visibility is real and constant across the entire -plane and only varies due to random noise. Furthermore, because , the -term vanishes and . Therefore, the dirty image of such a point source is given by
in which is the total flux density of the source. Note that here
so that
Assuming that the uncertainty of the flux density , the noise in the dirty image is given by
Furthermore, assuming an observing array that consists of identical telescopes and a naturally weighted , untapered image, simplifies to
Since , the uncertainty of the flux density is given by the rms noise of the measured visibilities which is, based on the radiometer equation, given by
in which is the Boltzmann constant and
is the quantization efficiency accounting for the quantization noise due to the conversion of analogue signals to digital signals.
Inserting into , the sensitivity for the synthesized image is finally given by
4.3. Sampling, weighting, gridding
Sampling
As previously mentioned, the visibilities are only measured at discrete locations . Therefore, the measured visibilities are given by the multiplication of the true visibility function and a sampling function . Hence, the dirty image is then given by the convolution of their Fourier transforms:
in which denotes the Fourier transform of and the double-star indicates a two-dimensional convolution.
Weighting
Furthermore, as also previously mentioned, the visibilities can be weighted by multiplying a weighting function to the visibilities :
This weighting function consists of three individual factors:
- accounts for different telescope properties within an array (e.g. different , , and ),
- is a taper that controls the beam shape,
- weights the density of the measured visibilities.
The -factor affects the angular resolution and sensitivity of the array. On the one hand side, the highest angular resolution can be achieved by weighting the visibilities as if they had been measured uniformly over the entire -plane. Therefore, this weighting scheme is called uniform weighting. Since the density of the measured visibilities is higher at the center of the -plane, the visibilities in the outer part are over-weighted leading to the highest angular resolution. On the other hand, the highest sensitivity is achieved if all measured visibilities are weighted by identical weights. This weighting scheme is called natural weighting. The main properties of both schemes are summarized in the following:
- uniform weighting:
With this weighting function , the dirty image is given by
Gridding
To use the time advantage of the fast Fourier transform (FFT) algorithm, the visibilities must be interpolated onto a regular grid of size , resulting in an image of size . This interpolation, also called gridding, is done at first by convolving the weighted discrete visibility with an appropriate function to obtain a continuous visibility distribution. This continuous visibility distribution is then resampled at points of the regular grid with spacings and by multiplying a two-dimensional Shah-function , given by
After these modifications, the visibility is given by
and the dirty image reads
This gridding process is illustrated in the Fig. 2.29, in which real -tracks measured by the IRAM interferometer at Plateau de Bure in France are shown as black dotted ellipses. Furthermore, a regular grid, onto which the visibilities have been interpolated, is shown as red dots.
Since the sampling intervals and in the -plane are inversely proportional to the sampling intervals and in the image plane ( and for a grid of size ), the maximum map size in one domain is given by the minimum sampling interval in the other. Therefore, it is important to choose appropriate sampling intervals and . If and are chosen to be too large for example, this will result in artefacts in the image plane produced by reflections of structures from the map edges, which is called aliasing.
The most effective way to deal with this aliasing is to use a convolution function for which the Fourier transform in the image plane decreases rapidly at the image edges and is nearly constant over the image. Therefore, the simplest choice for the convolution function is a rectangular function, but with this choice the aliasing would be strongest. A better choice would be a sinc function, however a Gaussian-sinc function (product of a Gaussian and a sinc function) leads to the best suppression of aliasing.
Finally, the gridding modifications must be corrected after the Fourier transform by dividing the dirty image by the inverse Fourier transform of the gridding convolution function . Therefore, the so-called grid-corrected image is given by
4.4. Bandwidth smearing
As already seen in Chapt. 2 Sect. 4.2, the sensitivity of radio-interferometrical measurements depends on the effective area of the telescopes and the bandwidth . The larger , the higher the sensitivity. However, using a large bandwidth is problematic, since the Fourier relation between brightness distribution and visibility is only valid for monochromatic signals. Therefore, the effects of the finite bandwidth on the images obtained from radio-interferometrical observations, especially the inevitable effect called bandwidth smearing or chromatic aberration, will be investigated in this section, following the textbook of Taylor et al. (1999).
For a single observing frequency the brightness distribution is given by
in which the "~" indicates the influence of a bandpass. The frequency-independent coordinates and are given by
where and are the actual spatial coordinates of the visibility at the frequency .
Therefore, using the generalized similarity theorem for Fourier transformations in dimensions, which reads
the two-dimensional Fourier relation between visibility and brightness distribution is given by
Furthermore, since the smeared visibility is obtained by rescaling and weighting the true visibility by a normalized bandpass function , where , and then integrating over the frequency band, another important effect must be taken into account. There will be a delay error of
for signals arriving from a direction at frequency . Therefore, the phase is shifted by
and the smeared visibility is given by
For simplicity, a point source with unit amplitude located at is assumed without restricting generality. Then, the true visibility is given by
and the smeared visibility reads
Furthermore, assuming that the bandwidth is sufficiently small, so that (in practice, ), the bandwidth-smeared brightness distribution is given by
using . Here, one can see that the term in squared brackets is the Fourier transform of the normalized bandpass function over , to an argument that represents a delay corresponding to the positional offset . Therefore, it is helpful to define a delay function , given by
Using this delay function, the bandwidth-smeared brightness distribution then reads
Therefore, the bandwidth-smeared brightness distribution is the Fourier transform over of the product of the true visibility with the delay function. Using the convolution theorem, the brightness distribution is given by
which is the convolution of the true image with a position-dependent bandwidth distortion function , which is the Fourier transform of the delay function
over . Since this distortion function varies with the radial distance from the phase center and is always oriented along the radius to the phase center, the final image of an extended source can be interpreted as a radially-dependent convolution.
The effect of finite bandwidth is illustrated in Fig. 2.30. Here it is assumed that the observed emission of a source at position positively interferes exactly at frequency . A signal from a neighboring position shifted by with respect to measured by one telescope at a frequency deviating from by an amount of , for example, can then also produce maximum interference with a signal from measured by the other telescope at frequency , if the delay difference matches. Therefore, a source observed with finite bandwidth can also produce structures away from its nominal position.
This finite bandwidth effect cannot be removed, because it is described by a mathematical functional. However, there are two methods to minimize the effects of bandwidth smearing. The first method is to limit the field size by dividing the observed area into subfields and produce separate images for all such subfields. To cover the complete area, the images of the subfields can then be fitted together. This technique is called mosaicing. The second method is to split the full frequency band into narrower sections, which will be summed up after imaging each individual data set. This method is called bandwidth synthesis.
4.5. Calibration
The true visibility of a target source observed by two telescopes and of an array is related to the observed visibility by
in which is the time-variable and complex gain factor of the correlation product of the two telescopes and . This gain factor includes variations caused by different effects, e.g., weather effects, effects of the atmospheric paths to the telescopes, and instrumental effects. Therefore, to determine the true visibility, the gain factor has to be calculated. This can be done via frequent observations of calibrator sources with known flux densities and accurately known positions. Furthermore, these sources have to be point sources, meaning that their angular size is
Then, the complex gain of each interferometer can be determined by
in which is the observed visibility of a calibrator source. This complex gain factor can then be used to calibrate the observed visibility of the target source. The true visibility of a target source is then given by
Note that the gains can also be expressed by the voltage gain factors and of the individual i-th and j-th antennas by
This reduces the amount of calibration data, since there are many more correlated antenna pairs than antennas in large arrays. Furthermore, this makes the calibration procedure more flexible, since, for example, a source that is resolved at the longest baselines of an array can be used to compute the gain factors of the individual antennas from measurements made only at shorter baselines, at which the source appears unresolved.
Gibb's phenomenon
Fig. 2.34 Real and imaginary part of a frequency power spectrum for an idealized rectangular bandpass. |
---|
In continuum observations the correlation of the signals is usually measured for zero time lags, meaning that the correlation of the measured signal is averaged over the bandwidth, which is defined via a bandpass. In spectral line observations (see Chapt. 3 Sect. 4), measurements at different frequencies across the bandpass are required, which can be implemented by correlating the signals as a function of time lags. After Fourier transforming this correlation output, this yields the so-called cross power spectrum. This cross power spectrum is obtained for frequency channels (for more detailed information see Chapt. 3 Sect. 4).
To calibrate an interferometer over the bandpass, which represents the filter characteristics of the last IF stage, a continuum point source is used to determine the complex gains of the interferometer, which yields the true cross-correlation spectrum of the observed target source. For calibrator sources with a constant continuum spectrum over the bandpass, the behavior of the complex gains and across the bandpass are given by the correlated power as a function of frequency for each interferometer of the observing array. The real and imaginary parts of this correlated power, measured for a continuum source with a flat spectrum, are shown in Fig. 2.31 for an idealized rectangular bandpass.
The observed visibility of an unknown target source is given by
in which is the Fourier transform of the truncating function in the frequency domain, which is a sinc function given by
Here, is given by a convolution with , since the cross-correlation has only a finite maximum lag of . The observed visibility of a calibrator source with and is then given by
Finally, the true visibility of the observed target source is given by
in which the gains do not cancel out, due to the convolution. Therefore, the original frequency power spectrum, shown in Fig. 2.31, is convolved with the sinc function , due to the finite lag, which will lead to large errors, if the target source has a spectral line near . This is known as Gibbs' phenomenon.
Since the cross-correlation function has a complex power spectrum, Gibbs' phenomenon has strong influence on the imaginary part of the true visibility, because of the ratio in the equation for , which leads to non-negligible errors, especially for strong line emission near the step at . Since the step in the imaginary part is large, these errors have a strong effect on the phases, producing strong ripples.
5. Image reconstruction
Learning Objectives:
- Image reconstruction using the CLEAN and MEM algorithms
- Image defects
- Image reconstruction using self-calibration
6. Digital Beamforming
Learning Objectives:
- Geometric understanding of an array performing beamforming
- Introduction and mathematical concept of weighting elements
- Mathematical description of one and two dimensional array factors
- Visualization of array factors, beam forming and beam steering