Free Standard AU & NZ Shipping For All Book Orders Over $80!
Register      Login
Publications of the Astronomical Society of Australia Publications of the Astronomical Society of Australia Society
Publications of the Astronomical Society of Australia
RESEARCH ARTICLE (Open Access)

Basic Testing of the duchamp Source Finder

T. Westmeier A C , A. Popping A and P. Serra B
+ Author Affiliations
- Author Affiliations

A International Centre for Radio Astronomy Research, M468, The University of Western Australia, 35 Stirling Highway, Crawley WA 6009, Australia

B ASTRON, Postbus 2, 7990 AA Dwingeloo, the Netherlands

C Corresponding author. Email: tobias.westmeier@uwa.edu.au

Publications of the Astronomical Society of Australia 29(3) 276-295 https://doi.org/10.1071/AS11041
Submitted: 31 August 2011  Accepted: 9 December 2011   Published: 17 January 2012

Journal Compilation © Astronomical Society of Australia 2012

Abstract

This paper presents and discusses the results of basic source finding tests in three dimensions (using spectroscopic data cubes) with duchamp, the standard source finder for the Australian Square Kilometre Array Pathfinder. For this purpose, we generated different sets of unresolved and extended Hi model sources. These models were then fed into duchamp, using a range of different parameters and methods provided by the software. The main aim of the tests was to study the performance of duchamp on sources with different parameters and morphologies and assess the accuracy of duchamp's source parametrisation. Overall, we find duchamp to be a powerful source finder capable of reliably detecting sources down to low signal-to-noise ratios and accurately measuring their position and velocity. In the presence of noise in the data, duchamp's measurements of basic source parameters, such as spectral line width and integrated flux, are affected by systematic errors. These errors are a consequence of the effect of noise on the specific algorithms used by duchamp for measuring source parameters in combination with the fact that the software only takes into account pixels above a given flux threshold and hence misses part of the flux. In scientific applications of duchamp these systematic errors would have to be corrected for. Alternatively, duchamp could be used as a source finder only, and source parametrisation could be done in a second step using more sophisticated parametrisation algorithms.

Keywords: methods: data analysis

1 Introduction

With the advent of the Square Kilometre Array (SKA; Dewdney et al. 2009) and its precursors and pathfinders, including the Australian SKA Pathfinder (ASKAP; DeBoer et al. 2009), the Karoo Array Telescope (Meer-KAT; Jonas 2009), and the Aperture Tile In Focus (APERTIF; Oosterloo et al. 2009), the prospect of deep radio continuum and Hi surveys of large areas on the sky demands for new strategies in the areas of data reduction and analysis, given the sheer volume of the expected data streams, in particular for spectroscopic surveys.

Of particular importance is the automatic and accurate identification and parametrisation of sources with high completeness and reliability. Due to the large data volumes to be searched, source finding algorithms must be fully automated, and the once common practice of source finding `by eye' will no longer be feasible. Moreover, accurate source parametrisation algorithms need to be developed to generate reliable source catalogues free of systematic errors, as otherwise the integrity of scientific results based on the survey data could be compromised.

In this paper we will take a closer look at the duchamp source finder1 (Whiting 2012; Whiting & Humphreys 2012). duchamp has been developed by Matthew Whiting at CSIRO as a general-purpose source finder for three-dimensional data cubes as well as two-dimensional images and will serve as the default source finder in the processing of data from the ASKAP survey science projects. The software identifies sources by searching for regions of emission above a specified flux threshold. To improve its performance, duchamp offers several methods of preconditioning and filtering of the input data, including spatial and spectral smoothing as well as reconstruction of the entire image or data cube with the help of wavelets. In addition to source finding, duchamp provides the user with basic source parametrisation, including the measurement of position, size, radial velocity, line width, and integrated flux of a source. More information about the capabilities of the software is available from the duchamp User Guide (Whiting 2011). A brief overview of duchamp's basic functionality is provided in Section 3.

So far, duchamp's source finding and parametrisation capabilities have never been systematically tested on a large set of artificial sources with well-defined parameters. The aim of this paper is to bridge this gap by thoroughly testing the performance of duchamp on sets of artificial point sources and galaxy models as well as a data set containing real galaxies and telescope noise. The tests were originally motivated by the need to identify suitable source finding algorithms for the Widefield ASKAP L-Band Legacy All-Sky Blind Survey2 (WALLABY, Koribalski & Staveley-Smith 2009), one of the large, extragalactic ASKAP survey science projects currently in preparation (Westmeier & Johnston 2010). Hence, the tests presented here will focus on the detection of compact and extended Hi sources, in particular galaxies, in three-dimensional data cubes with ASKAP characteristics.

However, we believe that the results and conclusions presented in this paper will be of interest not only to those involved in SKA precursor science, but to a larger community of astronomers interested in the automatic detection and parametrisation of sources in their data sets, regardless of the wavelength range involved. For a comparison of duchamp's performance with that of other source finding algorithms we refer to the paper by Popping et al. (2012) in this issue.

This paper is organised as follows: In Section 2 we summarise the source finding strategies of other large Hi surveys in the past, followed by a brief overview of the duchamp source finder in Section 3. In Section 4 we present the outcome of our test of duchamp on point source models with simple Gaussian line profiles. Section 5 describes our testing of duchamp on models of disc galaxies with varying physical parameters. In Section 6 we apply duchamp to a data cube containing real galaxies and genuine noise extracted from radio observations. A discussion of our results is presented in Section 7. Finally, Section 8 summarises our main results and conclusions.


2 Source Finding in Previous Surveys

Some of the previous, large Hi surveys, including the Hi Parkes All-Sky Survey (HIPASS; Barnes et al. 2001), the Hi Jodrell All-Sky Survey (HIJASS; Lang et al. 2003), and the Arecibo Legacy Fast ALFA survey (ALFALFA; Giovanelli et al. 2005), already had to deal with the issue of (semi-)automatic source detection.

In the case of HIPASS, two different source finders were used and the results combined to maximise completeness (Meyer et al. 2004). The first algorithm, multifind (Kilborn 2001), used a simple 4σ flux threshold combined with smoothing of the data cube on different scales. The second algorithm, tophat, detected sources in the spectral domain by convolving each spectrum in the data cube with a top-hat function of varying width. Neither of the two algorithms alone managed to detect more than 90% of the final, combined source list. The two algorithms combined produced about 140,000 unique detections, all of which were inspected by eye to remove potential false detections. The final HIPASS catalogue included 4315 sources (Meyer et al. 2004), resulting in an overall reliability of the automatic source finding algorithms of only about 3%.

In the case of HIJASS, again two different methods were used and the results combined to improve completeness (Lang et al. 2003). The first method simply involved searching the cubes by eye to extract potential sources. The second algorithm, polyfind, first searched for signals above a given threshold in a smoothed version of the data cube and then ran a series of matched filters over the detected signals to decide whether a signal was likely genuine or false. As in the case of HIPASS, the source list produced by the automatic source finding routine was inspected by eye to further reject potential false detections. The positions of uncertain detections were re-observed to either confirm of refute them.

For the ALFALFA survey, a matched-filtering technique was applied to the data in the spectral domain (Saintonge 2007). The data were convolved with a set of template functions created by combining the first two symmetrical Hermite functions, Ψ0(x) and Ψ2(x). The resulting templates range from simple Gaussian profiles for narrow signals to double-peaked profiles for broader signals, covering the range of spectral shapes expected from Hi observations of real galaxies. In tests on 1500 simulated galaxies, 100% reliability and about 70% completeness are achieved at an integrated signal-to-noise ratio of S/N ≈ 6, while the 90% completeness level is exceeded at S/N 9.3


3 The duchamp Source Finder

duchamp has been implemented as a general-purpose source finder for three-dimensional spectral-line data cubes with two spatial axes and one frequency (or velocity) axis, although the software can also operate on one- and two-dimensional data sets (Whiting 2011). duchamp finds sources by applying a simple flux threshold to the data cube, specified by the threshold or snrCut keyword, and searching for pixels above that threshold. In a second step, the software attempts to merge detections into sources under specific circumstances that can be controlled by the user. One option is to simply merge adjacent pixels (flagAdjacent keyword). Alternatively, a maximum spatial and spectral separation can be specified for the merging of detected pixels into sources (threshSpatial and threshVelocity keywords, respectively). Once detected, sources can be `grown' to a flux level below the actual detection threshold, using the flagGrowth and growthThreshold/growthCut keywords.

Basic removal of false detections is achieved by requiring that sources comprise a minimum number of contiguous spatial pixels and spectral channels, using the minPix and minChannels keywords, respectively. To improve the reliability of the source finding even further, duchamp offers a powerful method of reconstructing the entire input data cube with the help of wavelets. Source finding is then performed on the reconstructed cube instead of the original input cube. Reconstruction can either be carried out in all three dimensions of the cube, or in the spatial (two-dimensional) or spectral (one-dimensional) domain only.

duchamp uses the so-called ‘à trous’ wavelet reconstruction method (Starck & Murtagh 2002). First, the input data set is convolved with a specific wavelet filter function (three different functions are offered to the user by duchamp). The difference between the convolved data set and the original data set is then added to the reconstructed cube. Next, the scale of the filter function is doubled and the procedure repeated, using the convolved array as the new input data set. Once the user-specified maximum filter scale is reached, the final convolved data set is added to the reconstructed cube, and source finding on the reconstructed data set commences.

The ‘à trous’ wavelet reconstruction of the data cube offers a powerful method of enhancing duchamp's source finding capabilities. First of all, the user can select the minimum (scaleMin keyword) and maximum (scaleMax keyword) filter scales to be used in the reconstruction, providing efficient suppression of small-scale and large-scale artefacts in the data, such as noise peaks, baseline ripples, or radio-frequency interference. Furthermore, the user can specify an additional threshold (snrRecon keyword) to be applied when adding wavelet components to the reconstructed data cube, thereby reducing even further the number of spurious signals in the data cube.

In comparison to simple data thresholding, the ‘à trous’ wavelet reconstruction method will greatly increase the completeness and reliability of duchamp's source finding procedure, and hence the method has been applied in all source finding tests presented in this paper.


4 Point Sources with Gaussian Spectral Profiles

For our first test of duchamp we generated models of 1024 point sources with simple Gaussian spectral line profiles. This will allow us to assess the fundamental performance of duchamp under ideal conditions and to investigate the accuracy of the software's source parametrisation algorithms. Point sources with Gaussian profiles are ideal for this test because — as a consequence of their simple morphology — their physical parameters can be exactly defined and calculated to serve as a benchmark for duchamp's parametrisation.

In order to create the model data set, the miriad (Sault, Teuben & Wright 1995) task uvgen was employed to generate visibility data of Gaussian noise at a frequency of 1.4 GHz with ASKAP characteristics and parameters similar to those anticipated for the WALLABY survey. The model parameters are summarised in Table 1.


Table 1.  Summary of parameters used to generate the visibility data set and noise image for the point source models
T1

The visibility data were Fourier-transformed using miriad's task invert to generate a noise image of 600 × 600 pixels and 31 spectral channels with characteristics similar to WALLABY (again, see Table 1 for details). The rms noise level in this image is σ = 1.95 mJy which is only slightly higher than the 1.6 mJy expected for WALLABY.

In order to generate images of point sources, the miriad task imgen was used to create 1024 data cubes each of which has a size of 31 × 31 pixels and 31 spectral channels and contains a single point source with Gaussian spectral line profile in the centre. Each source was randomly assigned a peak flux in the range of 1 to 20σ, resulting in an average of about 54 sources per 1σ interval. Spectral line widths (FWHM) range from 0.1 to 10 spectral channels, equivalent to approximately 0.4 to 38.6 km s–1, resulting in a density of about 27 sources per 1 km s–1 line width interval. While in reality sources with Hi line widths of as small as 0.4 km s–1 will not exist, the reason for including such narrow lines in our test is to study the performance of duchamp on sources that are spectrally unresolved, irrespective of the absolute line width.

Each of the 1024 cubes was convolved with the beam model produced by invert. Next, a random portion of 31 × 31 pixels of the original noise cube was selected and added to each convolved image to create the final images used for testing duchamp. To facilitate correct integrated flux measurements, we added information on the synthesised beam to the image header. The entire procedure is outlined in Figure 1. An example image and spectrum of one of the point source models is shown in Figure 2.


Figure 1  Outline of the procedure used to create the model data cubes of point sources with miriad.
F1


Figure 2  Example of a point source model generated for testing duchamp. The left-hand panel shows a single-channel map of the data cube (at the systemic velocity of the source), and the right-hand panel depicts the spectrum at the source position. The circle in the map illustrates the half-power beam width.
Click to zoom

4.1 Running duchamp

Next, we ran duchamp (version 1.1.8) on the data cubes. In order to find out which combination of control parameters provided the best performance in terms of completeness and reliability, we first ran duchamp several times with different flux thresholds and minimum wavelet scales to test the performance of each set of parameters (Table 2). An overview of the different completeness and reliability levels achieved in these runs as a function of integrated flux of the source is shown in Figure 3.


Table 2.  Relevant input parameters for the different test runs of duchamp
T2


Figure 3  Completeness as a function of integrated flux for different tests of duchamp with varying control parameters. The parameters employed in the different runs are listed in Table 2. The overall reliability for each run is listed in the legend.
F3

We then selected the best set of control parameters for the analysis presented in this section. In this best-performing run (number 5 in Figure 3 and Table 2) we used a 1.5σ flux threshold equivalent to 2.9 mJy. In addition, we made use of duchamp's ‘à trous’ wavelet reconstruction. We employed a full three-dimensional wavelet reconstruction with a minimum wavelet scale of 2 (i.e. the smallest scales were excluded to suppress noise in the reconstructed cube) and a flux threshold of 3σ for wavelet components to be included in the reconstructed cube. In addition, we required sources to cover a minimum of 5 contiguous pixels in the image domain and 3 contiguous spectral channels above the detection threshold to be included in the final source catalogue. This will further reduce the number of spurious detections. The duchamp input parameters explicitly set in the parameter file are listed in Table 3.


Table 3.  duchamp input parameters (Whiting 2011) explicitly set in the input-parameter file for point-source modelsa
T3

The 1024 output parameter files generated by duchamp were concatenated, and those source entries whose positions were within ±1 pixel of the nominal source position were considered as genuine detections and selected for further processing and analysis. The results of this analysis will be presented and discussed in the following sections.

For a number of reasons it is not possible to specify the typical time it takes for duchamp to process a certain amount of data. Firstly, the performance of duchamp strongly depends on the exact choice of input parameters, including detection threshold, wavelet reconstruction choices, or settings related to merging and discarding of initial detections. Three-dimensional wavelet reconstruction of the input data cube, for example, is particularly computationally expensive. Secondly, the running time of duchamp on a particular data cube will depend on a large number of details, including the number of sources in the cube, their size and morphology, and in principle even the number density of sources in the cube. Thirdly, recent updates of the software have resulted in a significant improvement of duchamp's performance, in particular compared to version 1.1.8 used for part of the testing presented in this paper.

To get a basic idea of the impact of the aforementioned parameters on the running time of duchamp we performed a few simple tests on a standard laptop computer with a state-of-the-art, dual-core 2.3 GHz CPU (only one core at a time was actually engaged) and 4 GB of physical memory. We ran duchamp several times with different parameters on our artificial noise data cube of 600 × 600 spatial pixels and 31 spectral channels without any sources in it. Using a 5σ detection threshold, duchamp takes about 0.64 s of CPU time to run, producing no detections. When performing a one-dimensional wavelet reconstruction in the spectral dimension prior to source finding, the running time increases by a factor of 30 to about 19 s. Full three-dimensional wavelet reconstruction is slower by a factor of 120, requiring about 77 s to complete.

As mentioned before, these numbers strongly depend on the number and nature of sources present in the cube. Decreasing the flux threshold to 3σ without wavelet reconstruction results in 966 detections (all of which are noise peaks) and increases the running time of duchamp to about 3.5 s. Processing time will also increase with cube size. Doubling the cube size to 62 spectral channels increases the running time by a factor of 2 without wavelet reconstruction, but by factors of 1.9 and 2.6 in the case of one-dimensional and three-dimensional reconstruction, respectively, indicating that an increase in cube size does not translate into a proportional increase in processing time when dealing with wavelet reconstruction.

In summary, the time duchamp needs to process a data cube is a complicated function of not only the machine specifications (e.g. CPU, memory, data transfer speed), but also the input parameters (e.g. flux threshold, wavelet reconstruction) and properties of the data set concerned (e.g. cube dimensions, number of sources). Hence, running times are almost impossible to predict and may have to be determined experimentally on a case-by-case basis. Instead of asking whether duchamp is fast enough for a particular problem, the user would have to determine the optimal set of conditions that would allow processing of the data in a given period of time. An alternative option would be to separate the problem into multiple, parallel processes.

4.2 Results

4.2.1 Completeness and Reliability

Two of the most important parameters in the characterisation of source finder performance are completeness and reliability. Completeness is defined as the number of genuine detections divided by the true number of sources present in the data. Completeness can either be calculated for the entire sample or more sensibly for a subset, e.g. for sources within a certain parameter range. Reliability is defined as the number of genuine detections divided by the total number of detections produced by the source finder. Reliability can only be calculated for the entire sample of sources and not for a subset of sources within a certain parameter range, because false detections do not possess physical parameters as such. Alternatively, the parameters derived by the parametrisation algorithm of the source finder can be used to derive reliability as a function of different source parameters, but it is important to note that for genuine sources those parameters can be affected by systematic errors and do not necessarily correspond to the original source parameters.

The ideal source finder would produce a completeness and reliability of 100%. In reality, however, we will have to find a compromise between good completeness and good reliability. In the case of duchamp, for example, decreasing the flux threshold for detections will lead to an increase in completeness, but at the cost of lower reliability.

In our test of duchamp on the set of 1024 point source models the software finds a total of 1103 sources of which 850 are genuine detections. The remaining 253 detections are false positives due to strong noise peaks in the data cube. These numbers translate into an overall completeness of 83.0% and an overall reliability of 77.1%.4

Completeness as a function of peak signal-to-noise ratio is plotted in the top panel of Figure 4. The detection list produced by duchamp is complete down to a peak flux level of Fpeak ≈ 10σ, but below that level completeness decreases to below 50% at about 3σ. The completeness curve shows a much steeper rise when plotted against integrated flux instead of peak flux (middle panel of Figure 4). The 100% completeness level is reached at Fint ≈ 0.3 Jy km s–1, corresponding to an Hi mass of about 7 × 104M at a distance of 1 Mpc, or 7 × 108M at 100 Mpc, for the expected 8-hour integration per pointing of the WALLABY project on ASKAP. Below that flux level there is a sharp drop in completeness.


Figure 4  Top panel: Completeness of the point source models as a function of true peak signal-to-noise ratio in bins of 1σ. Middle panel: Same, but as a function of true integrated flux in bins of 0.1 Jy km s–1. Bottom panel: Same, but as a function of true line width (FWHM) in bins of 2.5 km s–1.
F4

The bottom panel of Figure 4 shows completeness plotted as a function of true line width (FWHM), irrespective of the peak flux and integrated flux of a source. Over most of the covered line width range the completeness remains constant at approximately 90%, but it gradually decreases to about 50% below line widths of 10 km s–1. This decrease is presumably the result of the ‘à trous’ wavelet reconstruction of the data cube. By ignoring the smallest wavelet scales in the reconstruction we suppress the detection of noise peaks, but at the same time we are also less sensitive to genuine sources with narrow spectral lines.

Reliability as a function of measured peak signal-to-noise ratio, measured integrated flux, and measured line width (w50) is plotted in the top, middle, and bottom panels of Figure 5. duchamp achieves 100% reliability at a peak signal-to-noise ration of about 5 and an integrated flux level of approximately 0.1 Jy km s–1. Reliabilities range between about 80% to 100% over most of the covered line width range, but drop significantly for sources with narrow lines of w50 15 km s–1 due to the increasing number of false detections associated with noise peaks. For line widths of less than about 5 km s–1 the reliability increases sharply, because duchamp effectively filters narrow signals caused by noise peaks through wavelet filtering and minimum channel requirements.


Figure 5  Top panel: Reliability of the point source models as a function of measured peak signal-to-noise ratio in bins of 1σ. Middle panel: Same, but as a function of measured integrated flux in bins of 0.05 Jy km s–1. Bottom panel: Same, but as a function of measured line width (w50) in bins of 2.5 km s–1.
F5

However, as discussed earlier, reliability calculations are very difficult to assess and should be approached with great caution. First of all, reliability can only be specified as a function of measured source parameters because false detections do not have genuine physical parameters. Any errors in a source finder's parametrisation will therefore affect the calculated reliability curves. Secondly, the actual reliability numbers are entirely meaningless in the case of model data as they depend on how the sources were distributed across the model data cube. Increasing the volume of the cube (without increasing the number of sources therein) will result in lower reliabilities as the fraction of false detections increases. Consequently, reliabilities can only be compared on a relative scale, e.g. when testing different source finders on the same data set to determine which algorithm performs best.

4.2.2 Source Position

The resulting position errors are plotted in the left-hand panel of Figure 6. duchamp does an excellent job in determining accurate source positions, with a mean position error of 0.0 ± 1.6 arcsec in right ascension and 0.1 ± 1.5 arcsec in declination.


Figure 6  Left-hand panel: Position error of the point source models in right ascension and declination. Right-hand panel: Mean position error (black data points) and corresponding standard deviation (error bars) as a function of true peak signal-to-noise ratio in 1σ bins.
F6

The mean position error (in terms of angular separation from the nominal source position) as a function of peak signal-to-noise ratio, in bins of 1σ, is shown in the right-hand panel of Figure 6. For bright sources of Fpeak ≈ 20σ the mean position error is approximately 1 arcsec, increasing to about 5 arcsec for Fpeak ≈ 3σ. These numbers correspond to only about 4% and 19%, respectively, of the FWHM of the synthesised beam.

Two limitations should be noted at this point. First of all, in our models the source was always placed exactly on the central pixel of the data cube. We did not explicitly test placement of sources at positions in between the grid points of the cube, which — in the case of point sources — could result in reduced detection rates and less accurate source positions. Secondly, as with other source parameters, source positions will be inaccurate in cases where two or more sources are confused.

4.2.3 Radial Velocity

An overall histogram of radial velocity errors derived from the duchamp run is shown in the left-hand panel of Figure 7. As expected, velocity errors have an approximately Gaussian distribution centred on zero. The mean velocity error of all sources is 0.0 ± 1.7 km s–1. The red, dashed curve in Figure 7 shows the result of a Gaussian fit to the histogram. While the overall distribution of velocity errors follows the fitted Gaussian function, there are a few significant deviations, namely a somewhat higher and sharper peak in the centre (which is slightly shifted into the negative range) and conspicuous `wings’ between 2 and 3 km s–1 (both positive and negative) where source counts are systematically too high with respect to the fit. The FWHM of the fitted Gaussian is 1.94 ± 0.04 km s–1, and the centroid is –0.026 ± 0.017 km s–1 which deviates from zero by about 1.5σ, reflecting the aforementioned negative offset of the peak of the histogram. These deviations from a pure Gaussian distribution are possibly caused by digitisation effects related to the segmentation of the frequency axis into discrete bins of 18.3 kHz equivalent to 3.86 km s–1.


Figure 7  Left-hand panel: Histogram of radial velocity errors (black curve) of the point source models in bins of 0.1 km s–1. The red, dashed curve is the result of a Gaussian fit to the histogram. Right-hand panel: Standard deviation of the velocity error of the sources as a function of true peak signal-to-noise ratio in bins of 1σ.
F7

The standard deviation of the radial velocity error as a function of peak signal-to-noise ratio in 1σ bins is shown in the right-hand panel of Figure 7. As expected, the standard deviation from the mean (which is essentially zero for all peak flux intervals) increases with decreasing peak flux. While for bright sources of Fpeak ≈ 20σ the standard deviation is below 1 km s–1, it increases to almost 6 km s–1 for faint sources near the 3σ level.

4.2.4 Line Width

Figure 8 shows the ratio of measured line width versus true line width (FWHM of the original Gaussian model) as a function of peak signal-to-noise ratio in bins of 1σ. duchamp determines three different types of line width: w50 is the full width at 50% of the peak flux, w20 is the full width at 20% of the peak flux, and wvel is the full detected line width of the source, i.e. the width across all channels with detected flux. For a Gaussian line, w50 is equivalent to the FWHM, and the ratio of FWHM/w50 should therefore be 1. The relation between w20 and w50 in the case of a Gaussian line is given by the constant factor of


Figure 8  Ratio of measured versus true line width for the point source models as a function of true peak signal-to-noise ratio in bins of 1σ. The black data points show w50, the red data points w20, and the blue data points wvel (for a 1.5σ flux threshold), all of which have been divided by the original FWHM of the Gaussian line. The corresponding theoretical expectations for a Gaussian line profile are shown as the black, red, and blue dashed lines.
F8

E1

Finally, the relation between wvel and w50, again assuming a Gaussian line profile, is defined via

E2

where Fthr = n ×σ is the flux threshold used in the calculation of wvel. These theoretical relations are plotted in Figure 8 as the dashed lines for w50 (black), w20 (red), and wvel (blue; for Fthr = 1.5σ).

duchamp's measurement of w50 (black data points) is in excellent agreement with the expectation (black, dashed line) over a wide range of peak signal-to-noise ratios. Only for faint sources of Fpeak < 5σ are the measured line widths on average slightly smaller than the true widths, but by no more than about 10 to 15%.

In contrast, duchamp's measurements of w20 and wvel (red and blue data points, respectively) are systematically too large over most of the covered range of signal-to-noise ratio as compared to the theoretical expectation (red and blue dashed lines, respectively). Only for faint sources of Fpeak 5σ do the measured w20 fall short of the theoretical ones. This result suggests that w50 is the most accurate measurement of line width provided by duchamp and should be used instead of w20 and wvel for the characterisation of astronomical sources. However, both w50 and w20 systematically fall short of the true line width for faint sources below Fpeak ≈ 5σ.

4.2.5 Peak Flux

The ratio of recovered versus true peak flux of the model point sources is plotted in the left-hand panel of Figure 9 as a function of peak signal-to-noise ratio in 1σ bins. The dashed and dotted red lines indicate the theoretical ±1σ and ±2σ envelopes, respectively. The right-hand panel shows the same figure, but as a function of integrated flux in bins of 0.1 Jy km s–1. For bright sources of Fpeak 10σ duchamp accurately recovers the peak flux of the sources, although there is the general tendency of measured peak fluxes being slightly too high on average. For fainter sources of Fpeak 5σ there is a strong deviation, with measured fluxes being systematically too high by a significant factor. This is generally due to faint sources being more likely to be detected when their maximum coincides with a positive noise peak, whereas faint sources sitting on top of a negative noise peak will likely remain undetected, thereby creating a strong bias in the measurement of peak fluxes.


Figure 9  Left-hand panel: Ratio of measured versus true peak flux (black data points) and corresponding standard deviation (error bars) of the model point sources as a function of true peak signal-to-noise ratio in 1σ bins. The dashed and dotted red lines indicate the theoretical ±1σ and ±2σ envelopes, respectively. Right-hand panel: Same, but as a function of true integrated flux in bins of 0.1 Jy km s–1.
F9

Even for high peak signal-to-noise ratios the peak fluxes measured by duchamp tend to be slightly too large. duchamp determines the peak flux of a source by simply selecting the data element with the highest flux encountered. As mentioned before, this method is biased towards selecting data elements that have been affected by positive noise peaks. In sources with broad spectral signals there is a higher probability of finding a positive noise signal in one of the channels near the peak of the line that increases the signal beyond the actual line peak. This is a result of the source being well resolved in the spectral domain. Hence, peak fluxes measured by duchamp will generally be too high irrespective of source brightness as long as the source is spectrally (or spatially) resolved. For very bright sources, however, the relative error will be negligible.

4.2.6 Integrated Flux

The ratio of measured versus true integrated flux of the model point sources as a function of peak signal-to-noise ratio in bins of 1σ is presented in the left-hand panel of Figure 10. The right-hand panel shows the same figure, but as a function of integrated flux in bins of 0.1 Jy km s–1. Apparently, duchamp's measurement of the integrated flux of a source is systematically too low by a significant factor. Even for the brightest sources of Fpeak ≈ 20σ only about 90% of the true flux is recovered by duchamp, and that figure drops to well below 50% for faint sources of Fpeak < 5σ.


Figure 10  Left-hand panel: Ratio of measured versus true integrated flux (black data points) and corresponding standard deviation (error bars) for the model point sources as a function of true peak signal-to-noise ratio in bins of 1σ. The solid red curve shows the theoretical expectation for the 1.5σ flux threshold used in our test. The dotted red curve shows the best fit to the data points, corresponding to an effective flux threshold of 2.2σ. Right-hand panel: Same, but as a function of true integrated flux in bins of 0.1 Jy km s–1.
F10

This issue is likely caused by the fact that duchamp only considers pixels above the detection threshold when calculating the integrated flux. Pixels below the threshold, while potentially contributing significantly to the overall flux of a source, are not included in the summation carried out by duchamp, resulting in integrated fluxes being systematically too small.

In order to study the expected decrease in the integrated flux measurement, let us assume a point source with Gaussian line profile being observed with a telescope with radially symmetric Gaussian point spread function (PSF),

E3

with amplitude, Fpeak, velocity dispersion, σv, and PSF size, σPSF. The integrated flux measurement can then be considered as the integral under the three-dimensional Gaussian brightness profile across the frequency/velocity range, ±v0, and the spatial range, ±x0 and ±y0, over which the flux of the line is above the detection threshold, thus

E4
E5

where erf(x) is the error function. Inserting the appropriate integration limits and then dividing Equation 5 by the total flux (i.e. integrated over ±∞) leads to a theoretical integrated flux ratio of

E6

with a flux threshold of Fthr = n ×σ.

The resulting theoretical integrated flux according to Equation 6, assuming a 1.5σ threshold, is shown as the solid red curve in Figure 10. The integrated fluxes measured by duchamp are only slightly below what one would expect from a simple integration over a three-dimensional Gaussian. A fit to the data points instead yields an effective flux threshold of 2.2σ, shown as the dotted red curve in Figure 10, which is slightly larger than the 1.5σ used when running duchamp. It is not quite clear why duchamp performs worse than expected. The discrepancy could be due to the fact that the software sums over discrete pixels whereas we assumed continuous integration in our mathematical model. This will likely result in small differences, particularly in those cases where the number of elements across the Gaussian profile is small. In our case, as we are dealing with point sources, this is certainly true for the spatial dimension.

In summary, integrated flux measurements provided by duchamp are systematically too small and will need to be corrected substantially to compensate for the systematic offset.


5 Models of Disc Galaxies

In order to test the performance of duchamp on more realistic, extended sources, we generated 1024 artificial Hi models of galaxies with a wide range of parameters, using a programme written in C for direct manipulation of FITS data cubes. All galaxies were modelled as infinitely thin discs with varying inclination (0 to 89°), position angle (0 to 180°), and rotation velocity (20 to 300 km s–1). For any one galaxy, inclination and position angle were considered to be constant over the radial extent of the disc, while the rotation velocity increases linearly from 0 to vrot between the centre and 0.5 times the semi-major axis of the disc and remains constant beyond that radius. Individual spectral profiles across the disc were assumed to be Gaussian with a dispersion of 9.65 km s–1 (equivalent to 2.5 times the width of a spectral channel). The radial surface brightness profile was assumed to be Gaussian, too, resulting in an elliptical Gaussian brightness distribution on the sky.

Next, we again generated an artificial ASKAP visibility data set of pure Gaussian noise at a frequency of 1.4 GHz with characteristics similar to the WALLABY survey, using the miriad task uvgen with parameters as listed in Table 4. The visibility data were Fourier-transformed to create an image of the point spread function and a noise data cube with 1601 × 1601 spatial pixels of 10 arcsec size and 201 spectral channels. We then convolved the model galaxies with a clean beam derived from fitting a Gaussian to the central peak of the point spread function. Finally, the convolved galaxy models were placed on a regular grid of 32 × 32 galaxies and added to the noise cube to create the final data cube of model galaxies for the testing of duchamp.


Table 4.  Summary of the parameters used to generate the visibility data set and noise image for the galaxy models
T4

The moment-zero map, position-velocity map, and integrated spectrum of one of the model galaxies is shown in Figure 11 for illustration. As with the point sources, all galaxies were centred on a pixel, although for extended sources we do not expect any significant effect from shifting the source centre with respect to the pixel centre. Again, all sources are isolated, and we did not attempt to test duchamp in a situation where source crowding occurs.


Figure 11  Example of a model galaxy generated for testing duchamp. The left-hand panel shows the zeroth moment of the model, the middle panel shows the position-velocity diagram along the dashed, red line, and the right-hand panel depicts the integrated spectrum of the model galaxy.
Click to zoom

It is important to note at this point that the resulting model galaxies, while exhibiting some of the spatial and spectral characteristics of real spiral galaxies, have been simplified to a great extent, resulting in limitations that need to be kept in mind when interpreting the results presented in this section. Firstly, the assumption of an infinitely thin disc will result in unrealistic edge-on galaxies, with integrated fluxes as well as individual spectral line widths across the disc being too small. Secondly, parameters such as peak flux, angular size, or rotation velocity were all varied independently of each other, resulting in unrealistic combinations of galaxy parameters in some cases. The purpose of the models is to cover a vast parameter range of extended sources irrespective of whether that entire range is populated by real galaxies. Even if disc galaxies with a certain combination of parameters do not exist, other objects, such as irregular galaxies or high-velocity clouds, could still cover those regions of parameter space, and their exploration will therefore be meaningful.

5.1 Running duchamp

We ran duchamp (version 1.1.12) on the model galaxy cube several times with slightly different input parameters to compare the performance. The different input parameters explicitly set in the parameter file are listed and compared in Table 5. In all cases we employed a 1σ flux threshold, equivalent to about 1.9 mJy, and performed a three-dimensional `à trous’ wavelet reconstruction with a minimum scale of 3 and a flux threshold of 2σ for wavelet components to be included in the reconstructed cube. The slightly larger minimum scale as compared to the point source models is motivated by the fact that we are now dealing with spatially and spectrally much more extended sources. In addition, we varied the number of contiguous spectral channels required for detections and used duchamp's growth criterion in a few runs with a growth flux threshold of 0.5σ. The latter method will grow detections to flux levels below the original detection threshold, resulting in more accurate source parametrisation. As it turned out, the change from 5 to 3 consecutive spectral channels for detections (run 2 versus 3) did not have any major impact on the results. Hence, only the results of runs 1 and 3 will be presented and discussed here.


Table 5.  duchamp input parameters explicitly set in the input-parameter file for the galaxy modelsa
Click to zoom

In order to compare the outcome of duchamp with the original input catalogue, we wrote a short Python script that reads in and processes the different catalogues. The script first reads in the duchamp output catalogue, the original model catalogue, and a special mask data cube marking pixels with emission in the original model by assigning them a unique number characteristic to each input source. The script then cycles through all the detections made by duchamp and decides for each detection whether it is genuine or not by checking the value of the mask data cube at the same position. If the detection is found to be genuine, the script will cycle through the original model catalogue to extract the actual input parameters of the respective source for comparison with the parametrisation results of duchamp.

At the end of this process we get a match of detected sources with original input sources, allowing us to calculate parameters such as completeness, reliability, and the fraction of sources being broken up into multiple detections by duchamp. In addition, we are able to compare the original parameters of each source with those determined by duchamp to test the performance of duchamp's parametrisation algorithms.

5.2 Results

5.2.1 Completeness and Reliability

For run 1 (without growing of detections to flux levels below the threshold), 436 out of 1063 detected sources are genuine, resulting in an overall reliability of 41%. As many original sources got broken up into multiple detections, only 194 of the 1024 input galaxies were detected, yielding an overall completeness of only 19%. There is a significant improvement for run 3 (with growing of detections to a flux level of 0.5σ), where 542 out of 1051 detected sources are genuine (reliability of 52%), but this time 521 of the 1024 input galaxies were detected, resulting in a much improved overall completeness of 51%.

Completeness as a function of different galaxy parameters is shown in Figure 12 for runs 1 and 3 (black and red data points, respectively). As mentioned before, run 1 resulted in very low completeness values of typically only about 20% and no strong variation with either the integrated flux of a source or its inclination and rotation velocity. By growing detections to a flux level of 0.5σ (run 3) we achieved much higher completeness levels over a large parameter range. 100% completeness is achieved for sources of Fint 20 Jy km s–1, and completeness levels reach 50% at Fint ≈ 2.5 Jy km s–1. The latter corresponds to an Hi mass sensitivity of 6 × 105M at a distance of 1 Mpc, or 6 × 109M at 100 Mpc, for the expected 8-hour integration per pointing of the WALLABY project on ASKAP.


Figure 12  Completeness for the galaxy models as a function of true integrated flux in bins of 2.5 Jy km s–1 (top panel), galaxy inclination in bins of 5° (middle panel), and rotation velocity in bins of 19.3 km s–1 (bottom panel) for runs 1 (black) and 3 (red).
F12

As shown in the middle and bottom panels of Figure 12, there is a strong variation of completeness with both inclination and rotation velocity of the galaxies. While face-on galaxies are on average detected at completeness levels near 80%, duchamp struggles to find edge-on galaxies, yielding average completeness levels of only about 20% for galaxies with inclination angles greater than 80°. This effect is caused by the combination of two separate effects. Firstly, as a result of the limitations from our assumption of an infinitely thin disc, edge-on galaxies have typically lower integrated fluxes than face-on galaxies. Secondly, edge-on galaxies typically have a broader spectral signature as a result of their higher projected rotation velocity, making it more difficult for duchamp to pick up their extended signal.

The latter effect can also be seen in the bottom panel of Figure 12, where completeness levels systematically decrease as a function of increasing rotation velocity of a galaxy, irrespective of its inclination or integrated flux, confirming that on average duchamp is more likely to pick up face-on galaxies with narrow spectral lines. It is important to note, however, that at a given distance galaxies with higher rotation velocity will typically have a larger Hi mass and are therefore more likely to be detected than galaxies with lower rotation velocity at the same distance.

5.2.2 Break-Up of Sources into Multiple Components

Due to their rotation velocity, spiral galaxies often exhibit a large radial velocity gradient across their projected disc on the sky, resulting in the characteristic double-horn profile of their integrated spectrum. This, however, can result in the two halves of a galaxy being detected as two separate sources by duchamp, in particular in the case of faint, edge-on galaxies with large rotation velocities.

In Figure 13 we have plotted the fraction of detected model galaxies that were broken up into two or more separate detections by duchamp as a function of true integrated flux (top panel), inclination (middle panel), and rotation velocity (bottom panel). For run 1 (black data points) there is a very high fraction of multiple detections, typically about 60--80%, with no strong variation with either integrated flux of the galaxy or inclination and rotation velocity of the disc. In total, 136 out of the 194 detected galaxies, or 70.1%, were broken up into multiple components by duchamp.


Figure 13  Fraction of model galaxies being broken up into two or more separate detections by duchamp as a function of true integrated flux in bins of 2.5 Jy km s–1 (top panel), galaxy inclination in bins of 5° (middle panel), and rotation velocity in bins of 19.3 km s–1 (bottom panel) for runs 1 (black) and 3 (red).
F13

Growing detections to the 0.5σ level (run 3, red data points) results in a major improvement, with the number of multiple detections (in total 21 out of 521 detected galaxies, or 4.0%) dropping to zero over most of the covered parameter range. Only for faint sources of Fint 5 Jy km s–1 does the fraction of multiple detections gradually increase up to about 10% at the low end of the flux spectrum. Figure 13 also clearly shows the expected increase in multiple detections for galaxies of higher inclination (i 40°) and rotation velocity (vrot 150 km s–1), which is the result of the double-horn profile becoming wider and more pronounced as the radial velocity gradient in the plane of the sky increases.

A similar case, although more difficult to assess, is the detection of only one half of a galaxy (one horn of the double-horn profile), whereas the other half remains undetected. As there is only a single detection of each affected galaxy, such partial detections are much more difficult to identify. They should, however, result in a significant offset of both the measured position and radial velocity of the detected source with respect to the location of the originating galaxy.

In the case of run 3, 62 out of 500 single detections show velocity errors of more than 20 km s–1, with 28 even exceeding 150 km s–1. The former corresponds to a fraction of 12.4% of all single detections. Similarly, 62 out of 500 singly detected sources have a position error of more than 20 arcsec, which again corresponds to a fraction of 12.4%.5

These results suggest that, even when growing detections down to the 0.5σ level, there is a significant number of partial (approximately 66 sources) or multiple (21 sources) detections, corresponding to an overall fraction of about 16.7% of all genuine detections. Such cases need to be identified in the output catalogue produced by duchamp, as otherwise they will introduce a significant bias in the measurement of source parameters such as line width and Hi mass. Identification of broken-up sources will be a very difficult task in practice, as it may be impossible to decide whether two detections are part of the same source or two separate sources in close proximity. While the growing of detections to lower flux levels can in principle reduce the fraction of sources being broken up, an undesirable side effect will be the potential merging of neighbouring sources, e.g. close galaxy pairs in group or cluster environments.

5.2.3 Source Position

The left-hand panel of Figure 14 shows a scatter plot of position errors for the model galaxies (based on run 3) in right ascension and declination. The mean position errors in right ascension and declination are 1.7 ± 14.7 arcsec and 0.6 ± 12.7 arcsec, respectively. The standard deviation is fairly large because there are several sources with position errors of tens of arcsec, well beyond the central concentration in the plot. These are cases in which only one half of a galaxy was detected as a source, whereas the other half remained undetected, resulting in systematic offsets in position as well as velocity with respect to the original model.


Figure 14  Left-hand panel: Position errors (from run 3) for the model galaxies in right ascension and declination. Right-hand panel: Mean absolute position error (data points) and standard deviation (error bars) as a function of true integrated flux in bins of 2.5 Jy km s–1.
F14

When excluding such cases of partial detections by only considering detections with position errors of less than 15 arcsec in both right ascension and declination, we obtain corrected errors of 0.9 ± 3.6 arcsec in right ascension and 0.5 ± 3.6 arcsec in declination.

The combined, absolute position error as a function of true integrated flux is shown in the right-hand panel of Figure 14. For bright sources of Fint 10 Jy km s–1 source positions are very accurate with typical errors of about 2.5 arcsec. Towards the faint end of the diagram both mean error and standard deviation increase substantially, partly as a result of increasing statistical uncertainties, but also due to an increasing fraction of galaxies that are only partially detected.

5.2.4 Radial Velocity

The mean velocity error (based on run 3) for the galaxy models is –1.9 ± 54.5 km s–1. As in the case of source position, the large standard deviation about the mean is caused by galaxies that are only partially detected. By including only sources with position errors of less than 15 arcsec in both right ascension and declination and velocity errors of less than 20 km s–1 we can exclude such partial detections, resulting in a corrected mean radial velocity error of –0.8 ± 4.6 km s–1.

A histogram of radial velocity errors for the galaxy models is shown in Figure 15. As in the case of point sources, the distribution is not exactly Gaussian. Instead, there is a sharp peak near zero and an underlying broad distribution of errors, in particular in the negative range. Some of these non-Gaussian structures could again be the result of digitisation effects in conjunction with the spectral channel width of 3.86 km s–1, while we have no conclusive explanation for the noticeable asymmetry of the distribution.


Figure 15  Histogram of radial velocity errors (from run 3) for the model galaxies in bins of 0.5 km s–1.
F15

5.2.5 Line Width

In order to estimate the original line width of the input models, we calculated a `pseudo line width’ which balances the intrinsic width of an individual line profile with the overall, integrated line width resulting from the rotation velocity of the galaxy, thus

E7

where vrot is the rotation velocity of the model galaxy, i is the inclination of the disc, and wint = 22.7 km s–1 is the intrinsic FWHM of the Gaussian spectral line at each position across the galaxy.

The left-hand panel of Figure 16 shows the mean ratio of the measured line width, w50, over the calculated `pseudo line width', wmod, as a function of true integrated flux in bins of 2.5 Jy km s–1 (based on run 3). duchamp measures accurate line widths close to the true value over a wide range of fluxes. The small deviation from the value of 1 can be easily explained by the fact that wmod is just an approximation to the FWHM of the line profile. Only for fainter sources of Fint 5 Jy km s–1 does the line width ratio decrease and the standard deviation increase significantly, indicating larger errors in duchamp's measurement of line width.


Figure 16  Ratio of measured (w50, from run 3) versus true line width for the model galaxies as a function of true integrated flux in bins of 2.5 Jy km s–1 (left-hand panel) and true line width in bins of 50 km s–1 (right-hand panel). The error bars indicate the standard deviation about the mean.
Click to zoom

In the right-hand panel of Figure 16 we have plotted the ratio of w50/wmod as a function of wmod in bins of 50 km s–1. While line width measurements for sources with narrow lines of wmod 250 km s–1 are on average accurate, there is a systematic discrepancy for sources with broader lines, the line widths measured by duchamp being systematically too small. The large standard deviation suggests that this could have been caused by cases in which only one half of the galaxy was detected, whereas the other half remained undetected, resulting in a significantly lower value of the measured line width. Nevertheless, line width measurements for fully-detected sources should be accurate even if their line widths are large. This problem again demonstrates the need to identify partially detected sources to avoid systematic errors that would affect the scientific interpretation of the data.

5.2.6 Integrated Flux

The ratio of measured versus true integrated flux of the model galaxies, based on run 3, is shown in Figure 17. Similar to our previous tests on point sources (see Figure 10), the integrated flux measured by duchamp is systematically too low. For bright sources of Fint ≈20 Jy km s–1 a large fraction of approximately 95% of the flux is recovered, whereas this figure drops to below 60% for fainter sources of Fint 2 Jy km s–1. At the same time, the scatter significantly increases, suggesting larger uncertainties (on a relative scale) in the flux measurement of faint sources.


Figure 17  Ratio of measured (from run 3) versus true integrated flux of the model galaxies as a function of true integrated flux. The error bars indicate the standard deviation about the mean. The bin width decreases towards lower fluxes.
F17

As discussed previously, the reason for the failure of duchamp to accurately determine the integrated flux of a source is that the software only sums over data elements that are above the flux threshold and hence misses some of the flux. Even the growth of detections down to the 0.5σ level has not solved this fundamental problem, although the defect has become less severe than for the point source models without growing (see the right-hand panel of Figure 10 for comparison).


6 Model Cube Based on Real Galaxies

So far, we have tested duchamp on artificial sources embedded in perfectly Gaussian noise. While this is useful to study the basic performance of the software, real observations will be more challenging for any source finder due to the more complex morphology of real sources and the presence of various artefacts in the data, e.g. terrestrial and solar interference, spectral baseline instabilities, or residual continuum emission.

Unfortunately, it is not possible to simply test duchamp on a real Hi data cube, because we would not have a-priori knowledge of the sources in such a cube and would not be able to assess which of the detections made by duchamp are genuine. A solution to this problem would be to inject copies of real galaxies into a real data cube of `pure’ noise, i.e. a data cube extracted from telescopic observations that does not contain any Hi sources above the noise level. This method combines the advantages of artificial source models, where the source locations and parameters are exactly known, with those of real observations with realistic sources and artefacts.

For this purpose, we generated a data cube containing real noise extracted from an observation with the Westerbork Synthesis Radio Telescope (WSRT). We then added about 100 data cubes from the ‘Westerbork Observations of Neutral Hydrogen in Irregular and Spiral Galaxies’ (WHISP) survey (Kamphuis, Sijbring, & van Albada 1996; Swaters et al. 2002), each containing one or more galaxies. The selected WHISP data cube were artificially redshifted by scaling their size and flux level to match sources in a redshift range of 0.02 z 0.04, centred on the median redshift of 0.03 expected for the WALLABY project (Koribalski & Staveley-Smith 2009). The procedure for creating the test data cube is explained in more detail by Serra, Jurek, & Flöer (2012).

The final test data cube has a size of 360 × 360 spatial pixels and 1464 spectral channels. The pixel size of 10 arcsec (with a synthesised beam width of 30 arcsec) and channel width of 18.3 kHz (equivalent to about 4 km s–1) were chosen to reflect the expected specifications of WALLABY. Figure 18 shows an example image and spectra of two of the galaxies in the final cube. As the locations and properties of the injected galaxies are well-known, we can directly compare them to the output of duchamp to assess performance indicators such as completeness and reliability.


Figure 18  Left-hand panel: Moment-zero map of a small region of the WSRT cube with injected WHISP galaxies, showing two galaxies labelled A and B. Right-hand panels: Integrated spectra of the two galaxies.
Click to zoom

6.1 Running duchamp

We ran duchamp multiple times on the WSRT data cube with WHISP galaxies to probe different input parameter settings of duchamp. A summary of the runs and parameters used is given in Table 6. We mainly covered a wide range of flux thresholds between 1.0 and 3.5σ and tested one-dimensional (spectral domain only) versus three-dimensional (spatial and spectral domain) wavelet reconstruction of the cube. The output catalogue of each run was again cross-matched with the original source catalogue, using the Python script described in Section 5.1.


Table 6.  Relevant duchamp input parameters for the model based on real WHISP galaxies and WSRT noisea
Click to zoom

In order to obtain the original source catalogue, we ran duchamp once on the input model cube without noise, using a very low detection threshold of well below the final noise level and no wavelet reconstruction. This resulted in a list of 100 sources against which the output catalogue provided by duchamp can be judged. Since this method already introduces a strong bias in the catalogue of source parameters, we will only analyse the completeness and reliability of duchamp, but we shall not attempt to assess the parametrisation performance of the software, because we do not have an exact source catalogue against which we would be able to assess the source parameters as measured by duchamp from the final test cube.

6.2 Results

Completeness and reliability of the different runs of duchamp on the WSRT model cube with WHISP galaxies are listed in Table 6 and displayed in Figure 19 as a function of detection threshold. Generally, between about 40% to 60% of all galaxies in the cube were found by duchamp, while the overall reliability varies strongly from about 10% to 100% depending on detection threshold and wavelet reconstruction parameters.


Figure 19  Completeness (filled circles, solid lines) and reliability (open circles, dashed lines) of duchamp on the WSRT model cube with WHISP galaxies for different flux thresholds and input parameters. The colours, as shown in the legend, distinguish the different wavelet reconstruction modes (one-dimensional versus three-dimensional) and wavelet reconstruction thresholds (3σ versus 4σ) used in the tests (see Table 6 for details). The numbers alongside the data points refer to the corresponding runs as listed in Table 6.
F19

We achieve better results for one-dimensional wavelet reconstruction (black and blue data points in Figure 19) which generally yields higher completeness and reliability than three-dimensional wavelet reconstruction (red data points). This is presumably due to the small angular size of most galaxies in the model cube; there is not much to gain from performing a wavelet reconstruction in the spatial domain, whereas one-dimensional wavelet reconstruction in the frequency domain yields much better results because most galaxies are well-resolved and extended in frequency.

In Figure 20 we plot completeness as a function of integrated flux for selected runs of duchamp. Above a flux of Fint 3 Jy km s–1 duchamp consistently finds all sources irrespective of the input parameters chosen. At lower fluxes the different runs produce significantly different results, with the one-dimensional wavelet reconstruction (black and blue data points) generally performing better than the three-dimensional reconstruction (red data points), as noted before.


Figure 20  Completeness as a function of integrated flux for selected runs (see legend) of duchamp on the WSRT model cube with WHISP galaxies. The choice of colours is the same as in Figure 19.
F20

The best-performing parameter set in terms of completeness, run 7, produces a completeness of 50% at an integrated flux of Fint ≈ 0.7 Jy km s–1, corresponding to an Hi mass of 1.7 × 105M at a distance of 1 Mpc, or 1.7 × 109M at 100 Mpc. This is worse than what we achieved for the point sources with Gaussian line profiles in Section 4, but significantly better than the outcome for the model galaxies in Section 5. The reason for the better performance could be that the artificially redshifted WHISP galaxies are generally much more compact than the model galaxies created for the tests in Section 5. As with any threshold-based source finder, compact sources are easier to detect than extended sources, even with prior wavelet reconstruction or smoothing. As the spectral profiles of the WHISP galaxies are generally broad and complex, performance is worse than in the case of the point source models in Section 4 which had much simpler and narrower Gaussian lines.

The overall reliability in the case of run 7 is very low with only 10%. This figure, however, is the raw reliability achieved by duchamp and can be substantially improved by filtering sources based on their measured parameters. False detections are usually the result of noise peaks being picked up by the source finder. A large fraction of these false noise detections will be characterised by very low integrated fluxes and small line widths, and often a simple cut in flux--line width space will remove more than 95% of false detections while retaining more than 95% of genuine detections. This fact is illustrated and discussed in more detail in Section 7.1.

In summary, when running duchamp on a realistic data cube with real galaxies at a redshift of about 0.03 and genuine noise extracted from observational data taken with the WSRT, the software performs as expected with completeness levels ranging in between those achieved for the compact and extended model sources discussed in the previous sections. This result illustrates that the performance of duchamp, as with any source finder based on flux thresholding, will strongly depend on the morphology and extent of the sources to be detected. Even with multi-scale wavelet reconstruction, duchamp is more likely to uncover compact sources than sources that are significantly extended, either spatially or spectrally.

At the same time, the performance of duchamp does not seem to be hampered by the fact that we are dealing with real telescope data and noise, as the completeness and reliability levels reported in Table 6 are generally very similar to what we achieved with the model sources discussed in the previous sections. This is presumably due to the excellent quality of the Westerbork data which do not contain any obvious artefacts such as interference or residual continuum emission.


7 Discussion

In general, duchamp does what it promises to do. It is able to reliably detect sources down to low signal-to-noise ratios and accurately determine their position and radial velocity. These are the most fundamental requirements for any source finder. Our tests also demonstrated that by using and fine-tuning the options of ‘à trous’ wavelet reconstruction and growing of sources to lower flux levels the performance of duchamp can be greatly enhanced.

7.1 Improving Reliability

The reliability figures reported throughout this paper have all been `raw’ reliabilities, i.e. reliabilities as achieved by duchamp prior to any filtering of the output source catalogue. The user would normally wish to substantially improve these through appropriate filtering of the source catalogue based on the source parameters as measured by duchamp.

The left-hand panel of Figure 21 shows the measured integrated flux plotted against measured line width for all genuine (black data points) and false (red data points) detections found by duchamp in the point source models discussed in Section 4. It is obvious that genuine and false detections occupy largely disjunct regions of Fintw50 parameter space, with false detections generally occurring near the low end of the integrated flux spectrum. Similar plots can be generated for other combinations of source parameters, but Fint and w50 usually provide the best distinction between genuine and false detections.


Figure 21  Measured integrated flux, Fint, versus measured line width, w50, of all genuine (black) and false (red) detections made by duchamp in the point source models with Gaussian line profiles (left) and the test cube with artificially redshifted WHISP galaxies (right). The dashed, black lines indicate the flux levels of 0.04 and 0.5 Jy km s–1 used to filter false detections.
Click to zoom

The easiest way to improve the reliability of duchamp's source finding results is to simply apply a cut in Fint to exclude most false detections while retaining most of the genuine sources. In our example, applying a cut at Fint = 40 mJy km s–1 will discard 97.2% of all false detections while at the same time retaining 96.9% of all genuine sources, thereby increasing the overall reliability from 77.1% to 99.2% while only moderately decreasing the overall completeness from 83.0% to 80.6%.

A similar cut can be applied to the results from run 7 on the test data cube containing artificially redshifted WHISP galaxies, as plotted in the right-hand panel of Figure 21. Again, applying a simple flux threshold of 0.5 Jy km s–1 will improve reliability from 10% to 84%, while only moderately decreasing completeness from 59% to 50%. The method is not quite as successful as for the point sources, as we are now dealing with real galaxies and real noise with interference and artefacts, but nevertheless a significant improvement in reliability can be achieved without any severe impact on the number of genuine detections.

This simple example illustrates that the `raw’ reliability figures quoted throughout this paper should not be considered as the final numbers. Reliability can be greatly improved through very basic filtering in parameter space of the duchamp output catalogue. In principle, this applies to the output of almost any source finder. Alternatively, instead of removing sources from the output catalogue, it may be desirable to calculate a reliability number for each catalogue entry based on the source's location in parameter space and leave it to the catalogue's users to decide as part of their scientific analysis at which reliability level they wish to make the cut.

7.2 Source Parametrisation Issues

When it comes to source parametrisation, the measurements provided by duchamp are affected by several systematic errors. These systematic errors are not due to errors in the software itself, but a consequence of the presence of noise in the data as well as the methods and algorithms used for measuring source parameters.

Spectral line widths determined by duchamp are generally very accurate and not much affected by noise-induced, systematic errors as far as the w50 parameter is concerned. The two other line width parameters calculated by duchamp, w20 and wvel, appear to be systematically too large over a wide range of signal-to-noise ratios and should not be used unless explicitly required in special, well-defined circumstances.

Peak fluxes, as reported by duchamp, are in general slightly too large for bright sources and significantly too large (on a relative scale) for faint sources. This is due to the fact that duchamp determines the peak flux by simply selecting the value of the brightest pixel encountered. This method introduces a bias towards positive noise peaks sitting on top of the brightest region of a source, and hence, in the presence of noise, peak fluxes measured by duchamp will be systematically too high.

Integrated fluxes determined by duchamp are significantly and systematically too small, in particular for faint sources. This is likely caused by the fact that duchamp simply sums over the flux of discrete elements above a given threshold to determine the integrated flux, thereby missing some of the flux from elements below the flux threshold. Hence, the raw integrated flux measurements currently provided by duchamp are not useful and need to be corrected to compensate for the systematic offset. This issue is particularly sensitive as many scientific projects, including the ASKAP survey science projects WALLABY and DINGO6 (Meyer 2009), rely on accurate flux measurements, for example for determining the Hi mass function of galaxies.

Finally, a particular problem in the case of galaxies is that under certain circumstances galaxies either get broken up into multiple detections or only one half of a galaxy is detected. This problem mainly affects faint, edge-on galaxies with broad spectral profiles that are partly hidden in the noise and results in systematic errors in the measurements of essentially all source parameters, including basic parameters such as position and radial velocity. Such cases of multiple or partial detections must be identified and treated separately to prevent biases in any scientific analysis based on the source finding results.


8 Summary

In this paper we present and discuss the results of basic, three-dimensional source finding tests with duchamp, the standard source finder for the Australian SKA Pathfinder, using different sets of unresolved and extended Hi model sources as well as a data set of real galaxies and noise obtained from Hi observations with the WSRT.

Overall, duchamp appears to be a successful, general-purpose source finder capable of reliably detecting sources down to low signal-to-noise ratios and accurately determining their position and velocity. In the case of point sources with simple Gaussian spectral lines we achieve a completeness of about 50% at a peak signal-to-noise ratio of 3 and an integrated flux level of about 0.1 Jy km s–1. The latter corresponds to an Hi mass sensitivity of about 2 × 108M at a distance of 100 Mpc which is slightly better than what the WALLABY project is expected to achieve for real galaxies (Koribalski & Staveley-Smith 2009). The situation is less ideal for extended sources with double-horn profiles. In this case we achieve 50% completeness at an integrated flux level of about 2.5 Jy km s–1 for the model galaxies and 0.7 Jy km s–1 for the WHISP galaxies. The latter is equivalent to an Hi mass sensitivity of about 1.7 × 109M at a distance of 100 Mpc, illustrating that the performance of duchamp, as well as any other source finder, will strongly depend on source morphology. However, these figures may well be improved by carefully optimising the various input parameters offered by duchamp.

In its current state duchamp is not particularly successful in parametrising sources in the presence of noise in the data cube, and other, external algorithms for source parametrisation should be considered instead. It appears, however, that most, if not all, parametrisation issues are due to intrinsic limitations in the implemented algorithms themselves and not due to errors in their implementation, suggesting that most of the problems can in principle be solved by implementing more sophisticated parametrisation algorithms in duchamp. Alternatively, corrections would have to be applied to all parameters derived by duchamp to compensate for systematic errors. Such corrections, however, would have to be highly specialised and tailored to the particular survey and source type concerned.



Acknowledgments

We wish to thank Matthew Whiting for creating the duchamp source finder and for having had an open ear for our comments and suggestions regarding improvement of the software.


References

Barnes, D. G., 2001, MNRAS, 322, 486
Crossref | GoogleScholarGoogle Scholar |

DeBoer, D. R., 2009, IEEEP, 97, 1507
Crossref | GoogleScholarGoogle Scholar |

Dewdney, P. E., Hall, P. J., Schilizzi, R. T. and Lazio, T. J. L., 2009, IEEEP, 97, 1482
Crossref | GoogleScholarGoogle Scholar |

Giovanelli, R., 2005, AJ, 130, 2598
| 1:CAS:528:DC%2BD28XjvFCnsQ%3D%3D&md5=77eef773f102562371a62f14fcf6bb7eCAS |

Jonas, J. L., 2009, IEEEP, 97, 1522
Crossref | GoogleScholarGoogle Scholar |

Kamphuis, J. J., Sijbring, D. and van Albada, T. S., 1996, A&AS, 116, 15
| 1:CAS:528:DyaK28XislKmu7w%3D&md5=63f62c8f6291edce74db3bb1d4fa410dCAS |

Kilborn, V. A., 2001, The Distribution of Neutral Hydrogen in the Local Universe, PhD Thesis, University of Melbourne

Koribalski, B. S. & Staveley-Smith, L., 2009, WALLABY Proposal Abstract, available at http://www.atnf.csiro.au/research/WALLABY/proposal.html

Lang, R. H., 2003, MNRAS, 342, 738
Crossref | GoogleScholarGoogle Scholar | 1:CAS:528:DC%2BD3sXlvFeis7Y%3D&md5=d202651020d31ddd97047963455d9743CAS |

Meyer, M., 2009, Proc. Sci., PoS(PRA2009)015

Meyer, M. et al., 2004, MNRAS, 350, 1195
Crossref | GoogleScholarGoogle Scholar | 1:CAS:528:DC%2BD2cXls1Sjurc%3D&md5=a45ea5e988dc074b2b07256ef6e8123aCAS |

Oosterloo, T., Verheijen, M. A. W., van Cappellen, W., Bakker, L., Heald, G. & Ivashina, M., 2009, Proc. Sci., PoS(SKADS 2009)070

Popping, A., Jurek, R., Westmeier, T., Serra, P., Flöer, L., Meyer, M. and Koribalski, B., 2012, PASA, 29, 318
Crossref | GoogleScholarGoogle Scholar |

Saintonge, A., 2007, AJ, 133, 2087

Sault, R. J., Teuben, P. J. and Wright, M. C. H., 1995, ASPC, 77, 433

Serra, P., Jurek, R. and Flöer, L., 2012, PASA, 29, 296
Crossref | GoogleScholarGoogle Scholar |

Starck, J.-L. & Murtagh, F., 2002, Astronomical Image and Data Analysis (Berlin: Springer)

Swaters, R. A., van Albada, T. S., van der Hulst, J. M. and Sancisi, R., 2002, A&A, 390, 829

Westmeier, T. & Johnston, S., 2010, Proc. Sci., PoS(ISKAF2010)056

Whiting, M. T., 2011, duchamp User Guide Version 1.1.13, available at http://www.atnf.csiro.au/people/Matthew.Whiting/duchamp/downloads/UserGuide-1.1.13.pdf

Whiting, M. T., 2012, MNRAS, submitted

Whiting, M. T. and Humphreys, B., 2012, PASA, 29, 371
Crossref | GoogleScholarGoogle Scholar |




1 duchamp website: http://www.atnf.csiro.au/people/Matthew.Whiting/duchamp/

2 Principal investigators: Bärbel Koribalski and Lister Staveley-Smith,http://www.atnf.csiro.au/research/WALLABY/

3 In her calculation of S/N, Saintonge (2007) makes the implicit assumption that the sources are spatially unresolved.

4 Note that these numbers differ slightly from the ones quoted for run 5 in Figure 3 because a different realisation of the model was used in the initial tests. Reliability values will generally depend on the characteristics of the data cube under consideration (e.g. the size of the cube) and are therefore difficult to assess and compare.

5 There is no exact match between the 62 sources with large position error and the 62 sources with large velocity error. A total of 66 sources fulfil either of the two criteria.

6 Deep Investigation of Neutral Gas Origins; principal investigator: Martin Meyer; public website: http://internal.physics.uwa.edu.au/~mmeyer/dingo/