Issue 
A&A
Volume 576, April 2015



Article Number  A8  
Number of page(s)  11  
Section  Numerical methods and codes  
DOI  https://doi.org/10.1051/00046361/201425259  
Published online  13 March 2015 
ADAM: a general method for using various data types in asteroid reconstruction^{⋆}
^{1}
Department of MathematicsTampere University of Technology,
PO Box 553,
33101
Tampere,
Finland
email:
matti.viikinkoski@tut.fi
^{2}
Astronomical Institute, Faculty of Mathematics and Physics,
Charles University in Prague, V
Holešovičkách 2, 18000
Prague, Czech
Republic
Received: 1 November 2014
Accepted: 26 January 2015
We introduce ADAM, the AllData Asteroid Modelling algorithm. ADAM is simple and universal since it handles all diskresolved data types (adaptive optics or other images, interferometry, and rangeDoppler radar data) in a uniform manner via the 2D Fourier transform, enabling fast convergence in model optimization. The resolved data can be combined with diskintegrated data (photometry). In the reconstruction process, the difference between each data type is only a few code lines defining the particular generalized projection from 3D onto a 2D image plane. Occultation timings can be included as sparse silhouettes, and thermal infrared data are efficiently handled with an approximate algorithm that is sufficient in practice because of the dominance of the highcontrast (boundary) pixels over the lowcontrast (interior) pixels. This is of particular importance to the raw ALMA data that can be directly handled by ADAM without having to construct the standard image. We study the reliability of the inversion, using the independent shape supports of function series and controlpoint surfaces. When other data are lacking, one can carry out fast nonconvex lightcurveonly inversions, but any shape models resulting from it should only be taken as illustrative largescale models.
Key words: minor planets, asteroids: individual: 2000 ET70 / methods: analytical / minor planets, asteroids: general / methods: numerical / minor planets, asteroids: individual: Daphne
The code is only available at the CDS via anonymous ftp to cdsarc.ustrasbg.fr (130.79.128.5) or via http://cdsarc.ustrasbg.fr/vizbin/qcat?J/A+A/576/A8
© ESO, 2015
1. Introduction
Groundbased and other remotesensing data on asteroids are obtained with a variety of instruments that essentially sample regions on the surface of the target in various ways. These share some common mathematical characteristics of generalized projections (Kaasalainen & Lamberg, 2006; Kaasalainen, 2011; Viikinkoski & Kaasalainen, 2014). The most abundant source of data for asteroid shape and spin reconstruction is diskintegrated photometry, because even data that are sparse in time are often sufficient for modelling (Kaasalainen, 2004; Ďurech et al., 2006). Lightcurveinversion procedures (Kaasalainen et al., 2001) are available at e.g. the Database of Asteroid Models from Inversion Techniques (DAMIT) site^{1}. Because of the inherently limited information content of the diskintegrated data, the corresponding models are usually most reliably described in convex space (Ďurech & Kaasalainen 2003, and further discussed below). However, even partially diskresolved data offer a realistic possibility of more detailed modelling. Previously described approaches for such reconstruction are the SHAPE software (Ostro et al., 2002) for radar and lightcurve data, and the KOALA procedure (Kaasalainen & Viikinkoski, 2012; Carry et al., 2012) for optical images, occultation timings, and lightcurves.
The best way to reconstruct a model of an asteroid is to use all available data. To combine diskresolved data (adaptive optics or other images, interferometry, and rangeDoppler radar data) with diskintegrated data (photometric or infrared lightcurves) and occultation timings (“sparse silhouettes”), we need a general procedure for using any data sources in asteroid modelling. We call this ADAM: allData Asteroid Modelling. Concise accounts of the various data types and their modelling aspects are given in Kaasalainen & Lamberg (2006), Kaasalainen & Ďurech (2013), and Ďurech et al. (2015). This paper is intended as a technical companion to those reviews.
We present the ADAM algorithm in a highlevel format that includes all the necessary methods and formulae, either written here or given by references to the literature. We discuss and collect the essential techniques and aspects of a complete inversion procedure capable of handling all the major asteroid data sources and formats. The key point is that complementary data sources can facilitate a good reconstruction even when none of them is sufficient alone.
The paper is organized as follows. In Sect. 2 we describe the various shape supports we use in the reconstruction; some with the emphasis on global features, some concentrating on local details. This is intimately connected with the reliability estimate of the result, since independent shape representations help to reveal the most probable features. Section 3 introduces the Fourier transform method necessary for a simple and universal handling of data sources of diskresolved type. In Sect. 4, we present examples of these types (interferometry, radar, and optical images). The interferometric data from ALMA are of particular interest. We also discuss the special case of onedimensional projections (continuouswave radar and certain types of interferometry). In Sect. 5 we sum up everything in the form of the ADAM algorithm, and conclude in Sect. 6. Some basic ADAM functions are listed in Appendix A.
Using the methods and algorithm described here and in Kaasalainen et al. (2001), Kaasalainen (2011), and Kaasalainen & Viikinkoski (2012), writing an ADAM program from scratch is quite straightforward (for example, convex lightcurve inversion is inherently more complex). We have uploaded free ADAM code files and functions written in Matlab and C to a toolbox at the DAMIT site. These can be used for writing customized inversion software, and for browsing and understanding the computational methods. These methods, too numerous to be discussed here in detail, include techniques, such as the partial derivative chains for gradientbased optimization, raytracing procedures, projections and transforms, scattering and luminosity models, GPU acceleration, etc. (Note that we do not offer any user support: the files are presented as is.)
The ADAM procedure is a considerably more general package than the KOALA (Kaasalainen & Viikinkoski, 2012; Carry et al., 2012), which is based on extractable image contours. The KOALA contourfitting principle is necessary for including occultation data, so a full ADAM procedure inherits this function from KOALA. For fitting any pixel images, we recommend the ADAM Fouriertransform functions rather than KOALA.
We take asteroid reconstruction to mean here that the following output parameters are derived from input data: 1) shape (surface) definition; 2) rotational state (period and spin axis direction; possibly also terms for YORP acceleration, precession, or a binary orbit); 3) scattering or other luminosity parameters (often fixed a priori); and 4) image offset (alignment) and possible other auxiliary or normalization terms. Without the loss of generality, we do not discuss each item separately, but mostly take the shape parameters to represent all the free parameters since the optimization principle is technically the same for all parameter types. The speed, convergence, and reliability of gradientbased optimization methods are here superior to global optimization methods (such as genetic algorithms or Monte Carlo; see the discussion in Kaasalainen et al., 2001). We emphasize that the spin parameters, especially the period, usually have numerous local minima, so a dense enough comb of initial values of these parameters is a prerequisite for a good final reconstruction.
2. Shape
Given the diverse shapes of asteroids and the continuing progress in instrument technology, effective methods for shape representation are required for a general reconstruction scheme from observations. In inverse problems it is typically not clear a priori how well a given shape support will perform. In this section we present shape supports and corresponding regularization functions well suited for asteroidlike shapes.
2.1. Shape supports
An important part of shape modelling is the choice of shape representation. Assuming a typical asteroid surface is homeomorphic to the unit sphere, we can consider each coordinate as a function on the sphere, and choosing a suitable basis for functions, expand coordinate functions using this basis. This can be generalized to multiple bodies such as binaries in a straightforward manner. Typical such bases are spherical harmonics, spherical wavelets, and spherical splines. Our experiments suggest that parametrizations, which expand each coordinate function separately, tend to produce suboptimal results since they ignore the geometric dependencies and constraints between coordinates when considering surfaces represented by nontangled meshes. Thus we have found it useful to consider two wellregulated, but conceptually different, shape supports in practice: octantoids based on spherical harmonics, and subdivision surfaces.
2.1.1. Function series
An octantoid is a surface given by p ∈ R^{3} that can be parametrized in the form (1)where a, b and c are conveniently expressed as linear combinations of the (real) spherical harmonic functions , with coefficients a_{lm}, b_{lm} and c_{lm}, respectively. Note that (θ,ϕ), 0 ≤ θ ≤ π, 0 ≤ ϕ< 2π, are coordinates on the unit sphere S^{2} parametrizing the surface, but not describing any physical directions such as polar coordinates. As usual, the Laplace series for a,b,c are useful for keeping the number of unknowns; i.e. the coefficients of , small and the surface smooth. If b = c = 0, this representation is the usual starlike one with the radius exp(a), but we have found that even if the target is starlike, the octantoid form allows us to capture of detail better, and b and c we can represent with considerably fewer terms than the main function a. The number of shape parameters is thus between the (l_{max} + 1)^{2} of the starlike case and 3(l_{max} + 1)^{2}, when l_{max} is the largest degree of the function series. The drawback of this representation is its globality: one might want less smoothness regularization in some regions than in others. When more local control is desired (e.g. a feature clearly visible in flyby images or in radar), the representation (1) may be expanded with spherical splines or spherical wavelets to provide local detail without affecting the global shape. Depending on the desired level of resolution and the nonstarlike irregularity of the surface, the number of free function series coefficients is typically between 50 and 300 from low to midresolution. Function series are seldom useful for high resolution, where one may ultimately want to adjust each vertex separately by defining individual a_{i}, b_{i}, and c_{i}.
2.1.2. Subdivision control points
Subdivision surfaces offer local control more than global representations like function series. Beginning with an initial set of vertices and corresponding triangles, called a control mesh, the surface is iteratively refined by adding new vertices and computing new positions for old vertices. The vertex coordinates of the control mesh form the parameter set defining the surface. Each subdivision step smoothes out the surface in a higher level of resolution. Those subdivision schemes that are wellbehaved converge to a smooth limit surface.
In this paper, we use the Loop subdivision scheme (Loop, 1987). Considering a vertex p with immediate neighbours p_{0},...,p_{n − 1}, the subdivision method first creates new vertices by splitting each edge (2)where the indices should be interpreted as modulo n. After the vertex creation step, the position of the vertex p is refined as (3)The multiplier β is usually chosen to be (4)but other choices are also possible. The limit surface is continuous; C^{2} at the ordinary vertices (i.e. vertices that have six neighbours) and C^{1} at extraordinary vertices. The number of free control points for model rendering is similar to or somewhat lower than the number of function series coefficients (for a comparable level of resolution).
The main computational aspect with subdivision methods is that the number of facets increases exponentially with the number of divisions. After n subdivision steps, each facet that has been divided has produced 4^{n} subfacets. An alternative scheme to Loop subdivision is the subdivision (Kobbelt, 2000). Instead of splitting the edges, scheme subdivides facets by inserting a new vertex to the facet centroid and connecting it to the vertices of the facet (Fig. 1). The main attraction of the scheme compared to the Loop subdivision is the slower increase (3^{n}) of facets, while performing similarly in the limit.
In practice, it is usually a good idea to choose the initial control mesh to be an ellipsoid or a scaled convex surface obtained from lightcurve inversion, with a suitable number of vertices for the mesh. The number of subdivision steps should be chosen carefully: while each subdivision increases resolution and stability by spreading the influence of each parameter to a larger number facets, the computational burden grows exponentially. Instead of subdividing all the facets, better performance may be obtained with adaptive subdivision, where only facets benefiting from increased resolution are subdivided. However, it is not obvious how to do this automatically during optimization. A heuristic inclusion of surface regions to be refined based on a ranking of the improvement of the fit is one possibility (cf. the χ^{2}sensitivity map of Kaasalainen & Viikinkoski, 2012); visual inspection of the model fit and a graphical user interface can guide the refining process.
Fig. 1 Original control mesh (left) with 18 vertices (54 coordinates) as the shape parameter set and 32 facets, after two subdivision steps (middle), and after four subdivision steps (right). 

Open with DEXTER 
2.2. Regularization functions
In inverse problems, finding a feasible regularization method is typically the most delicate part of problem solving. Ideally, both the shape representation and regularization method should be chosen to complement each other. The shape support should be general enough to represent probable shapes, and the regularization should prevent unrealistic or degenerate shapes while, at the same time, reveal the features present in the data. For octantoids, the choice is remarkably easy. Assuming the basic shape is geometrically starlike, it is intuitively obvious to penalize the deviation from starlikeness. To this effect, we define (5)Every starlike surface has a representation for which η = 0, so η is a natural quantity to be included in the final χ^{2}function to be minimized (Sect. 5). The χ^{2}sum contains both the goodnessoffit measure and the regularizing functions that represent prior assumptions and expectations of the solution.
Subdivision surfaces have somewhat different smoothness properties in this regard. It is well known that the Loop subdivision converges to a smooth surface, so each subdivision step will produce a smoother result. However, it is computationally expensive to take a large number of subdivision steps. Therefore, it is advantageous to combine a few, usually two or three, subdivision steps with meshbased regularization methods.
While not strictly necessary, it is convenient to assume that the triangular mesh representing the shape forms a manifold. This assumption makes the checking of shadowing and illumination both conceptually and computationally simpler. Thus it is imperative to avoid selfintersections, as they introduce errors to the fitting process. One approach is to explicitly check for intersecting facets and retriangulate if required. However, triangulation and intersection tests are costly, and usually optimization steps leading to selfintersections are suboptimal. A better approach is to prevent selfintersections in the first place.
Regularization based on dihedral angles penalizes large angles between adjacent facet normals; i.e. the regularization prefers planar regions. We thus want to minimize (6)where are the facets of the mesh, and ν_{k} is the unit normal vector corresponding to the facet k. The sum is over all those facets j that are adjacent to the facet i, and the weights w_{ij} are usually chosen to be unity. As a special case, we may suppress only concave features, obtaining convex regularization (Kaasalainen & Viikinkoski, 2012) (7)where A_{i} is the area of the facet i and the sum is over those facets j that are adjacent to the facet i and tilted above its plane.
To prevent degenerate facets and maintain a homogeneous mesh, it is advantageous to inhibit large variations in facet areas (8)where is the mean facet area of the polyhedron.
In practice, the regularization functions η and γ_{2} are sufficient for octantoid surfaces, while γ_{2} and γ_{3} are useful for the subdivision surfaces. Unrealistically sharp angles can be prevented with γ_{1}, but a weight that is too large will inhibit convergence. In addition to geometric considerations, one can use regularization based on physical constraints, such as the requirement for the rotation axis to be close to the largest principal axis of the inertia tensor (Kaasalainen, 2011; Kaasalainen & Viikinkoski, 2012).
2.3. Reliability estimates
The octantoid representation or the subdivision mesh tend to produce aesthetically pleasing, “asteroidlike” surfaces, but it is not initially obvious which surface features of the model are actually present in the data, and which are the side effects of the shape support and the regularization used. Conventionally, Markov chain Monte Carlo (MCMC) methods are used to obtain a reliability estimate for the model parameters. However, in our case, modelling and systematic errors usually dominate (Kaasalainen & Ďurech, 2006), rendering the MCMC approach inefficient and inaccurate because the error distribution is not known (it is certainly not random Gaussian as in standard MCMC).
Moreover, the posterior distribution of shape parameters from MCMC will not really tell anything about the reliability of the model with respect to data, but only about the distribution of the estimate within the adapted shape support. We have found that this results in an overly optimistic conception of the reliability of the result, simply because the acceptable shape results cannot be probed widely enough using one shape support only. The Monte Carlo procedure focusses on regions of shape variation that are too small for both computational and geometric reasons. In addition, the computation of the model fit is time consuming if the data set and parameter space are large, making MCMC estimation computationally expensive.
To circumvent these obstacles, we have found the following approach fast and robust in practice. Any real feature of the model based on the data should also be present if another, independent model type such as a shape support is used. When model errors dominate, it is thus better to sample the “model space” within some χ^{2} than the χ^{2}space with some fixed model. As an example, shape models of the asteroid Daphne from adaptive optics images and photometry (Viikinkoski & Kaasalainen, 2014), using both the octantoid representation and subdivision surfaces, are shown in Fig. 2. The models are quite similar and fit the data equally well, and their difference gives an idea of the real level of resolution. The MCMC probing with either shape support leads to small differences that are unrealistic (insignificant compared to those in Fig. 2). Even the shapesupport test is likely to produce reliability limits that are too optimistic; the model error can be further enlarged by e.g. introducing random fluctuations in the scattering properties over the surface. This principle could be developed into a metalevel Monte Carlo procedure that probes the space of possible model types using latent parameters.
We conclude that shape sampling based on a fixed model type, no matter how diligently done with Monte Carlo or other methods, leads to overly optimistic resolution with artificial details. A typical example of this is the radar model of the asteroid Itokawa that portrayed imaginary detail at the resolution level expected from the data while not capturing even the largescale features. There was nothing wrong with the model fit to the data as such: the inverse problem was not unique (or very unstable) because of the restricted observing geometries and instrumental projection (Sect. 4.2), but the constrained shape support of the program did not reveal this (Ostro et al., 2005; Nolan et al., 2014).
Fig. 2 Model of asteroid (41) Daphne from adaptive optics images, reconstructed as a subdivision surface (left) and an octantoid (right). 

Open with DEXTER 
2.4. Inversion with photometry only
Since ADAM utilizes photometric data in addition to diskresolved data, we note that ADAM can be used to reconstruct a model using only photometric data (simply using only the photometric fit function from the toolbox). This is easy and fast to do (and the shape rendering is even faster than the standard convex inversion of lightcurves), but the result is inevitably unreliable: it is well known that even sizable nonconvex shape features require high solar phase angles to show in diskintegrated data (Kaasalainen et al., 2001; Ďurech & Kaasalainen, 2003; Kaasalainen & Ďurech, 2006). This can be probed with the shape reliability approach above.
As an example, we show reconstructed shapes of the asteroid Golevka in Fig. 3, based on the data in Kaasalainen et al. (2001). Both the subdivision method and the octantoidbased model display additional detail not seen in the convex model. However, the details are not supported by the data: the convex model gives at least as good a fit as the nonconvex, as is almost always the case with lightcurves (so far the only case of a better nonconvex model fit to photometry is that of the asteroid Eger in Ďurech et al., 2012). Indeed, with Golevka and other ground truth cases (maps from space probe missions), even the lightcurve fit with the correct shape and the scattering model assumed in inversion is not better than that with the convex model (Kaasalainen et al., 2001; Kaasalainen & Ďurech, 2006). This underlines the fact that, because of systematic errors, any bestχ^{2} optimized solution that relies only on photometry is likely to miss the details.
While the convex model yields the best overall agreement with the radarbased Golevka model (see the comparison in Figs. 3 and 4 in Kaasalainen et al., 2002), the nonconvex models portray much of the general sharpness and ruggedness of the body even though their details are not correct. The convex shape presents something of a softened error envelope within which numerous local shape variations are possible (as if the target were seen unfocussed), while the nonconvex representations are samplings of those variations. Their details coincide neither with each other nor with those in the radarbased model, but they are useful as illustrations and for probing the potential shape options (cf. the nonconvex examples in Kaasalainen et al., 2004).
Fig. 3 Asteroid (6489) Golevka reconstructed from diskintegrated photometry. From left to right: convex, octantoid and subdivision surface. 

Open with DEXTER 
3. Fourier transform and information content
As discussed in Viikinkoski & Kaasalainen (2014), the Fourier transform (FT) facilitates a natural interpretation for the pixel size as the maximum frequency present in the data, and makes it easy to incorporate the impulse response function of the imaging system. It also makes the optimization procedure fast and straightforward, without the cumbersome aspects related to pixellated image fields and binned model image distributions. The principle of the ADAM approach is to compare, instead of the images themselves, a set of FT samples (typically some thousands depending on the level of resolution) from the model image with those of the data image, and to iterate until the best fit is found. This is described in Sect. 5.
Letting be the set of facets forming a model polyhedron and a projection operator, the twodimensional Fourier transform of a projected polyhedron in the (ξ,η)plane is (9)where B_{i} is the luminosity value of the facet i, and the function I(ξ,η) is unity if the point projected on (ξ,η) is visible and zero otherwise. As shown in Viikinkoski & Kaasalainen (2014), we obtain by Green’s theorem, dividing a facet into subfacets if necessary so that we may assume I is constant within each subfacet, (10)where (11)for the jth boundary line segment (oriented counterclockwise) of the facet i, with the end points (a,b) and (c,d).
The summation over the interior edges of a projected polyhedron can be reordered by noting that each polygon edge in the interior is shared by two polygons, so a new factor can be taken to be the difference between the two B_{i}, and the edge term is computed only once. This explicitly shows why most of the information in the image is indeed from the limb and shadow boundary curves discussed in Kaasalainen (2011) and Kaasalainen & Viikinkoski (2012). The values of for interior triangle edges are usually close to zero (indeed, they vanish for the geometric scattering B_{i} = const.), so most of the weight is on the boundary edges. In practice, this is confirmed by the similar results for e.g. the asteroid Daphne obtained by KOALA and ADAM. There is little real information in the interior pixels of adaptive optics images, but on the other hand their errors do not distort the result either: the difference between the KOALA and ADAM models (for the same initial values and shape support) is negligible.
The role of boundary information can be understood when compared to the extreme case of lightcurve data: if we sum the pixel brightnesses over the image as in photometry, all the local shape information in the image is lost, so the remaining information is considerably more dependent on the lightscattering properties that are never very well known. With images the boundary contrast is always largest, however, so it is sufficient to have some kind of reasonable scattering (or thermal distribution) model to account for the interior pixel contrasts. Indeed, the uniqueness theorems on the image, interferometry, occultation, or radar data are based on robust boundary contour information (Kaasalainen, 2011; Kaasalainen & Viikinkoski, 2012; Viikinkoski & Kaasalainen, 2014). With diskintegrated data only, Minkowski stability is luckily on our side when using convex models (Kaasalainen et al. 2001, 2002).
4. Data sources
The versatility of the ADAM algorithm enables the handling of different data sources with only minor changes to the instrumentdependent part of the procedure (essentially just the definition of the instrumental projection plane and the adopted pointspread function). In this section, we present diverse examples of shape reconstruction with ADAM using both simulated and observed data.
4.1. Interferometry and ALMA
The interferometric imaging method differs radically from a typical telescope; instead of observing the sky brightness directly, the interferometer samples the Fourier transform of sky brightness. Each antenna pair of the interferometric array determines one sample on the Fourier plane. The maximum separation between antennas determines the maximum attainable resolution. The interferometer most relevant to asteroid shape studies is the Atacama Large Millimeter Array (ALMA) in the Chilean desert. In its full configuration, the interferometer will be capable of observing at the resolution of a few milliarcseconds at the wavelength of 0.3 mm, corresponding to the separation of 16 km between antennas.
Given the brightness distribution I(ξ,η) on the planeofsky, the visibility function is defined as the integral (12)which is a twodimensional Fourier transform of the brightness distribution. Each antenna pair, corresponding to the projected baseline on the planeofsky, samples the visibility function. When the visibility function is sampled on a sufficiently dense set, the Fourier transform can be inverted to obtain the brightness distribution I(ξ,η). Since the function V(u,v) is measured only at a finite number of points, the observed visibility function is (13)where F(u,v) is a sampling function corresponding to the sampled points on the (u,v)plane. Thus the obtained brightness distribution is actually (14)i.e., a convolution of the true brightness distribution with the inverse Fourier transform f(ξ,η) of the sampling function. Deducing the true brightness distribution I from the partially measured brightness is an inverse problem and there are several iterative algorithms to infer I from , see, e.g. Labeyrie et al. (2006).
While the images obtained from the interferometer are informative, the great advantage with ADAM is that the algorithm works directly with the values of the visibility function obtained from the instrument. This approach has several distinct advantages:

sparse data may be used (e.g. interferometry with a few baselines);

the distribution of antennas does not cause bias, since the Fourier transform is not inverted;

possible artefacts caused by the inversion process are avoided;

the dependence between different observations is taken automatically into account.
To obtain the luminosity values for the model surface (i.e. the brightness factor B_{i} for each facet) in the infrared regime of ALMA, we can use the Fourierseries approximation of Nesvorný & Vokrouhlický (2008) as discussed in Viikinkoski & Kaasalainen (2014). The fast analytical computations are then efficient in the optimization. A simple thermophysical model is sufficient for shape reconstruction, as the most relevant information is contained in the boundary data, which are quite robust with respect to our thermal model. This is in contrast to the diskintegrated thermal data that are more sensitive to both the surface properties and the thermal model.
For thermal infrared imaging, ALMA facilitates asteroid observations at resolution levels previously attained only by rangeDoppler radar. To explore the possibilities of ALMA for shape modelling, we use the Common Astronomy Software Applications (CASA) package developed by National Radio Astronomical Observatory (NRAO) to simulate observations.
Consider a hypothetical asteroid with geocentric and heliocentric distances of 1.5 and 1 AU, respectively. The thermal flux is observed at the 350 GHz band, a frequency located in an atmospheric window. There are 11 observation runs, each observation lasting 50 s with 10 s integration time. We choose an antenna configuration providing approximate resolution of 10 mas, a resolution which is well within the capabilities of ALMA. The antenna configuration and the corresponding uvplane sampling pattern are shown in Fig. 4. The uncorrupted planeofsky images, with a resolution of five milliarcseconds, are displayed in the column on the left in Fig. 5. We use the CASA software to add realistic atmospheric noise to the observations. The resulting dirty images, which are obtained by assuming that the unsampled frequencies are zero, are shown in the middle column. These images are provided for illustration purposes only, since ADAM uses the uvplane samples directly.
Fig. 4 Antenna locations of ALMA (left) and corresponding uvplane visibilities (right). Images generated with the CASA software package. 

Open with DEXTER 
Fig. 5 Simulated, uncorrupted images with 5 mas pixel size (left column), observed dirty images generated with CASA (middle) and the reconstructed lowresolution shape model (right). Note that the middlecolumn images are not needed in inversion; we use the direct FT data instead. The images are what would be seen if the raw data were deconvolved for viewing purposes as is usually done for ALMA targets. The test shape model is from Ostro et al. (2000). 

Open with DEXTER 
To test the ADAM reconstruction method, we use a lowresolution octantoid representation with 75 shape parameters. We also fit a scaling term, common to all observations. Usually it is a good idea to use scaling specific to each observation, but in this case we know that all the simulated observations are done in similar conditions, so the common scaling term is justified. The reconstructed shape is displayed in the right column in Fig. 5, with the same observation geometries as for the model images. The smallscale detail is lost, which is to be expected because of the added atmospheric noise and coarse instrument resolution. However, the bifurcated shape is well recovered despite the noisy data (note that we used ALMA data only). The computation time for this reconstruction was a few minutes. For real observations, complementary data are often provided by other observation methods e.g. diskintegrated photometric data are almost always available.
4.2. Radar data
The mathematical principles of the feasibility and uniqueness of the inversion of rangeDoppler images are discussed in Viikinkoski & Kaasalainen (2014). Here we consider some practical issues related to shape reconstruction. While other imaging methods rely on detecting the radiation of the sun that is reflected or reradiated from the asteroid, radar provides its own illumination, making it possible to observe an asteroid regardless of the position of the sun. Moreover, in contrast to the visible or infrared wavelengths, the frequencies used by the radar are not significantly distorted by the atmosphere. Additionally, the properties of the waveform may be carefully controlled to reveal structural details on the surface of the asteroid. These properties make it possible to obtain data resolution down to 10 m or less for nearEarth asteroids, but this does not immediately translate to the same model resolution because of the inverse problem (cf. the Itokawa example in Sect. 2.3).
RangeDoppler radar resolves an object both in the range and in the lineofsight velocity that translates to the Doppler shift of the reflected pulse. The frequency spectrum may be extracted by taking the fast Fourier transform of the pulses corresponding to a particular range gate. The actual hardware implementation and the signal processing are complicated as the detected signals are below the noise level of the instrument (Ostro et al., 2002). Fortunately, the technical specifics are not required for the actual shape reconstruction, since the radar performance may be modelled by the pointspread function of the system.
The point p = (x,y,z) on the asteroid’s surface can be transferred to the rangeDoppler frame (r,D) by the linear mapping where ω is the rotation rate of the asteroid around the zaxis, and (θ,φ) are the spherical radar direction coordinates as seen from the asteroid. In this mapping, the rangeDoppler radar image brightness L may be written as an integral over the asteroid surface S as (17)where h_{D} and h_{r} are the pointspread functions of the radar system, corresponding to the Dopplershifted frequency D and the range r, respectively. Here I is the visibility function, which is unity if the point is visible to the radar and zero otherwise. This form is similarly defined for all generalized projections (Kaasalainen & Lamberg, 2006). The mapping p → (r,D) is unique, but its inverse is manytoone, so the inherent information content of a rangeDoppler image is considerably smaller than that of an optical image of similar resolution. Thus, while the nominal resolution provided by radar may be unmatched by any other instrument, the drawback of radar imaging is the difficulty of the interpretation of the images.
The radar scattering function is given by B, which is usually a simple cosine law (18)where μ is the cosine of the angle between the surface normal and the radar direction. The constants C and n measure the surface reflectivity and the specularity of scattering, respectively. The validity of Eq. (18) for modelling the microwave scattering from the asteroid’s surface is a rather convoluted question. While the cosine law is quite simplified, it should be noted that as the reflected wave is formed in a complicated manner by the surface material whose properties and roughness are usually unknown, fully realistic modelling of the reflected wave is not computationally feasible. However, as in the other diskresolved cases, most of the information is contained in the boundary contours and is thus independent of the scattering model used.
Assuming the asteroid is modelled as a polyhedron with triangular facets , the integral (17) may calculated separately for each facet, after projecting each triangle T_{i} as a triangle on the rangeDoppler plane (19)where we have assumed that the visibility I and the scattering law B are constant within a triangle.
Taking the Fourier transform on both sides, applying the convolution theorem, and writing for the sum over the edges of a Fourier transformed triangle as in Sect. 3, we obtain (20)where H_{r}(u) and H_{D}(v) are the Fourier transforms of h_{r} and h_{D}, respectively.
Like any images, radar plots are seldom correctly aligned in some reference frame due to the errors in the centre of mass prediction, so the actual position of a radar image with respect to the twodimensional projection of the model must be determined during the optimization. The task of image alignment is further complicated by the peculiar asymmetric structure of radar images, especially the bright leading edge, other possible ridges of strong reflectivity, and the fading farthestrange pixels. If the alignment information is unknown, it is usually a good idea to fit the image offsets to a fixed shape first, obtaining better initial positions that can be used in the shape optimization.
To demonstrate the reconstruction method, we make a fast ADAM model of the nearEarth asteroid 2000 ET_{70}. Our goal is to get a quick first look at an initial model (to be refined at will). The asteroid was observed during February 2012 at Arecibo and Goldstone observatories using 2380 and 8560 MHz rangeDoppler radars (Naidu et al., 2013). The images obtained from Arecibo have a resolution of 15 m in range and 0.075 Hz in frequency. Goldstone images have a somewhat lower resolution, 15 to 75 m and 1 Hz, respectively. Our goal is to produce mediumscale detail in the reconstructed shape, so a typical model choice is an octantoid with l_{max} ~ 10 and around 1500 facets. Our example is “firstresult oriented” on purpose, so we assume no information about the instrumentspecific distortions, or more importantly, knowledge about the pointspread functions determined by the instrument and the processing routines of the radar signal. Thus the pointspread function used in the shape reconstruction is simply the twodimensional delta function.
Fig. 6 Midresolution shape model of the asteroid 2000 ET_{70} reconstructed from radar images. Viewing directions are from the positive x, y, and z axes, respectively. 

Open with DEXTER 
For each data image, we fit, in addition to the shape parameters, the offset with respect to the model centre of mass and the reflectivity term in Eq. (18). The reconstructed middleresolution shape is shown in Fig. 6 and the model fit to the data in Fig. 7. The shape model fits the boundary contours of the radar images satisfactorily, but there are some differences in the interior details. This is a consequence of the parametrization and facet size chosen for reconstruction. The interior could be reproduced in greater detail by choosing a different parametrization, for example locally adaptive subdivision surfaces, or by refining the positions of individual vertices. The model dimensions, shape features, and spin parameters agree with those published by Naidu et al. (2013; the spin parameters are identical except for a 2° difference in the pole latitude, well within error limits).
Fig. 7 Examples of rangeDoppler images of the asteroid 2000 ET_{70} from Arecibo Observatory (rows 1 and 3) and corresponding simulated images from the midresolution model (rows 2 and 4). The contrast scale of the model image is somewhat modified to reveal inner image features. 

Open with DEXTER 
The main point of the initial lowtomiddle resolution is that the speed of ADAM is considerable, and a detailed knowledge of the instrument or the surface scattering physics is not needed, so one obtains a first model very fast by just feeding in the images. The middleresolution radarbased reconstruction (using 82 radar images) was computed in less than an hour on a standard laptop, and GPU programming can reduce the computation time significantly. This makes possible a broad sampling of the parameter space or realtime experimenting with various models. Once a lowerresolution model has been adopted as the final frame, it is straightforward to refine it further. However, this requires accurate information about the pointspread and scatter functions.
4.3. Adaptive optics and other images
Model reconstruction from adaptive optics images in the Fourier approach is extensively covered in Viikinkoski & Kaasalainen (2014), along with an example of the reconstruction of the main belt asteroid Daphne from adaptive optics images and photometry (Fig. 2). Other imaging data may be incorporated into the framework using a similar approach. For instance, flyby images are, from the viewpoint of the reconstruction algorithm, conceptually identical to the AO images. This is one of the attractions of ADAM: at the bare minimum, the user does not need to know anything about the images except their projection matrix and epochs.
We note that the photometric data were actually not even needed in reconstructing Daphne (except for a better estimate of the rotation period than with AO images only). The shape results with or without photometry are similar. This shows that even sparse AO data are well sufficient for modelling asteroid spin states and shapes in detail.
4.4. Onedimensional projection operators
In the regime between diskintegrated and diskresolved observations there are onedimensional operators that project the planeofsky onto a line. Typical examples are the continuouswave (CW) Doppler spectra that measure the distribution of the reflected power in frequency only, and the fine guidance sensors (FGS) onboard the Hubble Space Telescope, measuring the brightness distribution along an instrument axis. Onedimensional projections are seldom sufficient for actual shape reconstruction, but they may contain useful information about the object’s size or indications about the bifurcated structure (Kaasalainen & Viikinkoski, 2012), and combined with other sources, they facilitate shape inversion.
In both examples, the measurement can be written in the form (21)where I(ξ,η) is the planeofsky brightness (optical or radar) distribution of an object, P is the pointspread function of the instrument, and the angle γ corresponds to the rotation of the sensor in the image plane. In Kaasalainen & Viikinkoski (2012), the integral was evaluated using a Monte Carlo method: the projected model was sprinkled with uniformly distributed sampling points, and the integral was approximated as a sum over the visible and illuminated sampling points. We demonstrate how the Fourier transform method can be used to interpret the integral as a tomographic operator on the Fourier plane.
Taking the Fourier transform on both sides and using the projectionslice theorem (a slice of a 2D FT along a line through the origin equals the 1D FT of the projection of the original 2D function onto a line in the same direction; see e.g. Bracewell, 2003), we get (22)where the calligraphic characters denote the Fouriertransformed functions. Now it is obvious that is a slice of a Fouriertransformed brightness distribution along a line through the origin, multiplied with the Fourier transform of the pointspread function. Moreover, this means that the same algorithm may be used to fit both FGS and adaptive optics data, and similarly both CW Doppler data and the rangeDoppler images. In other words, we extract a onedimensional Fourier transform from the 2D model FT, and compare this with the 1D FT formed from the data in the same manner as in the full 2D case.
5. ADAM algorithm
The flowchart in Fig. 8 describes the workings of ADAM. More specifically, the algorithm may be divided in five distinct steps:

1.
For each data image D_{i} and observation geometry ℰ_{i}, the twodimensional Fourier transform ℱD_{i}(u,v) of D_{i} is sampled at a set of points { (u_{ij},v_{ij}) }, j = 1...N_{i}, on the spatial frequency plane. The size of the set is chosen to correspond to the level of resolution. For pixel images, the transform can be computed with Eq. (10) when considering each pixel as a polygon, or with using fast Fourier transform functions for chosen grid points (but the time spent for ℱD_{i}(u,v) is irrelevant, as are most of the computations, for the trial models).

2.
The shape support and resolution level (number of parameters) are chosen. The parameters are initialized such that the initial shape is a sphere approximately equal in size to the target.

3.
For each observation geometry ℰ_{i}, the Fourier transform ℱM_{i}(u,v) of the corresponding projection image M_{i} of the model is calculated as described in the previous sections, together with the partial derivatives of ℱM_{i}(u,v) with respect to all optimized parameters. Raytracing, scattering or luminosity models, and coordinate transforms for the image plane are discussed in Kaasalainen et al. (2001), Kaasalainen (2011), and Viikinkoski & Kaasalainen (2014).

An objective function χ^{2} is formed, with the square norm of the complexvalued FT fit error (23)where is the offset between the data image D_{i} and the model image M_{i}, and, by the convolution theorem, is the Fourier transform of the pointspread function of the imaging system. The γ_{i} represent various regularization terms defined above.
Fig. 8 ADAM optimization algorithm as a schematic for one image type.
Open with DEXTER For brevity, we have written only one data mode in Eq. (23); any number of modes with their goodnessoffit functions can be added to the sum. These functions for photometry and silhouettes (occultations) are given in Kaasalainen et al. (2001), Kaasalainen (2011), and Viikinkoski & Kaasalainen (2014). The determination of the weights of the data modes (as λ_{i} for the regularization functions) is discussed in Kaasalainen (2011) and Kaasalainen & Viikinkoski (2012). Weights can be determined for any subsets of data (e.g. less reliable images) if necessary. In addition, the intensity level of each data and model image must be normalized. Often it is enough to divide both model M_{i} and data image D_{i} by their respective mean intensities. Equivalently, writing (24)we have (25)where the mean ⟨·⟩ is taken over { (u_{ij},v_{ij}) }, j = 1...N_{i}. However, sometimes it is better to allow the intensity level of each M_{i} to be a free parameter and use χ^{2}; this is useful in the case where the mean intensity of D_{i} is corrupted by excessive noise in the image background (this is typical for rangeDoppler images). This causes the based solution to have a slightly wrong size to compensate for the “diluted” normalized intensity level inside the actual object region of D_{i}.

5.
The shape and spin parameters and the offsets as well as the possible intensity level factors C_{i} minimizing χ^{2} are determined with a suitable method such as the LevenbergMarquardt algorithm. If there are several hundreds of parameters, as in the case of fitting all shape vertices directly (instead of using function series or control points) to produce maximal resolution, the conjugate gradient method is efficient (Kaasalainen et al., 2001).
6. Conclusions and discussion
The ADAM algorithm can handle radar data, images, interferometry (also in the thermal infrared), photometry, and occultations separately or in combinations. The ADAM procedure consists of a number of modules, and there are various options for each module that are customized to the enduser (e.g. the adopted optimization method, regularization functions, shape support and mesh structure, raytracing method, coordinate system, luminosity/scatter model, image formats, etc.). In this sense, ADAM is a toolbox and a set of building blocks rather than a readymade program.
The main idea behind ADAM is the efficient use of the Fourier transform in handling both images and onedimensional projection data. Fourier analysis has long been used in e.g. image compression because it conveniently captures the essential information in a hierarchy of resolution. In the same vein, the FT approach in ADAM is ideal for producing models of desired levels of resolution, especially in the low to mediumresolution category. In this framework, the goodnessoffit function between the model and the data is easy to compute and use in optimization. In addition, with this method, its convergence properties are more robust than if the images are used directly. In fact, one does not necessarily even have to look at the images or know much about the instrument that produced them. An analogy of this paradox is the simple onedimensional problem of realigning two phaseshifted copies of a dualfrequency signal. If one does this by minimizing the signal difference by optimizing the shift in the original amplitude space, there are multiple local minima, but in frequency space the offset is found immediately.
Despite its automatic character, ADAM should not be used as a black box: asteroid reconstruction is a complicated inverse problem, and one should be familiar with its mathematical principles to understand the limitations and information content of the data sources.
Appendix A: Sample ADAM functions
In order to make the structure of ADAM more concrete, we show an example of how the program is divided into subroutines. We consider only the part of the program that computes the heat flux density of an object and its Jacobian (i.e. the partial derivatives of each modelled flux data point w.r.t. the free parameters), as all the other modules are structurally similar. In the case of optical images, the flux is replaced by brightness from scattering; for radar, by the signal strength in the rangeDoppler plane.
The partial derivatives w.r.t. the shape parameters are initially calculated with respect to the vertex coordinates, making the routines independent of the parametrization we used. The Jacobian is determined using the chain rule only in the final phase. The functions are complexvalued, since the fitting is done on the frequency plane. In the optimization, the data are divided into real and imaginary parts and fitted separately. Usually, the LevenbergMarquardt or the conjugate gradient method is used to optimize the χ^{2}fit.
Three different coordinates systems are used in ADAM: the asteroidcentric coordinate frame with coordinate axes fixed to the asteroid, the asteroidcentric inertial frame, and the camera frame, which is determined by the instrumental orientation geometry. The planeofsky view of an asteroid is obtained by projecting the asteroid in the camera frame to the xyplane.
The Jacobian and the vector, consisting of simulated values corresponding to the observations, are computed using the following subroutines:

Generate_HF_Matrix calls the subroutine Calc_Heat_Flux for each observation, and then combines the Jacobian submatrices into a full Jacobian matrix corresponding to all the observations.

Calc_Heat_Flux calculates the Fourier transform of the twodimensional planeofsky flux density and its partial derivatives by calling the subroutines Calc_Temp, Rot_Matrix, Cam_Matrix, Calc_Vis and Calc_FT. After the flux density of each facet is determined, the routine transforms the triangular mesh to the camera frame and projects the visible part of the mesh onto the xyplane by discarding the zcoordinate. Finally, each vertex is transformed to the frequency plane using the Calc_FT subroutine and the contributions of the Fouriertransformed facets are summed.

Calc_Temp determines the temperature of the facets corresponding to the observation geometry, using the FFT method. The partial derivatives of the temperature with respect to the shape parameters are also calculated. This subroutine also calls subroutines Rot_Matrix and Calc_Vis.

Rot_Matrix calculates the rotation matrix needed to transform the object to the inertial frame. The rotation matrix is determined by the spin vector and the observation time.

Cam_Matrix determines the matrix needed to transform the inertial frame to the camera frame. This depends on the instrument location and orientation. The zcoordinate codes the relative distance from the instrument.

Calc_Vis determines the visibility of facets using raytracing. In contrast to the optical case, a facet can be visible to the observer even if it is not illuminated by the sun.

Calc_FT calculates the Fourier transform of a triangle projected onto the xyplane together with the corresponding partial derivatives.
The most important setup factors determining the computation time of ADAM are the numbers of facets and data points (the number of images and their pixels). The computation time increases approximately linearly with both numbers. The cost of actual optimization steps increases superlinearly with the number of free parameters, but with large data sets (such as the radar example above) most of the computation is spent on determining function values and their partial derivatives with respect to the vertex coordinates. In such cases, the number of shape parameters is not critical for the computational cost in the midresolution regime, so one is free to choose a number that best corresponds to the resolution level (and set the number of facets accordingly). When the data set is small, the computation time is short in any case, and the model is likely to be low resolution, so again the number of parameters is not an issue. The cost of visibility determination by raytracing is insignificant as the potential blocker facets can be precomputed (Kaasalainen & Torppa, 2001).
The shape reconstruction from observations is an easily parallelizable problem. There are two obvious levels of parallelism: each observation can be calculated independently; or, within each observation, the contribution of each facet may be determined simultaneously. The best choice depends on the computer’s architecture. The observationlevel parallelism may be easily exploited using the MATLAB parallel computing toolbox, or more effectively by using the OpenMP API in the C language. The reduction in execution time scales almost linearly with the number of CPU cores. This is the approach currently implemented in ADAM.
While it is possible to implement facetlevel parallelism on the CPU by dividing the facet computations between several CPU cores, a more natural approach is to use one thread per facet. This kind of implementation is inefficient on the CPU, since the threadswitching latency is high compared to the running time of a thread. However, the ability of the GPU to run thousands of lightweight threads simultaneously combined with the virtually costless thread switching makes it possible to attain orders of magnitude faster computation than with CPU. We will implement GPU acceleration in ADAM using the Nvidia CUDA programming platform.
Acknowledgments
This work was supported by the Academy of Finland project “Modelling and applications of stochastic and regular surfaces in inverse problems” and the CoE in inverse problems research. The work of J.Ď. was supported by the grant GACR P209/10/0537 of the Czech Science Foundation. We thank JeanLuc Margot for providing the radar images of asteroid 2000 ET_{70}.
References
 Bracewell, R. 2003, Fourier analysis and imaging (Springer) [Google Scholar]
 Carry, B., Kaasalainen, M., Merline, W. J., et al. 2012, Planet. Space Sci., 66, 200 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
 Ďurech, J., & Kaasalainen, M. 2003, A&A, 404, 709 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
 Ďurech, J., Grav, T., Jedicke, M., Kaasalainen, M., & Denneau, L. 2006, Earth Moon Planet., 97, 179 [NASA ADS] [CrossRef] [Google Scholar]
 Ďurech, J., Vokrouhlický, D., Baransky, A. R., et al. 2012, A&A, 547, A10 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
 Ďurech, J., Carry, B., Delbo, M., Kaasalainen, M., & Viikinkoski, M. 2015, Asteroids IV [arXiv:1502.04816] [Google Scholar]
 Kaasalainen, M. 2004, A&A, 422, L39 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
 Kaasalainen, M. 2011, IPI, 5, 37 [CrossRef] [Google Scholar]
 Kaasalainen, M., & Ďurech, J. 2006, Proc. IAU, 2, 151 [CrossRef] [Google Scholar]
 Kaasalainen, M., & Ďurech, J. 2013, in Asteroids, ed. V. Badescu (Berlin, Heidelberg: Springer), 131 [Google Scholar]
 Kaasalainen, M., & Lamberg, L. 2006, Inverse Probl., 22, 749 [NASA ADS] [CrossRef] [Google Scholar]
 Kaasalainen, M., & Torppa, J. 2001, Icarus, 153, 24 [NASA ADS] [CrossRef] [Google Scholar]
 Kaasalainen, M., & Viikinkoski, M. 2012, A&A, 543, A97 [NASA ADS] [CrossRef] [EDP Sciences] [Google Scholar]
 Kaasalainen, M., Torppa, J., & Muinonen, K. 2001, Icarus, 153, 37 [NASA ADS] [CrossRef] [Google Scholar]
 Kaasalainen, M., Mottola, S., & Fulchignoni, M. 2002, Asteroids III, eds. W. F. Bottke Jr., A. Cellino, P. Paolicchi, & R. P. Binzel (Tucson: University of Arizona Press), Asteroids III, 139 [Google Scholar]
 Kaasalainen, M., Pravec, P., Krugly, Y. N., et al. 2004, Icarus, 167, 178 [NASA ADS] [CrossRef] [Google Scholar]
 Kobbelt, L. 2000, in Proc. 27th annual conference on Computer graphics and interactive techniques (ACM Press/AddisonWesley Publishing Co.), 103 [Google Scholar]
 Labeyrie, A., Lipson, S. G., & Nisenson, P. 2006, An introduction to optical stellar interferometry (Cambridge University Press) [Google Scholar]
 Loop, C. 1987, Master’s Thesis, University of Utah [Google Scholar]
 Naidu, S. P., Margot, J.L., Busch, M. W., et al. 2013, Icarus, 226, 323 [NASA ADS] [CrossRef] [Google Scholar]
 Nesvorný, D., & Vokrouhlický, D. 2008, AJ, 136, 291 [NASA ADS] [CrossRef] [Google Scholar]
 Nolan, M., Bramson, A., & Magri, C. 2014, in Asteroids, Comets and Meteors, Proc. Conf., eds. K. Muinonen, et al. [Google Scholar]
 Ostro, S. J., Scott, R., Hudson, et al. 2000, Science, 288, 836 [NASA ADS] [CrossRef] [Google Scholar]
 Ostro, S. J., Hudson, R. S., Benner, L. A., et al. 2002, Asteroids III (Tucson: Univ. of Arizona Press), 151 [Google Scholar]
 Ostro, S. J., Benner, L. A. M., Magri, C., et al. 2005, Meteoritics, 40, 1563 [NASA ADS] [CrossRef] [Google Scholar]
 Viikinkoski, M., & Kaasalainen, M. 2014, IPI, 8, 885 [CrossRef] [Google Scholar]
All Figures
Fig. 1 Original control mesh (left) with 18 vertices (54 coordinates) as the shape parameter set and 32 facets, after two subdivision steps (middle), and after four subdivision steps (right). 

Open with DEXTER  
In the text 
Fig. 2 Model of asteroid (41) Daphne from adaptive optics images, reconstructed as a subdivision surface (left) and an octantoid (right). 

Open with DEXTER  
In the text 
Fig. 3 Asteroid (6489) Golevka reconstructed from diskintegrated photometry. From left to right: convex, octantoid and subdivision surface. 

Open with DEXTER  
In the text 
Fig. 4 Antenna locations of ALMA (left) and corresponding uvplane visibilities (right). Images generated with the CASA software package. 

Open with DEXTER  
In the text 
Fig. 5 Simulated, uncorrupted images with 5 mas pixel size (left column), observed dirty images generated with CASA (middle) and the reconstructed lowresolution shape model (right). Note that the middlecolumn images are not needed in inversion; we use the direct FT data instead. The images are what would be seen if the raw data were deconvolved for viewing purposes as is usually done for ALMA targets. The test shape model is from Ostro et al. (2000). 

Open with DEXTER  
In the text 
Fig. 6 Midresolution shape model of the asteroid 2000 ET_{70} reconstructed from radar images. Viewing directions are from the positive x, y, and z axes, respectively. 

Open with DEXTER  
In the text 
Fig. 7 Examples of rangeDoppler images of the asteroid 2000 ET_{70} from Arecibo Observatory (rows 1 and 3) and corresponding simulated images from the midresolution model (rows 2 and 4). The contrast scale of the model image is somewhat modified to reveal inner image features. 

Open with DEXTER  
In the text 
Fig. 8 ADAM optimization algorithm as a schematic for one image type. 

Open with DEXTER  
In the text 
Current usage metrics show cumulative count of Article Views (fulltext article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 4896 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.