Methods for visual quality assessment of a digital terrain model

CRAZIEST Sagawa1gou Funny TikTok Compilation | Try Not To Laugh Watching Cactus Dance Challenge
CRAZIEST Sagawa1gou Funny TikTok Compilation | Try Not To Laugh Watching Cactus Dance Challenge

1A digital terrain models (DTMs) is a continuous surface that, besides the values of height as a grid (known as a digital elevation model—DEM), also consists of other elements that describe the topographic surface, such as slope or skeleton (Podobnikar, 2005). Different techniques for the generation of DTMs have been developed since their inception more than fifty years ago (Miller and Laflamme, 1958; Doyle, 1978). The first decades focused mainly on models’ reliability. The common techniques for quality assessment were based on the statistical comparison of small reference areas of higher quality with the created DTM in order to find outliers. Until the end of the 90s, high quality terrain data were acquired mainly photogrammetrically using aerial photographs and manual stereo measurements or matching techniques, or by vectorisation of contour lines from topographical maps and attribution.

  • 1 e.g. NASA’s SRTM (Shuttle Radar Topography Mission) with a horizontal /planimetrical/ resolution o (…)
  • 2 airborne LIDAR for local DTMs with resolution of around 1 m
  • 3
  • 4
  • 5
  • 6

2The quality of DTMs significantly increased over the last decade due to three significant factors:
The first was the introduction and development of new methods for data acquisition, especially from satellites and airplanes. At small scales (coarser spatial resolution) radar interferometric techniques (IfSAR) had been applied to generate global DTMs1 (Burrough and McDonnell, 1998; Maune, 2001). For larger scales and more local usage, airborne laser scanning (ALS) techniques have been applied2 (e.g. Kraus and Pfeifer, 1998).
The second factor is the increasing availability of additional data sources that are useful for the DTM quality assessment or enhancement. In addition to the aerial photographs and contour lines, different point datasets with height attributes could also be applied, such as fundamental geodetic network points, boundary points of land-cadastre, databases of buildings, spot elevations, and other related datasets such as highway construction or hydrological network measurements. Even datasets without height attributes such as lines of a hydrological network, roads, railways, and standing water polygons can be used (Podobnikar, 2005). These additional data sources can provide valuable input for integrated DTM production, as exemplified in Slovenia (Podobnikar, 2005) and in Europe (EuroGeographics, 2008).
Thirdly, applications using DTMs are now part of our everyday lives (e.g., Google Earth3, Microsoft Virtual Earth4, NASA World Wind5, Radrouten Planer6…). This trend can also have some impact on the quality of the DTMs used if it influences usability significantly.

3The higher the resolution, the more difficult the evaluation of input data quality and the assessment of the resulting DTM are. Experience indicates that the effort is proportional to the square of the inverse value of horizontal resolution. High resolution DTMs are thus more prone to errors. Visual methods can be very important for the evaluation of spatial data and can balance some weaknesses of statistical methods. They are still underused for at least three reasons. Visual approaches being qualitative are generally more neglected than statistical ones which are considered to be more objective. The other reasons for the lower acceptance of visual methods lie in the insufficient graphical capabilities of computers until recently and in the longer tradition of using statistical methods. Finally, visualisation of spatial data has traditionally been part of cartography. The main emphasis of this paper is to focus attention on visual methods as a powerful tool for quality assessment.

4The quality of spatial analysis depends on data quality, (data) model relevance and on the way they interact (Burrough and McDonnell, 1998). The model (or nominal ground) is a conceptualisation and representation (abstraction) of the real world, i.e., a selected representation of space, time, or attributes (Aalders, 1996). The datasets—in our case the DTMs—are realised by the type of spatial object to which variables refer on the level of measurement of these variables. The model relevance is a semantic quality of the representation by which a complex reality is captured. Data quality refers to the performance of the dataset given the specification of the data model (Haining, 2003).

5The DTM dataset is an approximation of the reality, based on a nominal ground. A semantically reliable and high quality data model (as a base for the DTM generation) should be carefully defined. The DTMs might vary depending on their purpose, the quality of data sources or interpolation algorithms, the experience of the operators, etc.

  • 7 A related data model is the digital surface model (DSM). The term refers, on the one hand, to a ge (…)

6A basic distinction can be made between digital elevation models (DEMs) and digital terrain models (DTMs) (Burrough and McDonnell, 1998; Podobnikar, 2005; Sutter et al., 2007)7. The DEM is one of the most used ‘raster datasets’ (a grid or a matrix) in geographical information systems (GIS). An elevation value (height) is attributed to each square cell of the grid. The set of cell heights can then be interpreted in two ways: In the first approach, each cell represents a discrete area, hence the entire cell area is assumed to have the same value, the changes occur only at the edges of the cells. In the second approach, the area between the cell centres is assumed to have some intermediate values. This approach is closer to the DTM definition. The DTM is considered as a continuous, usually smooth surface which, in addition to height values (as DEMs), also contains other elements that describe a topographic surface: slope, aspect, curvature, gradient, skeleton (pits, thalwegs, saddles, ridges, peaks), and others. In this study, we focus on DTM but the methods and results are largely applicable to DEM also.

7Quality assessment methods can be distinguished a priori or a posteriori. Before generating the DTM, one can know the expected quality that result from our capacity and what quality is required with regard to the respected standards. These two factors enable regular production and usability of the DTM. The a priori assessments are based mostly on analyses of the datasets and methods for the DTM production while the a posteriori methods are based on the final DTM as described in this paper.

8One of the DTM quality assessment goals is to fulfil the requirements of spatial data standards. The ISO (International Organization for Standardization) distinguishes five elements of data quality: completeness; logical consistency; and three types of accuracy (positional, temporal, and thematic). This paper is concerned with accuracy, defined as a difference between the value of a variable, as it appears in a dataset, and the value of the variable in the data model (or “reality”). More specifically, we are referring to positional accuracy. We can distinguish between absolute and relative accuracy in terms of nature of the data. The position (horizontal or vertical) of the objects (e.g. ridges or sink holes as part of the DTM) could be assigned to absolute accuracy and the irregularity of the shapes of objects to the relative accuracy, that is, morphologically relative to a general position. The term precision is considered as a component of accuracy, related to the scale, resolution, and also to the generalisation of datasets (Podobnikar, 2008).

9The term error is used for lack of quality, or little or no accuracy. In addition to mistakes—in its widest meaning—it also refers to the statistical concept of variation (Burrough and McDonnell, 1998). The variation corresponds to random errors, thus incorrect spatial variation can be considered as systematic or gross error. According to these definitions, a level of accuracy (or error) can be described with a root mean square error (RMSE) and precision with a standard deviation or a standard error (σ).

  • 8
  • 9 This standard determines a grid size and accuracy according to different levels, from 0 to 5 (from (…)
  Send BMAC Down Under

10Most data quality standards for the DTMs encompass several quality requirements, but methods for quality control are seldom used. Visual quality control methods are even less often included. A certain level of standardisation is provided by USGS (1998). National Geospatial-Intelligence Agency (NGA)8 developed a “Digital Terrain Elevation Data” (DTED) standard for uniform matrix DTMs. It provides basic quantitative data for applications that require terrain elevation, slope, and/or surface roughness information9. The metadata of quality are roughly described with absolute horizontal (circular)/vertical (linear) error.

11EuroGeographics10 is currently developing a pan-European grid called EuroDEM11. Since the DTM is produced from various national DTMs, an important part of the project consists in the standardisation/harmonisation of the various coordinate systems, resolutions, and accuracies.

12The proposed procedure for quality assessment of the spatial datasets, especially of a DTM, comprises the following steps: (1) preparing the datasets; (2) processing with statistical or visual methods; (3) obtaining results as numbers, thematic maps, graphs, etc.; (4) analysis (comparison with expected results); and (5) obtaining metadata or corrected datasets (see figure 1).

Figure 1: The five-step procedure for quality assessment of a DTM

Figure 1: The five-step procedure for quality assessment of a DTM

13The procedure for quality assessment is based primarily on one (single) or multiple spatial datasets. In the case illustrated on Figure 1, one dataset is a tested DTM, while multiple datasets denote a DTM + (independent) reference datasets. The approach with one dataset uses a DTM alone without any reference data. This case is the most subjective and requires a high level of knowledge of the generation processes. The operator also needs to be experienced to recognise deviations from expected outputs and to predict the most useful kind of analysis. The approach with multiple datasets uses a DTM and additional reference datasets. The reference datasets can be the DTMs as regionally continuous data, lines, or points. The basic criterion for selecting the appropriate reference data is that the quality should be at least as high as expected from the tested DTM. The reference data should be representative (of sufficient quantity), therefore distributed with a certain degree of regularity and significance with respect to the whole area. These methods are not convenient for areas where availability of the reference datasets is very low (e.g. currently, Mars datasets).

14Processing with both statistical and visual methods is the primary focus of this research. The methods addressed differ according to whether they use one or multiple datasets and by their expected outputs. The single dataset method may allow more techniques for processing. These techniques may be used one after another. We classified them into two complexes: techniques using numerical processing and those using visualisation (Figure 1). Those in the first complex apply statistical and visual methods, while those in the second complex additionally apply visual methods only. With respect to visual methods, multiple techniques from complex 1 may be followed with single techniques of complex 2, and vice versa. Furthermore, some techniques of complex 1 can generate input for statistical methods but not for visual ones, some of them are useful just for visual methods, and the others for both statistical and visual methods. Statistical methods are denoted by /S/ and the visual by /V/. We propose the following classification of the methods:

15The results of the processing include numbers for the statistical assessment methods, and thematic maps, various non-spatial visualisations, and other approaches to visualisations for the visual assessment methods.

16The next step is the comparison of the results with what can be expected from the quality of the data model. This is done via statistical methods (e.g. calculated RMSE with allowed RMSE). The analysis of results of the visual methods is more complex and less objective. In this case the results are compared with the “thresholds” and already “established” models. The visual methods require experience obtained through training. Fortunately some visual methods are generated fairly effortlessly and are easily understandable by a wide audience (as in Figure 2).

17 As a final result, the datasets (DTMs) are evaluated by statistical or visual methods within the reports. Parameters of quality control are assessed and presented as extended standard metadata. An additional advantage is the opportunity for correction of the datasets—DTMs (Podobnikar, 2005).

18The statistical methods for quality control are also known as geometrical (when a topographic description of particular DTM objects is applied), stochastic (non-deterministic), or even mathematical (using mathematical methods). The most common approaches are analytical and empirical. The analytical approaches are primarily used when reference data is not available (Martinoni and Bernhard, 1998).

19The following parameters for quality assessment can be considered (descriptive statistic): arithmetical mean of heights, slopes, etc.; standard deviation σ; covariant function for heights, slopes, and volumes (Östman, 1987), rang (minimum/maximum), and Koppe’s formula adapted with other coefficients (Ackermann, 1978; Kraus, 1994); and autocorrelation analysis (Lee and Marion, 1994). The local methods entail description with variograms and correlograms (Wood, 1996; López, 2000) and measurement of the fractal dimensions of terrain (Wood, 1996) and terrain curvature.

20To analyse the estimated uncertainty of height data, Monte Carlo methods can be applied (Goodchild, 1995; Fisher, 1996; Podobnikar, 2005). The robust estimation method is based on statistical elimination of data that are not well enough autocorrelated to a certain threshold (Kraus and Pfeifer, 1998). Additionally, error assessment for the surroundings of a selected point on a surface may employ the “perfect inspector” hypothesis (López, 2000). A complex analytical method of spectral terrain analysis has been developed by Tempfli (1980; 1999), Frederiksen and Jacobi (1980), Russel et al. (1995), and Russel and Ochis (1995). The sensitivity analysis method was developed by Martinoni and Bernhard (1998). Accuracy can be also estimated by considering the density of the original datasets and local terrain curvature (Kraus et al., 2004).

21Another series of assessments includes various topological controls using vector contour lines developed to correct data in the following manners: nodes between two lines should have identical attributes; crossed lines should be eliminated; different heights of points and lines with identical coordinates should be unified; and contours with only one point (node) should be eliminated (Podobnikar, 2005; Figure 6). Other methods can be used to eliminate gross errors such as determination of the slopes that are too steeply inclined, and methods for determination of the height differences on the basis of control of neighbour contours (Larson, 1996).

22Possible methods using a DTM and additional reference DTM(s) include: computing a mean error (M) (indicator for a systematic error), root mean square error (RMSE) (indicator for a random error after the systematic component has been eliminated), range (minimum/maximum), and others. Furthermore, the following tests are proposed: statistical covariance, regression, histograms, volume differences, and others.

23The methods for comparison of the DTM with reference lines and points are similar to the methods described using continuous reference DTMs. The main difference is that their quality is expected to be much higher than that of the continuous reference data. Unfortunately, there is a high possibility that the reference data would not be available for areas where the quality of the DTM is already low. Another difficulty is that it is generally not possible to compute derivative surface, e.g. slope from the reference lines and points.

24The visual (or graphical, where the term is often applied in relation to geomorphological and semantical analysis) methods require a higher level of adaptation to particular problems than the more objective statistical ones. They are based on particular spatial analysis or modelling. Similar to cognitive mapping (Held and Rekosh, 1963), the use of visual method depends on the expertise and experience of the operator. The rule of thumb is more commonly applied with visual methods than with statistical methods. Visual methods actually offer the first assessments of the spatial data—DTMs. In the past they were carried out on a sheet of paper, nevertheless today they are primarily applied interactively with digital monitors (Burrough and McDonnell, 1998) and other equipment for the digital data visualisation (e.g. Drecki, 2002).

  D.T.M: What does DTM mean in Academic & Science? Department...

25This category of methods utilise the visual appearance of the dataset and is associated with thematic cartography and our ability to graphically express the studied problem. These methods can be roughly split into those that concern plasticity impression (embossing) and those that use geometric methods. For example, analytical shading as a plastic-oriented method (i.e. producing a three-dimensional impression) is based on visually effective presentation of the landform. In contrast, geometric methods like producing contour lines are better for a higher accuracy presentation of the landform. The methods of /V11/ may have some similarities with the methods /V1n/. Similar techniques may be used when comparing the DTM with its derivatives (reference datasets in /V1n/), but for this category only one dataset is used.

Figure 2. Shaded DTM

Figure 2. Shaded DTM

A shaded DTM with the original resolution of 100 m (A), and condensed to a resolution of 20 m using a spline interpolation algorithm (B). The red circle marks a gross error that is more easily recognised in the right picture. The visualisation is based on the /V11/.

26Visual controls of the basic derivatives of DTM include visualisation of slope, aspect (sensitivity to small errors especially on flat terrain), curvature (sensitivity to high frequency changes of the surface; Wood and Fisher, 1993), terrain roughness, dimension (characteristics) of the surface in a fractal sense (Li, 1998; Cheng et al., 1999), and visualisation of the condensed grid cells (Figure 2) or cost surfaces. These methods use different colour cast schemes, analytical shading with different parameters, or a dichromatic colour scheme (applying bipolar differentiation) with linear or non-linear cast (Wood, 1996; Rieger, 1992; Figure 3). The bipolar differentiation technique (or modulo approach, relative height-coding, “continuous” contour lines) can be described as a combination of contour lines (consecutive lines in the same colour of the dichromatic colour set) and repeated height-coding. Bipolar differentiation is similar to contours, but with different casts between them: a transition from light to dark or through a series of hues, which enables portrayal of even small details within the contour intervals. Depending on the chosen height interval, some tiny oscillations (possible errors) within “contour line” intervals can be clearly assessed, independently on the chosen particular azimuth as with analytical shading.

Figure 3. Example of dichromatic colour visualisation

Figure 3. Example of dichromatic colour visualisation

Figure 3. Visualisation based on /V11/ with a bipolar differentiation method with linear cast applying a certain height interval (20 m).

27The other methods are based on detection of seemingly impossible existing structures (e.g. the edges of the connection zone of the neighbour datasets) by applying high-pass filters; characterising the characteristic points, lines, and areas (peaks, pits, etc., or contour lines; Li, 1998); and searching for their false patterns (Figure 4).

Figure 4. Utilisation of false pattern to detect structures

Identification of the ridges and thalwegs based on /V11/. A: crossed contour lines (in circle) caused a false combination of ridge/thalweg (green and red areas are associated). B: incorrect attributes were assessed with a sensitive interpolation that presents analytical shading and ridges (red dots)/thalwegs (green dots) that are in unlikely positions.

28Further quality control methods include visualisation of the DTMs that were previously generalised. Additional techniques for generalisation make possible a multi-scale presentation. A combination of the proposed quality control methods in various scales can improve the reading and understanding of the landform features and therefore the finding of possible errors (Figure 5).

Figure 5. Morphological detection on Mars

Figure 5. Morphological detection on Mars

Detecting morphologically artificial (impossible) features on Mars (Candor Chasma) and labelling them as possible gross errors by applying different visualisation methods based on /V11/. A: analytical shading; B: bipolar differentiation with an interval of 100 m; C: curvatures visualisation; D: curvatures visualisation using a generalised DTM.

29The proposed methods are intended for checking consistency of the datasets when using reference data for the analyses. The reference data might be a better quality DTM, an orthophoto, contour lines from the maps, etc. For visualisation purposes the datasets can be previously reclassified, overlaid in different ways (e.g. transparently, using operations), or even placed alongside each other.

30This paper proposes and selects the following methods of spatial analytical operations with the multiple datasets visualisations: (1) difference between the overlaying DTMs; (2) combination of different type of derivatives of the DTMs (hypsometry, analytical shading, contour lines from the maps, contour lines from a DTM, etc.); (3) and contour lines from the maps overlaid over the following DTM derivatives: hypsometry, analytical shading, aspect, slope, curvature, or contour lines interpolated from the DTM (Ackermann, 1978; Hutchinson and Gallant, 1998; Carrara et al., 1997). The hydrological network can be assessed in a way similar to contour lines.

31The next methods use (4) contour lines vectorised from the maps which have been overlaid with characteristic points and lines derived from the contour lines (Figure 6)—the contours may be hierarchically coloured by applying a colour alternation method; (5) overlaying the hydrological network, generated from the DTM (Hutchinson and Gallant, 1998; Wood, 1996) over the pits and from hydrography acquired from the maps; (6) overlaying the contour lines from maps with the DTM generated from them (Carrara et al., 1997) or DTMs generated by other means; (7) overlaying the automatically generated characteristic points, lines, and contour lines; (8) overlaying the DTM with datasets that are basically not connected with DTM generation—satellite images, maps, orthophotos (Wiggenhagen, 2000); (9) overlaying considering Bayes theorem (Skidmore, 1997) where preliminary and actual knowledge is considered (Eastman, 1997); and (10) a perspective view applying the previously described methods for better recognition of the specific problems.

Figure 6. Contour lines obtained with Visual Methods

Figure 6. Contour lines obtained with Visual Methods

Visual methods based on /V11/ and /V1n/ (and on the statistical methods based on one dataset /S1/ that is not presented here) for detection of gross errors from the contour lines. A: contour lines from the original map (grey) and generated by a DTM (red). B: contour lines from the original map and an analytical shaded DTM generated from them. In both examples, a consequential gross error from the attributes (i.e. height of contour line) is easily perceived according to different methods.

32This set of methods is based on generating a selected statistical test of the dataset (DTM) and presenting the results in a way similar to the one described for the both classes of /V1/ methods. Firstly, we propose a group of methods based on Monte Carlo simulations: (1) visibility (Figure 7), slope and aspect, or optimal path simulation is applied by an appropriate error model of the DTM (Fisher, 1996; Podobnikar, 2005; Burrough and McDonnell, 1998; Heuvelink, 1998; Nackaerts et al., 1999; Felicísimo, 1994; Heuvelink, 1998; Canters, 1994; Ehlschlaeger and Shortridge, 1996; Ehlschlaeger et al., 1997); and (2) simulation of positional error of the hydrological network, watersheds, contour lines, characteristic features, and other vectors which have a significant influence on quality in certain circumstances (Burrough and McDonnell, 1998; Hutchinson and Dowling, 1991; Wood, 1996; Veregin, 1997; Lee, 1996; Openshaw, 1992; Podobnikar, 2005).

  Airborne LiDAR-Derived Digital Elevation Model for Archaeology

Figure 7. Monte Carlo simulation

Figure 7. Monte Carlo simulation

Monte Carlo simulation from a selected viewpoint (Krim) based on /V21/ and /V2n/ (comparing two different datasets). Two different models of error simulation on different DTMs were used. The DTM on A is a higher quality, especially on the plain. The Monte Carlo simulations applied specific error models (continuously varying error distribution surfaces) to the evaluated quality of DTMs with a resolution of 25 m—interferometric radar (IfSAR, A), and integrated DTM 20 m (B). The probability viewshed was converted to a fuzzy viewshed with a semantic import model (Burrough and McDonnell, 1998; Podobnikar, 2008), therefore to the fuzzy borders. Red indicates shadows, with a lower possibility of visibility. Hill shadows of tested DTMs are transparently overlaid;

33The next method entails (3) construction of fractal surfaces (Wood, 1996) similar to Monte Carlo approaches, where changing of the fractality allows controlled changing of the surface; (4) visualisation of precision and uncertainty of the contour lines, calculated with analytical methods (Tempfli, 1980; Kraus, 1994); and (5) visualisation of reference point difference according to the terrain surface, presented as deviation plots, that describes and portrays the quality of the DTMs’ surfaces.

34This class of visualisation methods is based on similar or completely different algorithms as for /V1/ and /V2/ classes. The outputs are histograms, graphs, diagrams, matrices, etc. Histograms as among the well known visual (graphical) presentation methods for certain statistic tests can be applied for DTM’s heights (Li, 1998) or derived aspects, curvatures, etc. (Hutchinson and Gallant, 1998). Histograms are then visually assessed: the DTM is expected to be of high quality if the transition between the columns is smooth enough or exhibits no repetitive pattern. Another possibility is a histogram of relative heights (so called relative histogram). If the DTM is interpolated from the contour lines then the values of DTM will tend to accumulate around the contour interval values. Higher perpendicularity (homogeneity) of the histogram signifies a higher quality of the interpolated surface (Carrara et al., 1997; Figure 8).

35The next proposed visualisation is a co-occurrence matrix calculation, used generally for analyses in a grey colour scheme. Using the DTM, the height values are assigned to the abscissa, and mean values of near surroundings to the ordinate. The autocorrelation of the surface can be inspected visually as it is higher when the values are closer to the principal diagonal (Wood and Fisher, 1993). Low autocorrelation signifies a very rough surface or a gross error.

Figure 8. Relative histogram for DTMs

Figure 8. Relative histogram for DTMs

Relative histogram for DTMs produced on a repetitive height interval of 10 m (0 to 9 m) based on /V31/ and /V3n/ (comparing two different datasets). On the left is a relative histogram for a DTM produced from contour lines (with interval 10 m) and on the right for a photogrammetrically generated DTM.

36There are many other possibilities for visually assessing a DTM’s quality. Several examples are presented below. The first is a path simulation between the selected points using different DTMs (Figure 9). This visualisation is actually bases on spatial analytical operations described in /V1/ but require some additional information besides the DTM (in this case the starting and the ending points). A very effective method is presenting terrain profiles (Figure 10) or terrain silhouettes from selected viewpoints. Another method demands motion picture techniques: attribute errors on the contour lines can be assessed, while the counter lines are presented sequentially according to their attributes or hierarchically from main to auxiliary ones. Another possibility is to label the contour lines according to their height (Hutchinson and Gallant, 1998).

Figure 9. Optimal path simulation

Figure 9. Optimal path simulation

Optimal path simulation using the same algorithm applied on three DTMs of different quality based on /V4n/. The black path is simulated on the highest quality DTM while blue one on the lower quality dataset. Similar results using DTMs produced from different sources signify (but do not prove) a higher quality.

Figure 10. Production of profile using DTMs

Figure 10. Production of profile using DTMs

Profiles over the same area on DTMs of different precision based on /V4n/. The appearance of the DTM on the A is very rough. It contains many gross errors and the overall quality is much lower than the one of the DTM on the B. These visualisations reflect the methods of the DTM production.

37Several methods have been developed, described and analysed, to assess DTM quality. This paper presents both statistical and visual methods, used for one (DTM) or multiple (DTM + reference) datasets. In particular, visual methods are presented in four classes: visualisations according to spatial analytical operations based on one dataset /V11/ or multiple datasets /V1n/; visualisations according to spatial statistical analysis /V2/; non-spatial visualisations /V3/; and other visualisation techniques/other algorithms /V4/. The first two classes result in thematic maps, while the third produces non-spatial visualisation.

38The visual methods (especially analytical shading) provide a first impression of the DTM quality. Although the methods for visual quality assessment of a DTM or other spatial datasets are less objective, they support statistical methods with their mutual combinations and combination with the other assessments, and allow understanding of even complex problems which may negatively influence the DTM quality and which otherwise would not be easily discovered. We can say that the statistical methods are well accepted for quality assessment, but they provide incomplete results, and vice versa. The examples are a quantification of the fuzzy viewsheds that would be additionally processed (see Figure 7) and quantifying/visualisations of the histograms (see Figure 8). The same examples also show a potential problem where the quality assessment is largely driven by a specific application. Additionally, more error types (e.g. random, systematic, and gross) could be assessed using the same visualisation method (see figures for examples).

39Results of the tests allow description of and improvement in quality in a sophisticated way considering the higher level of description and integrity of the processes. Consequently, the usability of the carefully checked and possibly corrected data can increase significantly. The proposed and applied methods considerably exceed available standards for the quality control used for the national or international DTM production (e.g. ISO/TC 211). The standards change frequently, and they are often based on the lowest common denominator—especially the subjective visual assessments. However, extensive experience combined with the complex knowledge thus acquired could be the most important factor in understanding the entire process of data acquisition, processing, etc. Furthermore, these checks provide an ideal opportunity to improve and extend the information content of standard metadata.

40In the future, more complex studies that include comprehensive simulation methods (Podobnikar, 2008) will be needed for visual quality assessment (ontologically, epistemologically, and pragmatically) to integrate outcomes of technical, natural, and social sciences and to reach a higher level of simplicity—as an ultimate level of sophistication (after Leonardo da Vinci).

Various techniques for quality assessment by visualisation have been carried out on different DTMs. Some of them were kindly provided by the Mapping Authority of the Republic of Slovenia through my doctoral thesis and others are DTMs of Mars available though the research project series TMIS (plus, plus.II, morph) funded by the Austrian Research Promotion Agency in the frame of the ASAP program. I am very grateful to Prof. Josef Jansa who performed a systematic review of my ideas.

Bạn đang xem bài viết: Methods for visual quality assessment of a digital terrain model. Thông tin do THCS Nguyễn Thanh Sơn chọn lọc và tổng hợp cùng với các chủ đề liên quan khác.

Leave a Comment