![]() | The topic of this article may not meet Wikipedia's general notability guideline .(July 2025) |
Accuracy assessment of land cover maps is the process of evaluating the reliability and quality of land cover maps. These maps are typically derived from remote sensing or other geospatial data sources using classification techniques. They play an important role in environmental monitoring, urban planning, and climate change studies, and accuracy assessment is essential for ensuring their reliability and usability. [1] [2] [3] [4] [5]
The accuracy of land cover maps is often assessed by comparison with reference data. These data are usually ground-based data or high-resolution imagery that is considered to represent the "true" land cover. Comparison of land cover maps with reference data can help identify misclassifications, and is often quantified using metrics such as overall accuracy, user's and producer's accuracy, and the Kappa coefficient. [5] [6]
In addition to validating individual maps with reference data, accuracy assessments may involve comparing different land cover products to evaluate their relative accuracy and suitability for various applications. [7]
Reference data (also called ground truth data or validation data) is used for assessing the accuracy of land cover maps. These data serve as the benchmark against which the classified land cover labels are compared, and their quality directly affects the effectiveness of the assessment. [5]
Reference data can be obtained from a variety of sources, including: [8]
Sampling refers to the procedure of selecting reference data. There are several common sampling strategies: [8] [9]
Selecting an appropriate sample size is an essential step in the validation design of land cover mapping. Two common ways to decide sample size are: [8] [10]
Sample interpretation refers to the assignment of a land cover class to each sample unit. There are several common sampling interpretation approaches: [8]
There are many quantitative metrics used to assess the accuracy of land cover maps. These metrics are usually derived from a confusion matrix (or error matrix), which summarizes the agreement between the classified map labels and the reference (ground truth) labels for a sample set. [5] [6]
Overall accuracy (OA) is an overall indicator, calculated as the proportion of correctly classified samples to the total number of samples. [5] [6]
Sometimes, it is valuable to report class-wise accuracy as well. [14]
User's accuracy and producer's accuracy are class-wise indicators. [6]
User's accuracy represents the probability that a pixel classified as a specific land cover class on the map actually corresponds to that class on the ground. Its complementary measure corresponds to the commission error. [6]
Producer's accuracy indicates the probability that a reference pixel of a specific land cover class is correctly classified on the map. Its complementary measure corresponds to the omission error. [6]
UA and PA can also be averaged separately to provide an overall perspective of classification performance from the user's and producer's perspectives. [15]
The F1-score combines UA and PA into one metric to measure the trade-off between them. It is the harmonic mean of UA and PA, where the relative contributions of the two metrics are equal. [6]
The Kappa coefficient [16] accounts for both omission and commission errors, as well as the possibility of chance agreement between the land cover maps and the reference data. Kappa values range from -1 to 1, and common rules of thumb for its interpretation are as follows: [17]
Kappa value | Strength of agreement |
---|---|
< 0 | Poor agreement |
0–0.20 | Slight agreement |
0.21–0.40 | Fair agreement |
0.41–0.60 | Moderate agreement |
0.61–0.80 | Substaintial agreement |
0.81–1.0 | Perfect agreement |
Since accuracy metrics are often sample-based, they are subject to uncertainty. The uncertainty of an estimate can be expressed by calculating its standard error or reporting a confidence interval. A confidence interval provides a range of values for a parameter, accounting for the uncertainty of the sample-based estimate. [18]
In addition to assessing the accuracy of a single land cover product, many studies [19] [20] [21] also conduct comparative evaluations across multiple land cover products. These products often differ in input data, classification schemes, or classification algorithms. Therefore, comparative evaluation is particularly important for understanding the consistency, differences, complementarity, and usability of these datasets. [7] [22]
Comparative evaluation is usually conducted in the following ways: [7] [22] [23] [24] [25]
Recent studies have compared high-resolution land cover products such as ESA WorldCover, Esir Land Cover, and Google's Dynamic World to assess their relative accuracy and thematic consistency across different regions and land cover types. These efforts help users make informed choices when selecting products for specific purpose. [7] [22]