Mapping of Land Cover with Open‐source Software and Ultra‐high‐resolution Imagery Acquired with Unmanned Aerial Vehicles

Reference
Horning, N., Fleishman, E., Ersts, P. J., Fogarty, F. A., & Wohlfeil Zillig, M. (2020). Mapping of Land Cover with Open‐source Software and Ultra‐high‐resolution Imagery Acquired with Unmanned Aerial Vehicles. Remote Sensing in Ecology and Conservation, 6(4), 487–497. https://doi.org/10.1002/rse2.144
Abstract

The use of unmanned aerial vehicles (UAVs) to map and monitor the environment has increased sharply in the last few years. Many individuals and organizations have purchased consumer-grade UAVs, and commonly acquire aerial photographs to map land cover. The resulting ultra-high-resolution (subdecimeter-resolution) imagery has high information content, but automating the extraction of this information to create accurate, wall-to-wall land-cover maps is quite difficult. We introduce image-processing workflows that are based on open-source software and can be used to create land-cover maps from ultra-high-resolution aerial imagery. We compared four machine-learning workflows for classifying images. Two workflows were based on random forest algorithms. Of these, one used a pixel-by-pixel approach available in ilastik, and the other used image segments and was implemented with R and the Orfeo ToolBox. The other two workflows used fully connected neural networks and convolutional neural networks implemented with Nenetic. We applied the four workflows to aerial photographs acquired in the Great Basin (western USA) at flying heights of 10 m, 45 m and 90 m above ground level. Our focal cover type was cheatgrass (Bromus tectorum), a non-native invasive grass that changes regional fire dynamics. The most accurate workflow for classifying ultra-highresolution imagery depends on diverse factors that are influenced by image resolution and land-cover characteristics, such as contrast, landscape patterns and the spectral texture of the land-cover types being classified. For our application, the ilastik workflow yielded the highest overall accuracy (0.82–0.89) as assessed by pixel-based accuracy.