Vegetation cover recognition has mostly been studied with near-surface remote sensing images, using UAV or digital camera equipment to obtain high-resolution images, and then constructing vegetation indices in the visible band to extract the vegetation cover. Most researchers selected a variety of commonly used vegetation indices to compare the extraction effect; for example, the study of grassland in the school district selected four vegetation indices, and it was found that the cover extraction accuracy of the Vegetative Index (VEG) and EXG exceeded 93% (Fu et al.,
2021). In karst and rocky desertification areas, the accuracy of the Excess Green Minus Excess Red Index (EXGR) reached 99.174% (Yin et al.,
2020). Researchers have also proposed new vegetation indices, such as the New Green Red Vegetation Index (NGRVI), which was developed based on the construction principle of the Green Red Vegetation Index (GRVI) and Modified Green Red Vegetation Index (MGRVI), achieving an extraction accuracy of over 90% (Zhang et al.,
2019). Additionally, the Excess Green-Red-Blue Difference Index (EGRBDI) was constructed by drawing on the Red Green Blue Vegetation Index (RGBVI), and its overall accuracy, applicability, and stability outperformed 18 other vegetation indices (Gao et al.,
2020). The Difference Enhanced Vegetation Index (DEVI) was proposed with a supervised classification method based on Support Vector Machine (SVM) as the baseline for accuracy evaluation, showing significantly better extraction accuracy compared to eight other vegetation indices (Zhou et al.,
2021). Furthermore, a novel approach was introduced by discarding traditional index construction forms and proposing a method to find the best index through function optimization, resulting in the DeepIndices model within a deep learning framework. This model is unaffected by external factors and monitoring shapes, offering better segmentation effects and stability (Vayssade et al.,
2021). In addition to the vegetation index method, the Red, Green, Blue (RGB) decision tree method (Zhang et al.,
2013), the Hue, Saturation, Value (HSV) discriminant method (Chen et al.,
2014), and the Lab color space a-component method (Xu et al.,
2018) are commonly used, all of which have good extraction accuracy and can be quickly determined. All of the above belong to the image-oriented method, which is suitable for scenes with high image resolution and has the advantages of rapid determination, cost-effectiveness, efficiency, and accuracy compared with satellite remote sensing. However, this method separates image acquisition from cover extraction and requires centralized processing after acquiring images with a low degree of automation.