29 January 2015 Framework for video-based monitoring of forest insect defoliation and discoloration
Author Affiliations +
Pest damage is a general problem that disturbs the growth of forests, influencing carbon sequestration and causing economic losses. In the past decades, many studies have been conducted to monitor and detect forest insect damage using satellite remote sensing technology. Satellite remote sensing has a satellite or aerial vision allowing the monitoring of extensive forest areas, but it usually requires constant time periods and is prone to cloud interference. To enable more efficient and effective monitoring of forest pest damage, a video-based monitoring framework is presented. This framework comprises three key parts: (1) video positioning of forest insect damage based on digital elevation model (DEM) and the parameters obtained from the pan-tilt-zoom camera, (2) integration of two-dimensional/three-dimensional geographic information system and video surveillance to provide more intuitionistic monitoring and assistance for positioning, (3) on-site verification conducted by ground surveys and guided through global positioning system (GPS) integrated in the embedded devices. The experiment was carried out over two forest areas to validate the proposed method. Results showed that the framework bears a sound positioning accuracy and high detection ratio, which could be effectively used in detecting and monitoring forest insect defoliation and discoloration.



There are clear ecological, economic, and social benefits associated with forests1 which are very important to biological diversity as well as soil and water conservation. However, disturbances are pivotal processes in forest ecosystem dynamics2 which strongly affect the composition, functioning, and structure of forest ecosystems,3 and the temporal and spatial patterns of forested landscapes.4 Forest monitoring and protection remains a significant and urgent task in China.5 Insect damage is an important forest disturbance that disturbs the growth of trees, influencing the production and quality of timber over large areas.6 Foliage discoloration and defoliation are significant variables by which to evaluate forest damage. Although there are multiple agents causing color alteration and loss of foliage, it is assumed that insect pests are the most common cause of foliage discoloration and defoliation.7

Currently, the methods used for monitoring of forest insect defoliation and discoloration mainly include field survey and remote sensing technologies. Field survey can diagnose insect damage more correctly than other approaches, but there is a problem if people cannot work continually, so it is difficult to collect damage information in a timely and comprehensive manner. For the monitoring using satellite remote sensing, researchers have explored numerous studies. For example, De Beurs and Townsend8 used MODIS imagery to estimate the magnitude of defoliation caused by the gypsy moth and concluded that MODIS data can be used to monitor insect defoliation on an annual time scale for large patches. Olsson et al.9 utilized SPOT and MODIS data to detect damage caused by Physokermes inopinatus, but insect attacks can be detected when the resulting defoliation or discoloration is sufficiently severe. Multitemporal Landsat TM images were used to estimate and classify defoliation in boreal coniferous forests,10 but a slight defoliation of coniferous forests was difficult to be detected from the remote sensing data used. In summary, due to low spatial and temporal resolutions of satellite images, it is difficult to identify forest insect damage in small area in a timely fashion. In addition, aerial videography has been proven to be an effective remote sensing tool for the detection of forest insect infestations.11 Video data provide a permanent record at the time of detection and, hence, eliminate the subjectivity of the observers. However, the background is dynamically changing for an airborne based video,12 and aerial videography has a high budget and cannot work continually in practical applications. In this situation, the development of distant video surveillance (DVS) technology provides a new opportunity for the monitoring of forest insect defoliation and discoloration.

DVS is an important new remote sensing method and has been widely used for monitoring. As a major application field, forest monitoring has played an important role in forest resource protection and management. In comparison to the watchtower observation, video monitoring can work continuously. In comparison to the satellite image monitoring, video monitoring has a low cost and is real time. For forest fires, various video-based monitoring systems have been proposed by researchers, and most of these systems can achieve reliable results based on smoke recognition during the day and flame recognition during the night.13,14 Compared with forest fires, the features of insect defoliation and discoloration are subtle, and this also brings some new challenges. For example, the forest insect damage monitoring has a higher requirement for positioning accuracy than forest fire. Therefore, it is difficult to directly apply forest fire video monitoring systems to forest insect damage monitoring.

To effectively reduce the damage of forest pests, the ability to enable command and control centers to have an intuitionistic view of the occurrences in the forest area in real time is very important. Although video surveillance technology is playing an increasingly important role in monitoring systems in various security and military applications,15 most of them do not support geospatial technology integration and analysis from a spatial perspective. Based on the above, the main objective of this study is to design and implement a framework for forest insect defoliation and discoloration monitoring, relying on the use of geospatial technology and video monitoring technology. In this framework, satellite remote sensing is mainly responsible for identifying susceptible areas of forest insect infestations. The susceptibility maps can be used for guiding monitoring strategy in the video surveillance stage. Remote sensing technology has been extensively studied and described by other authors, so we will not elaborate on it. Video monitoring is utilized to visualize and identify forest insect damage, to guarantee real-time monitoring by using a pan-tilt-zoom (PTZ) camera with three degrees of freedom because of its ability to pan, tilt, and zoom. On-site verification is conducted by ground surveys and guided through global positioning system (GPS) integrated in the embedded devices. Furthermore, the proposed approach bridges the gap between aerial monitoring and ground surveys to form a comprehensive monitoring framework for forest insect defoliation and discoloration.





Forest pest damage has obvious visual and dynamical features, and they are usually characterized by defoliation and discoloration. Visual detection of an infested stand is not straightforward in many cases, particularly in large and dispersed forests. Video surveillance technology has been called to address this issue mainly due to two reasons: first, DVS can visualize the forests continually to check the health of woods beyond our own eyes; second, DVS has a wider vision and can monitor extensive forest areas and record these data in real time. In addition, to our knowledge, the application of video surveillance technology for forest insect defoliation and discoloration is rarely reported. This paper focuses on the integrated monitoring framework and explores a new application which will be very helpful for further study.

Compared with previous methods, the video-based method can provide continual and real-time monitoring, and the video data can be stored for a long time. Video surveillance techniques have many attractive attributes, but the most prominent are the real-time monitoring capability and the immediate availability of the electronic signal for both visual interpretation and digital processing.16 Visual images can convey more memorable and meaningful information than written words, numbers, or other types of media.17 Visualization refers to any technique for creating images, diagrams, or animations to communicate a message.18 The visualization of forest scene is essential for forestry departments to detect disturbances and make decisions for pest control. With the development of computer hardware and software, it is possible to visualize complex natural phenomena, such as a three-dimensional (3-D) scene or landscape. Visualization means to represent data in an intuitionistic and perceptible form that can be sensed by people’s vision. The 3-D visualization for forest insect defoliation and discoloration monitoring is supported by a geographic information system (GIS) and is characterized by multiperspective views, depending on the relationships with observers from different directions, distances, or locations.

In the proposed method, satellite remote sensing monitoring is mainly used to identify the susceptible areas of forest insect infestations in order to provide support to make appropriate monitoring strategies that reduce the impact of disturbances.19 The remote sensing sources used to identify the susceptibility areas should be high-resolution images, because forest insect infestation is often studied at the individual tree scale.20 DVS is used to capture video images to integrate with GIS, to provide intuitionistic display for the command and control center, and to obtain the geographic coordinates of insect damage for field survey. Then ground survey by human walkthrough is mainly responsible for conducting field verification and to simultaneously provide complementary details that video images lack. In addition, for those forest areas difficult to access, satellite imagery and video images are often the only means of gathering information. The main framework diagram is shown in Fig. 1. A detailed description of the above-mentioned procedures will be given in the following part of the article.

Fig. 1

Framework diagram of the proposed comprehensive monitoring method for forest insect defoliation and discoloration.



Video Positioning of Forest Insect Damage

For the system implementation, we used a high-definition CCD camera, digital PTZ, zoom lens, protective slab, decoder, watchtower, and power supply facilities. In the process of video surveillance, PTZ can control the camera to rotate at the horizontal (pan) and vertical planes (tilt) and to change the level of magnification (zoom). The camera, fixed on the PTZ, is responsible for capturing video images. The PTZ can control the camera to rotate smoothly in the horizontal direction and in the vertical direction, to guarantee complete coverage. Briefly, the front-end equipment and the monitoring center are connected through a wireless network.

The geographical position of forest insect damage is traditionally determined by comparing the video scene with a corresponding map. There are some unavoidable subjective factors that affect the accuracy. To improve the accuracy, we designed an automatic positioning algorithm for forest insect damage based on digital elevation model (DEM) and returned parameters from the PTZ camera. The principle of the disaster positioning is to build a ray emitting from the monitoring camera according to the orientation parameters returned from the PTZ. From the perspective of geometry, the process of the disaster positioning is to get the intersection point of the DEM and the ray. The DEM cannot be directly expressed by a mathematical function. Thus, in a discrete space, disaster positioning can be seen as a process of finding the optimal solution regarding the constraint satisfaction problem. To improve the efficiency of the algorithm, the viewshed analysis was used to generate the viewshed layer, which could reduce the search space of the optimal solution. To generate the ray, the Bresenham algorithm21 was used which adopts incremental computation and is a classical algorithm to generate a line in computer graphics. Thus, for each column, just check the sign of an error item, and then the pixel of this column is selected. This algorithm does not include the operation of real numbers, so it can generate a line quickly.

Figure 2 illustrates the spatial model for disaster positioning, where M represents the monitoring point and P is the disaster point. α and β are the azimuth angle and pitch angle, respectively. θ is a vertical angle in the direction of α. L is a ray from the monitoring point. Based on this model, the flow chart of the positioning algorithm is shown in Fig. 3, where H represents the height from the camera to the ground. (Xc, Yc, Zc) are the projective coordinates of the monitoring point.

Fig. 2

Spatial model for forest insect damage positioning. The meaning of parameters: M, monitoring point; P, disaster point; L, a ray from M; α, the azimuth angle of P; β, the pitch angle of P; θ, a vertical angle in the direction of α.


Fig. 3

Flow chart of positioning forest insect damage. The meaning of parameters: (Xc, Yc, Zc), the projective coordinates of the monitoring point; H, the height from the camera to the ground; Di (i=1n), the discrete points of the ray L; θDk (k=1m), a vertical angle in the direction of α.



Integration of GIS and Video Surveillance

The hardware architecture depicting the system context is shown in Fig. 4. For real-world image capture, several PTZ cameras are utilized. A PTZ camera has three degrees of freedom because of its ability to rotate in the horizontal (pan) and vertical planes (tilt) and to change the level of magnification (zoom).

Fig. 4

Hardware structure of distant video monitoring.


Three servers are deployed at the server-side and are connected to the Internet with a speed of 10GB/s. The video server uses the D-Link DNS-726-4 to provide the load balancing service and is utilized to distribute front-end video streams and to provide load balancing service for an increasing number of users. The data server can provide basic spatial data service and pest information storage service for the application server. The DVS system is deployed on the application server using a middleware of Internet Information Services 7.0. As shown in Fig. 5, the software structure of the DVS system consists of two-dimensional (2-D) GIS, 3-D GIS, video control, disaster positioning, and communication control. In comparison to previous research, our method integrates 2-D GIS and 3-D GIS synergistically. This can make the monitoring process more intuitionistic.

Fig. 5

Software structure of distant video monitoring.


Functionalities of the 2-D GIS component are related to 2-D visualization, layer management, geospatial data access, feature querying, and identification. Furthermore, the 2-D GIS can navigate monitoring points and aid disaster positioning. In addition, the 2-D GIS can serve as an eagle-eye map of the 3-D GIS component. The 3-D GIS component is responsible for the visualization of the DEM and 2-D geospatial data of the forest area. The 2-D geospatial data are displayed as a texture that is laid over the virtual terrain. Moreover, the disaster points can be added and displayed in the 3-D GIS context.

The video control component is responsible for controlling the camera to pan, tilt, and zoom as well as for setting the window size and the bitrate of the camera video. The communication control component is responsible for the acquisition of the camera’s orientation parameters as well as for communications between different components. The component for disaster positioning is responsible for calculating the geographic coordinates of the forest pest disasters.


Field Survey

In order to evaluate the method, ground survey by human is used to collect field data for on-site verification. The field survey is used for both forward validation and reversed validation (Fig. 6). Forward validation is used to assess the accuracy of damage detection using video surveillance. Reversed validation is used to evaluate the completeness of damage detection. Due to the limited memory and screen size of the embedded equipment, the massive spatial data need to be preprocessed, such as establishing the image pyramid and segmenting the vector data.22 In the system, the ShapeFile format is used for storing all the vector data because it uses less storage space. The pyramid images are managed in FGI format, which consists of a master file (.fgi), a secondary file (.aux), and a coordinate file (.txt). After downloading the processed data into the embedded GIS, field surveys can be conducted to collect the spatial data, attribute information of the pest disasters, and upload the results to the server in time.

Fig. 6

Flow chart of field surveys for validation.


Field verification is aided by smartphones or personal digital assistant (PDA) with GPS embedded (Fig. 7). After the geographical coordinates of the forest pest disasters are obtained, the thematic map of disasters can be output and imported into PDA, and afterward the foresters are able to conduct on-site verification following the optimal path planned in advance. In addition, to guarantee the reality and authenticity of the field data, the proposed approach can provide supervision and inspection to enable managers to avoid fake data to some extent.

Fig. 7

Field survey for data acquisition.



Experiments and Analysis


Application Effect

The DVS system is implemented in Microsoft Visual Studio 2010, as a Visual C# project. Browser/Server structure is utilized in this system. The main user interface is composed of three components: 2-D view, 3-D view, and video view (Fig. 10). Besides these views, the application interface contains the disaster information table (bottom-left) to record the disaster information including the monitoring point names, the latitude and longitude of disasters, damage degree, and date, which is simultaneously stored in the database. The first component, 2-D view, provides the overall perspective of the forest area based on the forestry data and other geographic spatial data. The component is implemented purely by the web pattern. Basic GIS functions related to the view are map zooming, map panning, object identification, and layer management. Users can choose to display the selected layers of geospatial data according to actual needs.

The second component, 3-D view, is responsible for the generation and visualization of the virtual terrain. The component is implemented based on OpenSceneGraph, which is an open source 3-D graphic engine. At the data layer, Geospatial Data Abstraction Library 1.8.0 is used to analyze the spatial data. At the network layer, the library of libcurl 7.21.3 is used to transmit the Web Map Service terrain/image pyramid models. At the function layer, the terrain rendering library of libMini 10.2 is utilized to implement 3-D terrain visualization. Functions for the 3-D view include roaming, zooming, and visibility analysis. Visibility analysis is used to determine whether a given object located on a terrain is visible from a viewpoint and how much of the object is visible.23 In addition, when the coordinates of disaster points are obtained, they can be simultaneously added to the 3-D scene.

Finally, the third component, video view, displays the on-site video of a selected surveillance camera. With this view, it is feasible and flexible to manipulate the camera by changing the parameters of pan, tilt, and zoom, or utilizing the mouse as a virtual joystick (Fig. 8).

Fig. 8

Video monitoring in application mode that displays video control.


During the process of scanning, when finding dubious phenomena, the PTZ parameters can be adjusted to clearly observe and analyze the video images to detect forest pest disasters. Then, combined with DEM and the angles of azimuth and pitch of the PTZ camera, the geographical coordinates of the disasters can be obtained through the positioning algorithm. Then the disaster information is stored into the disaster database. Users can query disaster by single field or multifield, and the query results can be simultaneously visualized in the 2-D view. In addition, the thematic map can be conveniently output and printed, to aid forestry surveyors to conduct field verification. Furthermore, the 2-D view, 3-D view, and video view can interact with each other. For example, when the user selects a monitoring point, the 3-D view and the video view will transform accordingly. Thus, the user can monitor the on-site situation in real time to detect forest pest disasters timely and accurately. In addition, the thematic map can be generated and printed, providing great convenience for the foresters to conduct on-site verification.

For on-site verification, the system for ground surveys (Fig. 9) is used, which includes the mobile terminal and the server terminal. The disaster information is transmitted to the system for ground surveys by extensible markup language (XML). The mobile terminal is developed by VC++. NET, and MSXML3.0 is used to parse the XML schema documents. The server terminal is developed by VC++6.0, using data access object to operate the database and MSXML6.0 to establish the dynamic form. Web services are used in the server terminal to receive the field survey data from the mobile terminal. The embedded system is connected with the server through a general packet radio service and the survey results are uploaded using these services. The implementation of the system for ground surveys relies on the previous work summarized in Ref. 22.

Fig. 9

Video monitoring integrated with two-dimensional geographic information system (GIS) and three-dimensional GIS.



Experiment Results and Validation

To evaluate the availability of the positioning algorithm, accuracy verification was conducted. First, differential GPS was used to measure the geographic coordinates of simulated disaster points. Combined with local three parameters, the geographic coordinates were transformed into projective coordinates. Next, the simulated disaster points were positioned through the proposed video positioning method. Finally, these two measurement results were compared. The projective coordinates and elevation of the simulated monitoring point are (547,366, 4,466,675, 1420).

From Table 1, we can see that the maximum error for X is 49 m and for Y is 97 m. The minimum discrepancy for X is 21 m and for Y is 45 m. These discrepancies are acceptable for forest pest damage monitoring.24

Table 1

Video monitoring positioning, GPS measuring, and their discrepancies.

IDVideo monitoring positioning (m)GPS measuring (m)Discrepancies (m)

Field verification is often considered the preferred method for evaluating insect damage detection derived from video remote sensing. In order to better assess the potential of the proposed approach, two forest areas are selected for field validation. Jinghai County is located in southwestern Tianjin (Fig. 10), with an area of 1476km2. Longitude ranges from 116°42′E to 117°12′E, and latitude ranges from 38°35′N to 39°04′N. The vegetation types of this area are mainly broadleaved deciduous forest. The Hyphantria cunea damage in this area is representative of this city and was, therefore, selected as the test site for this study. Qianshan County is located in southwestern Anhui (Fig. 10), with an area of 1686km2. Longitude ranges from 116°14′E to 116°41′E, and latitude ranges from 30°27′N to 31°04′N. The vegetation types of this region are mainly pinewood and mixed deciduous forest. Planted forests account for 88.4% of the whole county, resulting in single forest form and poor biodiversity, which facilitates the occurrence and spread of insect pests. The Dendrolimus punctatus walker damage in this region is representative of the province, so it was selected as one experimental area. Both Hyphantria cunea and Dendrolimus punctatus walker occur with three generations in a year.

Fig. 10

Forest areas for field validation. One is in Jinghai County of Tianjin, and the other is in Qianshan County of Anhui Province.


To quantitatively evaluate the proposed method, we employ the forward detection ratio (fDR) and reversed detection ratio (rDR) as the assessment principles, which are defined as

where Numfield represents the number of detected real damage patches from the field survey and Numvideo represents the number of detected damage patches from video images.

According to the different stages of insects, we selected sampling points (Fig. 11) according to the visual features of discoloration and defoliation (Fig. 12). Then, field surveys were carried out to validate the detection results. The data are obtained from the local forest bureau. To further demonstrate the method, we also carried out a reversed field validation task. We selected 20 sampling points by field surveys, respectively, in the two areas and validated them by video monitoring. Tables 2 and 3 show the field validation results.

Fig. 11

Sampling points for accuracy validation: (a) test site in Jinghai County and (b) test site in Qianshan County.


Fig. 12

Pictures from field surveys showing forest insect defoliation and discoloration.


Table 2

Accuracy validation of site 1.

Test siteObjectiveDateVideo detectionField surveyfDR
JinghaiForward validationMay 15, 201319160.84
July 15, 201330250.83
15 September, 201320170.85
ObjectiveDateField surveyVideo detectionrDR
Reversed validationMay 15, 201320180.90
15 July, 201320190.95
15 September 201320180.90

Table 3

Accuracy validation of site 2.

Test siteObjectiveDateVideo detectionField surveyfDR
QianshanForward validation23 April, 201317150.88
23 June, 201330260.87
23 August, 201320170.85
ObjectiveDateField surveyVideo detectionrDR
Reversed validation23 April 201320160.80
23 June, 201320170.85
23 August, 201320160.80


Comparative Experiment

Here, an assessment of accuracy was conducted for the detection and positioning of forest insect damage. Currently, the identification of forest insect damage by video cameras is usually conducted by a forest fire video monitoring system. Due to low definition of the camera, it is difficult to identify forest insect damage. In our method, we selected a high-definition camera, so the identification rate is improved, as shown in Table 4. The fDR and rDR of the current method are 0.79 and 0.85 in site 1, respectively. The fDR and rDR of the proposed method are 0.84 and 0.92 in site 1. The fDR and rDR of the current method are 0.81 and 0.76 in site 2. The fDR and rDR of the current method are 0.87 and 0.82 in site 2. Experiments show that the accuracy of the proposed method is higher than the current method.

Table 4

Comparison of the forest insect damage detection method by current method and proposed method.

Test siteCurrent methodProposed method

Traditionally, the geographical position of forest insect damage is determined by comparing the video scene with the corresponding map. There are some unavoidable subjective factors that affect the accuracy. We utilized the traditional method to determine the position of the simulated disaster points, and the comparison results are shown in Table 5. The overall accuracy is calculated as Overall=(ΔX)2+(ΔY)2+(ΔZ)2. Experiments show that the proposed positioning method has better overall accuracy than the traditional method. In addition, the proposed method is automatic and avoids the subjectivity of observers to some extent.

Table 5

Comparison of the positioning method by traditional method and proposed method.

IDTraditional methodProposed method
ΔX (m)ΔY (m)ΔZ (m)Overall (m)ΔX (m)ΔY (m)ΔZ (m)Overall (m)



Three are some other types of GIS-based video monitoring systems. Zhang et al.25 presented a prototype system VSGIS integrating 2-D GIS and video surveillance, and it was applied in city safety management. Zong et al.26 combined GIS and video monitoring technology to design a GIS-based video surveillance system by constructing an electronic map with video linkage to display control points. In comparison to them, the proposed method integrates 2-D GIS and 3-D GIS simultaneously, and the three modules are geographically linked and displayed in the same window. This is the basis of implementing the fusion of video images and 3-D GIS in future studies.

The study clearly highlights the potential for video surveillance and geospatial technologies to monitor forest pest disturbances, and the ability for positioning and mapping damage in a timely manner. This monitoring technique tends to be proactive rather than passive aiming to complement the existing methods. An operational surveillance system may detect forest pest disasters as early as possible to reduce losses. Although the approach is designed for monitoring forest insect defoliation and discoloration, it could be modified for monitoring forest resources, ecological environment, and other disturbances. Forest pest damage surveillance is established as part of forest health management to ensure the early detection of pest threats to tree health and to respond in time before there is significant damage. From the perspective of ecology, in the long run, this can facilitate sustainable forestry development.

The study focuses on the monitoring and detecting of the damage characterized by defoliation and discoloration in the sense of human–computer interaction. Little effort, however, was spent on researching on the damage detection and recognition from the video images automatically. Nonetheless, according to the literature retrieval, there is hardly any research on this issue. This issue is a complicated problem that refers to various subjects and fields, such as computer vision, intelligent video analysis, image processing, and data mining. Furthermore, forest pest damage detection and recognition also involves a substantial amount of heuristic knowledge, requiring much effort for domestic and international researchers to carry on a great deal of research.

This study aimed to present an integrated application to monitor forest insect damage. It was not, nor could it be our intention, to evaluate the cost benefit of the system. This procedure should be performed by the competent authorities.



In this study, an integrated framework for video-based monitoring of forest insect defoliation and discoloration was presented. By using video monitoring, we could extend the spatial coverage and number of samples far beyond what was obtained with field surveys. Video images make the monitoring procedure more intuitive and perceptual. The foresters can remotely control the video cameras at monitoring spots and observe the video images displayed on the screens in the monitoring center. Moreover, video images are obtained in real time, thus the forest insect disturbances can be detected as early as possible. Compared with traditional human ground observation, using video cameras the human observer is capable of monitoring a wider area. Cameras equipped with power zoom facilitate the observer to inspect suspected areas. Video data can be stored and this is quite useful for postdisaster analysis. In addition, there are some other issues worthy of consideration. For example, what are the spatial distribution and density of the cameras needed for the DVS system to adequately monitor the forest area depending on landscape traits? This will be our future studies.


The authors would like to thank the reviewers whose comments were very helpful in improving this paper. The work is supported by National Science and Technology Major Project of the Ministry of Science and Technology of China (21-Y30B05-9001-13/15), the National High Technology Research and Development Program of China (863 Program No. 2008AA12Z203), and the National Science and Technology Support Project (No. 2006BAD23B04).



Cao C. et al., “Extraction of forest structural parameters based on the intensity information of high-density airborne light detection and ranging,” J. Appl. Remote Sens., 6 (1), 063533 (2012). http://dx.doi.org/10.1117/1.JRS.6.063541 1931-3195 Google Scholar


Bengtsson J. et al., “Biodiversity, disturbances, ecosystem function and management of European forests,” For. Ecol. Manage., 132 (1), 39 –50 (2000). http://dx.doi.org/10.1016/S0378-1127(00)00378-9 FECMDW 0378-1127 Google Scholar


Franklin J. F. et al., “Disturbances and structural development of natural forest ecosystems with silvicultural implications, using Douglas-fir forests as an example,” For. Ecol. Manage., 155 (1), 399 –423 (2002). http://dx.doi.org/10.1016/S0378-1127(01)00575-8 FECMDW 0378-1127 Google Scholar


Kennedy R. E. et al., “Spatial and temporal patterns of forest disturbance and regrowth within the area of the Northwest Forest Plan,” Remote Sens. Environ., 122 117 –133 (2012). http://dx.doi.org/10.1016/j.rse.2011.09.024 RSEEA7 0034-4257 Google Scholar


Zhang F. et al., “Forest and deforestation identification based on multitemporal polarimetric RADARSAT-2 images in Southwestern China,” J. Appl. Remote Sens., 6 (1), 063527 (2012). http://dx.doi.org/10.1117/1.JRS.6.063527 1931-3195 Google Scholar


Eklundh L., Johansson T., Solberg S., “Mapping insect defoliation in Scots pine with MODIS time-series data,” Remote Sens. Environ., 113 (7), 1566 –1573 (2009). http://dx.doi.org/10.1016/j.rse.2009.03.008 RSEEA7 0034-4257 Google Scholar


Ciesla W. et al., “Aerial signatures of forest damage in the eastern United States,” (2015) http://www.fs.fed.us/foresthealth/technology/pdfs/AerialSignaturesEast.pdf January ). 2015). Google Scholar


De Beurs K. M., Townsend P. A., “Estimating the effect of gypsy moth defoliation using MODIS,” Remote Sens. Environ., 112 (10), 3983 –3990 (2008). http://dx.doi.org/10.1016/j.rse.2008.07.008 RSEEA7 0034-4257 Google Scholar


Olsson P. O., Jonsson A. M., Eklundh L., “A new invasive insect in Sweden—Physokermes inopinatus: tracing forest damage with satellite based remote sensing,” For. Ecol. Manage., 285 29 –37 (2012). http://dx.doi.org/10.1016/j.foreco.2012.08.003 FECMDW 0378-1127 Google Scholar


Heikkilä J., Nevalainen S., Tokola T., “Estimating defoliation in boreal coniferous forests by combining Landsat TM, aerial photographs and field data,” For. Ecol. Manage., 158 (1), 9 –23 (2002). http://dx.doi.org/10.1016/S0378-1127(00)00671-X FECMDW 0378-1127 Google Scholar


Honggan W., Zhennan X., Jin S., “Detecting technique of forest diseases and insects by airborne video,” Remote Sens. Technol. Appl., 19 (1), 10 –15 (2004). http://dx.doi.org/10.3969/j.issn.1004-0323.2004.01.003 Google Scholar


Liu K. et al., “Fast motion detection from airborne videos using graphics processing unit,” J. Appl. Remote Sens., 6 (1), 061505 (2012). http://dx.doi.org/10.1117/1.JRS.6.061505 1931-3195 Google Scholar


Surit S., Chatwiriya W., “Forest fire smoke detection in video based on digital image processing approach with static and dynamic characteristic analysis,” in IEEE First ACIS/JNU Int. Conf. on Computers, Networks, Systems and Industrial Engineering, 35 –39 (2011). Google Scholar


Kim D. K., Wang Y. F., “Smoke detection in video,” in IEEE WRI World Congress on Computer Science and Information Engineering, 759 –763 (2009). Google Scholar


Sebe I. O. et al., “3D video surveillance with augmented virtual environment,” in First ACM SIGMM Int. Workshop on Video Surveillance, 107 –112 (2003). Google Scholar


Everitt J. H. et al., “Using airborne video, global positioning system, and geographical information system technologies for detecting and mapping citrus blackfly infestations,” Southwest. Entomol., 19 (2), 129 –138 (1994). SENTDD 0147-1724 Google Scholar


Biljana D., “The role of visualization in forest planning,” Publications Oboulo.com, (2008) http://www.oboolo.com/politics-international/ecology-environment/presentation/the-role-of-visualization-in-forest-planning-61713.html Google Scholar


Wang X. et al., “Challenges in visualizing forests and landscapes,” J. For., 104 (6), 316 –319 (2006). Google Scholar


Wulder M. A. et al., “Digital high spatial resolution aerial imagery to support forest health monitoring: the mountain pine beetle context,” J. Appl. Remote Sens., 6 (1), 062527 (2012). http://dx.doi.org/10.1117/1.JRS.6.062527 1931-3195 Google Scholar


Bone C. et al., “Integrating high resolution remote sensing, GIS and fuzzy set theory for identifying susceptibility areas of forest insect infestations,” Int. J. Remote Sens., 26 (21), 4809 –4828 (2005). http://dx.doi.org/10.1080/01431160500239180 IJSEDK 0143-1161 Google Scholar


Bresenham J. E., “Algorithm for computer control of a digital plotter,” IBM Syst. J., 4 (1), 25 –30 (1965). http://dx.doi.org/10.1147/sj.41.0025 IBMSA7 0018-8670 Google Scholar


Qiao Y. et al., “Field investigation system and its application in forest pest survey,” Proc. SPIE, 8203 820302 (2011). http://dx.doi.org/10.1117/12.910360 PSISDG 0277-786X Google Scholar


De Floriani L., Magillo P., “Intervisibility on terrains,” Geogr. Inf. Systems, 1 543 –556 (1999). Google Scholar


Wang Z. et al., “A positioning system for forest diseases and pests based on GIS and PTZ camera,” IOP Conf. Ser.: Earth Environ. Sci., 17 (1), 012254 (2014). http://dx.doi.org/10.1088/1755-1315/17/1/012254 1755-1307 Google Scholar


Zhang X. et al., “Video surveillance GIS: a novel application,” in IEEE 21st Int. Conf. on Geoinformatics, 1 –4 (2013). Google Scholar


Zong W. J. et al., “Research on GIS-based video surveillance system,” Comput. Eng. Des., 32 (2), 745 –748 (2011). Google Scholar


Feifei Zhao is currently a PhD candidate in the Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences. Her current research interests include remote sensing for forestry, video monitoring, and geographic information system.

Yafei Wang is currently a PhD candidate in the Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Sciences. His major research interests include remote sensing data mining and geographic information system (GIS) application.

Yanyou Qiao is a professor in the Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences. He focuses on forest GIS and spatial decision support.

© The Authors. Published by SPIE under a Creative Commons Attribution 3.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Feifei Zhao, Feifei Zhao, Yafei Wang, Yafei Wang, Yanyou Qiao, Yanyou Qiao, "Framework for video-based monitoring of forest insect defoliation and discoloration," Journal of Applied Remote Sensing 9(1), 096093 (29 January 2015). https://doi.org/10.1117/1.JRS.9.096093 . Submission:


Back to Top