<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. ICT</journal-id>
<journal-title>Frontiers in ICT</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. ICT</abbrev-journal-title>
<issn pub-type="epub">2297-198X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fict.2017.00004</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>ICT</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Algorithmic Identification of Looted Archaeological Sites from Space</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Bowen</surname> <given-names>Elijah F. W.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/396731"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Tofel</surname> <given-names>Brett B.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/407982"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Parcak</surname> <given-names>Sarah</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://frontiersin.org/people/u/425457"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Granger</surname> <given-names>Richard</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="cor1">&#x0002A;</xref>
<uri xlink:href="http://frontiersin.org/people/u/2512"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Psychological and Brain Sciences, Dartmouth College</institution>, <addr-line>Hanover, NH</addr-line>, <country>USA</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Anthropology, University of Alabama at Birmingham</institution>, <addr-line>Birmingham, AL</addr-line>, <country>USA</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Florence Forbes, Institut national de recherche en informatique et en automatique (INRIA), France</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: K. C. Santosh, University of South Dakota, USA; Peter Savadjiev, Harvard Medical School, USA</p></fn>
<corresp content-type="corresp" id="cor1">&#x0002A;Correspondence: Richard Granger, <email>richard.granger&#x00040;dartmouth.edu</email></corresp>
<fn fn-type="other" id="fn002"><p>Specialty section: This article was submitted to Computer Image Analysis, a section of the journal Frontiers in ICT</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>24</day>
<month>04</month>
<year>2017</year>
</pub-date>
<pub-date pub-type="collection">
<year>2017</year>
</pub-date>
<volume>4</volume>
<elocation-id>4</elocation-id>
<history>
<date date-type="received">
<day>08</day>
<month>12</month>
<year>2016</year>
</date>
<date date-type="accepted">
<day>09</day>
<month>03</month>
<year>2017</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2017 Bowen, Tofel, Parcak and Granger.</copyright-statement>
<copyright-year>2017</copyright-year>
<copyright-holder>Bowen, Tofel, Parcak and Granger</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>In response to widespread looting of archaeological sites, archaeologists have used satellite imagery to enable the investigation of looting of affected archaeological sites. Such analyses often require time-consuming direct human interpretation of images, with the potential for human-induced error. We introduce a novel automated image processing mechanism applied to the analysis of very high-resolution panchromatic satellite images and demonstrate its ability to identify damage at archaeological sites with high accuracy and low false-positive rates compared to standard image classification methods. This has great potential for large-scale applications whereby countrywide satellite datasets can be batch processed to find looting hotspots. Time is running out for many archaeological sites in the Middle East and elsewhere, and this mechanism fills a needed gap for locating looting damage in a cost and time efficient manner, with potential global applications.</p>
</abstract>
<kwd-group>
<kwd>machine vision</kwd>
<kwd>archaeology</kwd>
<kwd>heritage</kwd>
<kwd>looting</kwd>
<kwd>automation</kwd>
<kwd>computational analysis</kwd>
<kwd>high resolution</kwd>
<kwd>Egypt</kwd>
</kwd-group>
<contract-num rid="cn02">OISE-1103878</contract-num>
<contract-sponsor id="cn01">National Geographic Society<named-content content-type="fundref-id">10.13039/100006363</named-content></contract-sponsor>
<contract-sponsor id="cn02">National Science Foundation<named-content content-type="fundref-id">10.13039/100000001</named-content></contract-sponsor>
<counts>
<fig-count count="6"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="55"/>
<page-count count="11"/>
<word-count count="7757"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="introduction">
<title>Introduction</title>
<p>Recent estimates indicate that looting at archaeological sites is a worldwide problem (Proulx, <xref ref-type="bibr" rid="B44">2013</xref>): 98% of 2,358 field archaeologists surveyed reported looting at or near their research sites. Recent and ongoing conflicts in Egypt (Ikram, <xref ref-type="bibr" rid="B26">2013</xref>; Ikram and Hanna, <xref ref-type="bibr" rid="B27">2013</xref>) as well as in Syria and Iraq (Pringle, <xref ref-type="bibr" rid="B43">2014</xref>; Gill, <xref ref-type="bibr" rid="B21">2016</xref>) create opportunities for organized looting for profit. While the exact numbers and amounts require more in-depth research (Chulov, <xref ref-type="bibr" rid="B14">2014</xref>; Felch, <xref ref-type="bibr" rid="B19">2014</xref>), it is clear that widespread damage to ancient sites and monuments is occurring as a result of looting and damage from war. In response, such looting and damage is monitored manually using space-based radar (Tapete et al., <xref ref-type="bibr" rid="B50">2016</xref>), repeat visits from commercial imaging satellites (Parcak, <xref ref-type="bibr" rid="B39">2007</xref>, <xref ref-type="bibr" rid="B41">2009</xref>; Hritz, <xref ref-type="bibr" rid="B25">2008</xref>; Stone, <xref ref-type="bibr" rid="B49">2008</xref>; AAAS, <xref ref-type="bibr" rid="B1">2014</xref>; Casana and Panahipour, <xref ref-type="bibr" rid="B8">2014</xref>; Newson and Young, <xref ref-type="bibr" rid="B36">2015</xref>), or from free public resources such as Google Earth (Thomas et al., <xref ref-type="bibr" rid="B51">2008</xref>; Contreras and Brodie, <xref ref-type="bibr" rid="B15">2013</xref>). However, manual analysis can be expensive, time intensive, difficult to replicate, and gappy. In addition source data can contain hundreds of megapixels. For example, a new looting mapping study used all the publicly available satellite imagery of Egypt to map looting from 2002 to 2013 across 1,200 &#x0201C;sitescapes&#x0201D; (with each site landscape containing one to hundreds of sites) (Parcak et al., <xref ref-type="bibr" rid="B40">2016</xref>). It took 6&#x02009;months of manual assessment to draw over 200,000 individual looting pits, with thousands of additional polygons around sites or features of interest. Even with high-resolution data at &#x0003C;1&#x02009;m/pixel resolution, expertise is required to discriminate between looting pits and modern holes or excavation units. Since looting is increasing across the globe, we need an algorithmic early warning system to focus human monitoring and intervention toward sites of interest in protection and preservation of ancient sites.</p>
<p>As others have done in illegal construction monitoring (Hofmann, <xref ref-type="bibr" rid="B23">2001</xref>; Hofmann et al., <xref ref-type="bibr" rid="B24">2008</xref>; Ioannidis et al., <xref ref-type="bibr" rid="B28">2009</xref>), we cast illegal looting pit detection as a change detection task, wherein the goal is to detect new looting pits and not preexisting pits or new instances of other categories. Approaches to satellite-based change detection often fall into one of two categories: those which classify pixels in the difference image and those which compare two images post-classification [for details, see Singh (<xref ref-type="bibr" rid="B47">1989</xref>), Mas (<xref ref-type="bibr" rid="B33">1999</xref>), Coppin et al. (<xref ref-type="bibr" rid="B16">2004</xref>), and Lu et al. (<xref ref-type="bibr" rid="B32">2004</xref>)]. Difference-image-based methods require very strict pixel alignment between images (e.g., Dai and Khorram, <xref ref-type="bibr" rid="B17">1998</xref>). This requirement is problematic when using very high-resolution imagery, multiple sensors, and differing photographic angles. Difference-image-based methods are commonly insensitive to variations in the types of change that occur (Bruzzone and Prieto, <xref ref-type="bibr" rid="B6">2000</xref>), making them susceptible to moving shadows, seasonal changes, and displaced objects such as sand dunes or re-parked trucks. Additionally, difference images are greatly impoverished by the removal of non-changing structures. This is fundamentally limiting; non-changing structures may still be good exemplars of the emergent targets and distractors, and modeling these structures may enhance the identification of targets.</p>
<p>Post-classification analysis exposes the entire scene of each image to processing and has become a popular approach. We take this approach to change detection here. Human-supervised classifiers, such as support vector machines (SVMs), have proven popular for the classification step, as well as for other topics in remote sensing. For example, Pal and Mather (<xref ref-type="bibr" rid="B38">2005</xref>) evaluated SVMs against maximum likelihood and artificial neural network classifiers trained to recognize land cover types in multispectral satellite data. They found that each classifier was successful, while SVMs were the superior choice. Mountrakis et al. (<xref ref-type="bibr" rid="B35">2011</xref>) describe a sharp increase in the application of these methods in recent years, particularly to ground cover classification using multispectral data. Ballabio and Sterlacchini (<xref ref-type="bibr" rid="B3">2012</xref>) go so far as to train an SVM classifier to predict regions susceptible to landslides in the future. Despite the popularity of supervised classifiers for change detection tasks on geographic information system (GIS) images, we will show that, for our particular task, our proposed partially supervised algorithm is superior: it produces better accuracy and false-positive rates while modeling unsupervised categories in the data, as well as localizing them.</p>
<p>Sophisticated unsupervised and partially supervised alternatives have recently gained traction. If each pixel or pixel neighborhood in the image is considered to be a data point, <italic>K</italic>-means clustering (Hartigan and Wong, <xref ref-type="bibr" rid="B22">1979</xref>; Rekik et al., <xref ref-type="bibr" rid="B45">2006</xref>) or a linear regression model (Galton, <xref ref-type="bibr" rid="B20">1894</xref>; Zhang et al., <xref ref-type="bibr" rid="B55">2011</xref>) can be used to separate out groups of changed and unchanged pixels, and principal component analysis can be effectively used to prepare the data (Celik, <xref ref-type="bibr" rid="B9">2009a</xref>). Scientists have used the kernel trick (Camps-Valls and Bruzzone, <xref ref-type="bibr" rid="B7">2009</xref>) to improve performance by creating non-linear classifiers. For example, one group clustered a preselected set of changed and unchanged pixels using kernel <italic>K</italic>-means, then assigned new pixels to categories using their one nearest neighbor (1NN) (Volpi et al., <xref ref-type="bibr" rid="B53">2012b</xref>). Others studies have used <italic>K</italic>-means on the undecimated discrete wavelet transform of the images, as well as a genetic algorithm to minimize the within-class errors of &#x0201C;changed&#x0201D; and &#x0201C;unchanged&#x0201D; assignments with equivalent effectiveness (Celik, <xref ref-type="bibr" rid="B10">2009b</xref>, <xref ref-type="bibr" rid="B11">2010</xref>). Each of these algorithms fails to build a generative model of the data and, therefore, is unlikely to build the detailed decision boundaries required for the classification of looting pits against a landscape of myriad structural distractors.</p>
<p>Partially supervised or unsupervised approaches can be improved with a variety of models. Bruzzone and Bovolo (<xref ref-type="bibr" rid="B5">2013</xref>) propose the manual creation of a tree-like hierarchical model of all changes expected to occur across images for the purpose of modeling both desired and undesired changes. Modeling the underlying data structure using unsupervised clustering often improves supervised classifier performance. For example, one group used <italic>K</italic>-means clusters to select data points for which human feedback might improve an SVM classifier (Liu et al., <xref ref-type="bibr" rid="B30">2008</xref>). More recently, Volpi et al. (<xref ref-type="bibr" rid="B52">2012a</xref>) used a similar approach to select data points for human feedback using hierarchical <italic>K</italic>-means and an adaptive heuristic. The search for looting pits among multiple instances of desert and urban targets using a minimal dataset is especially demanding of a model-based approach. However, modeling the entire set of observed objects requires a prohibitively large dataset. The software of eCognition (Hofmann, <xref ref-type="bibr" rid="B23">2001</xref>; Hofmann et al., <xref ref-type="bibr" rid="B24">2008</xref>) is capable of simultaneously modeling color histogram objects at multiple scales and segmenting the image to localize them. Localization aids in unsupervised object discovery by removing irrelevant pixels from each category prototype, performing feature selection. However, because eCognition&#x02019;s approach to this task ignores structural information and is fundamentally unsupervised, it cannot be guided to identify relatively few looting pit structures among numerous instances of so many distractor classes. Ioannidis et al. (<xref ref-type="bibr" rid="B28">2009</xref>) describe an algorithm to identify new buildings in a rural environment. The authors identified changed buildings using a knowledge-based approach on stereo images to categorize &#x0201C;ground&#x0201D; vs &#x0201C;above-ground,&#x0201D; but only after they used a number of tricks to eliminate false positives (FPs) by increasing the ratio of target and non-target pixels. In a common pattern, these algorithms either attempt to model all observed classes or only the classes of interest.</p>
<p>We propose a novel categorization scheme and its use as the central component of a post-classification analysis technique for change detection. Our algorithm is derived from hierarchical clustering performed by the brain for use in the identification of repeated structurally consistent small-scale targets among many repeated distractors. Our approach improves the identification of a specific category of observation by modeling numerous categories. It avoids modeling all potential object categories by only modeling those relevant to the distinctiveness of the target. It localizes instances of each observational category in order to ignore uninformative pixels. It accomplishes this using only the positions of members of the relatively rare target category as supervised labels and avoids the need for costly human intervention for parameter tweaking or resolution of high false-positive rates. Finally, it may exemplify the first algorithm for satellite-based identification of structurally consistent ground feature changes for archaeology.</p>
<p>In the following sections, we introduce our algorithm based on recently developed image analysis methods for recognition and localization (Chandrashekar and Granger, <xref ref-type="bibr" rid="B12">2012</xref>; Chandrashekar et al., <xref ref-type="bibr" rid="B13">2014</xref>). We apply our algorithm to a large (239 million pixel) satellite image of the pyramid field zone in Egypt, where organized looting of burial sites is ongoing (Figure <xref ref-type="fig" rid="F1">1</xref>). We show cross-validated success and false-alarm rates for looted pits and compare these directly against SVM on bags of visual words, a standard method for classification in satellite imagery and other fields.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>Satellite image from Saqqara desert, Egypt</bold>. Mean ground sample distance 0.71&#x02009;m. <bold>(A)</bold> Full image (10,148&#x02009;&#x000D7;&#x02009;23,561 pixels) of border between unsettled desert and populated regions (see text). <bold>(B)</bold> Close-up of region indicated by outline in panel <bold>(A)</bold>, in 2011, in which evidence of looting can be seen (small dark spots in circled regions). <bold>(C)</bold> Close-up of several looting pits. All relevant permissions have been obtained for figure publication.</p></caption>
<graphic xlink:href="fict-04-00004-g001.tif"/>
</fig>
</sec>
<sec id="S2">
<title>Algorithm Description</title>
<p>The new hierarchical categorization and localization (HCAL) algorithm introduced here is a method for partially supervised classification and localization, derived from previous work on partially supervised classification alone (Chandrashekar and Granger, <xref ref-type="bibr" rid="B12">2012</xref>) and on unsupervised localization alone (Chandrashekar et al., <xref ref-type="bibr" rid="B13">2014</xref>). The resulting algorithm clusters images by similarity while simultaneously identifying image regions likely to contain examples of these clusters. It then successively narrows category descriptors for these images until they identify specific features that constitute target images of disrupted archaeological burial sites or usefully discriminable non-target objects.</p>
<p>Intuitively, the HCAL algorithm first identifies repeated featural motifs occurring in the set of images using unsupervised methods, thus localizing candidate feature sets. It then (again without supervision) categorizes images based on similarity. Finally, the members of these similarity-based categories are checked for mismatching (supervised) labels&#x02014;the first time supervised information is referenced. When mismatched labels are found (such as a category containing both looting pits and other objects), the algorithm hierarchically splits categories again in a purely unsupervised manner. This method is repeated until clusters contain purely pit images, purely non-pit images, or very few (e.g., 7) images. The resultant data structure contains hierarchical categories of image regions, with leaves of the hierarchical tree corresponding to subcategories each sharing a common supervised label. The use of unsupervised splits causes the HCAL algorithm to learn the structural model of the data, enabling more intelligent decision boundaries than supervised learning alone. Figure <xref ref-type="fig" rid="F2">2</xref> provides a flowchart and pseudocode for the algorithm.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p><bold>Hierarchical categorization and localization (HCAL) flowchart and pseudocode</bold>. <inline-formula><mml:math id="M1"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DisruptedSites</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>Archived</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> contains a list of coordinates for known disrupted burial sites. The joint localization and classification algorithm (JLC) receives a document map (<inline-formula><mml:math id="M2"><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>), a set of subphotos from preselected training satellite images. JLC returns <inline-formula><mml:math id="M3"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>Loc</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>, the location of those labeled items in global coordinates, and <inline-formula><mml:math id="M4"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>, the histogram of foreground pixel values. CSLTrain returns C<sub>Forest</sub>, the data structure representing the hierarchical clustering solution. HCALTest receives C<sub>Forest</sub> as a representation of what was learned, and <inline-formula><mml:math id="M5"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>New</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>, a set of tiled subphotos from new &#x0201C;test&#x0201D; satellite observations&#x02014;the data to analyze. When HCALTest completes, it has generated <inline-formula><mml:math id="M6"><mml:mrow><mml:msubsup><mml:mrow><mml:mover accent='true'><mml:mtext>L</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow><mml:mrow><mml:mtext>Pr&#x02009;ed</mml:mtext></mml:mrow></mml:msubsup></mml:mrow></mml:math></inline-formula>, a list of labels for each subphoto.</p></caption>
<graphic xlink:href="fict-04-00004-g002.tif"/>
</fig>
<sec id="S2-1">
<title>Training Steps</title>
<p>The four training steps of the HCAL algorithm can be summarized as follows:
<list list-type="order">
<list-item><p><inline-formula><mml:math id="M7"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>Archived</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x02190;</mml:mo><mml:mtext>FeatureExtr</mml:mtext><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>GISImages</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>Archived</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></inline-formula> <inline-formula><mml:math id="M8"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>New</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x02190;</mml:mo><mml:mtext>FeatureExtr</mml:mtext><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>GISImages</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>New</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></inline-formula> GIS data from a site of interest are split into numerous small images. For each image, we extracted features from each image using a process similar to Chandrashekar et al. (<xref ref-type="bibr" rid="B13">2014</xref>). Dense scale-invariant feature transform (SIFT) (Lowe, <xref ref-type="bibr" rid="B31">2004</xref>), histogram of oriented gradients (HOG) (Dalal and Triggs, <xref ref-type="bibr" rid="B18">2005</xref>), or speeded up robust features (SURF) (Bay et al., <xref ref-type="bibr" rid="B4">2008</xref>) created a high-dimensional representation of each pixel. Clustering each image using <italic>K</italic>-means produced a 40 cluster solution in pixel feature space. Finally, we clustered the set of all means (40 per image) with <italic>K</italic>-means to produce a 40 cluster vocabulary of visual words in pixel feature space for the entire training dataset and renumbered the pixels for each image with its closest cluster as identified via nearest neighbor. Each image can be thus viewed as a document of visual words (Chandrashekar et al., <xref ref-type="bibr" rid="B13">2014</xref>). A document map (<inline-formula><mml:math id="M9"><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>) was then created wherein each pixel was labeled based on the vocabulary word representing its feature. Based on the comparison of overall algorithmic performance on several different mean sizes, we determined our choice of 40 means analytically, and may be increased for application to more complex urban datasets.</p></list-item>
<list-item><p><inline-formula><mml:math id="M10"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>Loc</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x02190;</mml:mo><mml:mtext>&#x000A0;JLC(</mml:mtext><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover><mml:mtext>)</mml:mtext></mml:mrow></mml:math></inline-formula> The (unsupervised) joint localization and classification algorithm (JLC) is performed on the set of archived and newly processed images (<inline-formula><mml:math id="M11"><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>), producing three data structures: (i) <inline-formula><mml:math id="M12"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>Loc</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>, which is the location of an image&#x02019;s region that is believed to contain the &#x0201C;foreground&#x0201D; object (expected to recur in the images) and (ii) the histogram of features <inline-formula><mml:math id="M13"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> occurring within that <inline-formula><mml:math id="M14"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>Loc</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> region (a bag of visual words). This expectation&#x02013;maximization algorithm finds <italic>K</italic>-means cluster centroids in the feature histograms <inline-formula><mml:math id="M15"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> for each image, then redefines the foreground of each image <inline-formula><mml:math id="M16"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>Loc</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> to best match its cluster&#x02019;s centroid. Each image&#x02019;s histogram <inline-formula><mml:math id="M17"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> is redefined using only foreground pixels, and the process repeats; see Chandrashekar et al. (<xref ref-type="bibr" rid="B13">2014</xref>).</p></list-item>
<list-item><p><inline-formula><mml:math id="M18"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>L</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x02190;</mml:mo><mml:mtext>&#x000A0;CalculateTrainLabels(</mml:mtext><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>Loc</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DisruptedSites</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>Archived</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></inline-formula> Foreground locations identified by the JLC subroutine are compared with a list of known locations <inline-formula><mml:math id="M19"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DisruptedSites</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>Archived</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> for disrupted burial sites (DBSs) and labeled based on whether one or more looted sites are present in the foregrounds.</p></list-item>
<list-item><p><inline-formula><mml:math id="M20"><mml:mrow><mml:msub><mml:mtext>C</mml:mtext><mml:mrow><mml:mtext>Forest</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x02190;</mml:mo><mml:mtext>&#x000A0;CSLTrain(</mml:mtext><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>L</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow><mml:mrow><mml:mtext>Train</mml:mtext></mml:mrow></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></inline-formula> The semi-supervised CSL classification algorithm (Chandrashekar and Granger, <xref ref-type="bibr" rid="B12">2012</xref>) is applied to the set of pairs of image-foreground histograms <inline-formula><mml:math id="M21"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> and supervised labels <inline-formula><mml:math id="M22"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>L</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> (DBS/non-DBS) corresponding to each such histogram. The CSL algorithm generates a tree (or forest of trees) in which each node performs unsupervised <italic>K</italic>-means clustering. For each cluster in the root node containing members of more than one supervised label, a child node is created. This child node is passed as data only the members of that cluster. The process recurses until leaf nodes each contain members of only one category label. Refer to Figure <xref ref-type="fig" rid="F3">3</xref> for pseudocode. Since the maximum number of child nodes that can be created at any node is equal to the number of clusters, the branching factor (BF) of this tree is equal to <italic>K</italic> (the number of means) in <italic>K</italic>-means. The time complexity of this training algorithm is O(<italic>NK</italic>), where <italic>N</italic> is the number of images and <italic>K</italic> is the number of means. Space complexity is the same (Chandrashekar and Granger, <xref ref-type="bibr" rid="B12">2012</xref>).</p></list-item>
</list></p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p><bold>Pseudocode for the CSLTrain component of the algorithm, derived with permission from Algorithm 1 of Chandrashekar and Granger (<xref ref-type="bibr" rid="B12">2012</xref>)</bold>.</p></caption>
<graphic xlink:href="fict-04-00004-g003.tif"/>
</fig>
</sec>
<sec id="S2-2">
<title>Testing Step</title>
<p>The HCAL procedure to evaluate new data can be summarized as follows:
<list list-type="order">
<list-item><p><inline-formula><mml:math id="M23"><mml:mrow><mml:msubsup><mml:mover accent='true'><mml:mtext>L</mml:mtext><mml:mo>&#x021C0;</mml:mo></mml:mover><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow><mml:mrow><mml:mtext>Pred</mml:mtext></mml:mrow></mml:msubsup><mml:mo>&#x02190;</mml:mo><mml:mtext>CSLTest</mml:mtext><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mtext>C</mml:mtext><mml:mrow><mml:mtext>Forest</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo>&#x021C0;</mml:mo></mml:mover><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow><mml:mrow><mml:mtext>Test</mml:mtext></mml:mrow></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></inline-formula> For each run, we tested the CSL algorithm on hold-out images not originally part of the training set; what is returned is the set of predicted labels <inline-formula><mml:math id="M24"><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mrow><mml:mover accent='true'><mml:mtext>L</mml:mtext><mml:mo>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow><mml:mrow><mml:mtext>Pred</mml:mtext></mml:mrow></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></inline-formula> that correctly and incorrectly classified as DBS. In the root node, each data point is assigned to its nearest cluster and is recursively passed to the corresponding child node. This repeats until the data point enters a leaf node, where it is assigned a label based on the label of training points assigned that leaf node during training. Since this approach amounts to a traversal from the root node to a leaf, in a full tree the time complexity is log-linear. Using a simple voting scheme, multiple trees can be combined in a random forest. For this experiment, we used repeated random subsampling to generate validation/test sets (see &#x0201C;<xref ref-type="sec" rid="S4">Materials and Methods</xref>&#x0201D;).</p></list-item>
</list></p>
</sec>
<sec id="S2-3">
<title>Contrasts with Prior Art</title>
<p>Many categorization techniques customized to change detection in very high-resolution satellite imagery are not directly applicable to the task at hand. Looting pits are defined by their structure, so pixel-based, texture-based, or multispectral methods will not work. Distractors are numerous, and both targets and distractors are varied, so fully unsupervised methods are likely to fail. These limitations leave few options in the prior art fit for comparison. However, our method can be contrasted with supervised methods such as the SVM, one of the most widely used statistical classification methods for remote imaging and other data, in two key ways: (i) SVMs classify entire input images, without localizing those features within the image which are most responsible for the classification and (ii) whereas SVMs laboriously (and computationally expensively) construct a &#x0201C;separator surface&#x0201D; containing all of one category on one side of the separator (in high-dimensional input feature space) and the other category on the other side, the HCAL algorithm simply identifies regional volumes in the space containing similar looking data, and iteratively sub-partitions those volumes whenever they contain mismatching labels. More fine-grained splits are performed only where necessary. The HCAL&#x02019;s CSLTrain mechanism is estimated to cost roughly an order of magnitude less in terms of time complexity than SVMs [for evidence, see Chandrashekar and Granger (<xref ref-type="bibr" rid="B12">2012</xref>)]. Here, we will compare against an SVM with linear, quadratic, and cubic polynomial kernels (see &#x0201C;<xref ref-type="sec" rid="S4">Materials and Methods</xref>&#x0201D;), and against Na&#x000EF;ve Bayes and linear discriminant analysis (LDA) classifiers.</p>
</sec>
</sec>
<sec id="S3">
<title>Data</title>
<p>The primary dataset used in the current study included a set of satellite images from the pyramid fields region of Egypt, captured by the EROS-B1 satellite at a panchromatic mean ground sample distance (spatial resolution) of 0.71&#x02009;m per pixel. A large (3,200&#x02009;&#x000D7;&#x02009;4,300 pixels) region was tested. This region contains multiple instances of looted burial sites, substantial expanses of open desert, and multiple distractors such as farmland, buildings, modern (non-archaeological) graveyards, and military bases with bomb craters. Like many similar approaches to the monitoring of structural objects (e.g., Wei et al., <xref ref-type="bibr" rid="B54">2004</xref>; Sirmacek and Unsalan, <xref ref-type="bibr" rid="B48">2009</xref>), we restrict analysis to grayscale images to demonstrate broad applicability across high-spatial-resolution commercial satellites many of which are panchromatic in high resolution (e.g., GeoEye-1, WorldView-2 and 3, and Ikonos). Luminance was normalized across the entire dataset to improve dynamic range.</p>
</sec>
<sec id="S4" sec-type="materials|methods">
<title>Materials and Methods</title>
<p>Our team split the very high-resolution satellite data into tiled images of dimensions 30&#x02009;&#x000D7;&#x02009;30 pixels with an overlap of 10 pixels to ensure that features on the scale of looting pits (generally smaller than 10 pixels in diameter) could be intact in at least one image. From these images, 300 looting sites and 2,000 other locations were randomly sampled for use (visualized in Figure S1 in Supplementary Material). This unbalanced ratio better mimics real world conditions, wherein looting pits cover only a small fraction of the land mass. To better take advantage of the limited number of training examples, we reduced the number of training images without looting pits to equal the number of training images with pits for the supervised components of the algorithm (CSLTrain). Human experts identified initial sets of DBSs for training and validation manually, with crosschecking by multiple trained participants.</p>
<p>We ran JLC with an <italic>a priori</italic> expectation of 32 clusters of images being present in the dataset. JLC-utilizing bounding boxes incorporated a tree-based pyramid search for the optimal rectangle. Using superpixel-based localization (sample localizations in Figure <xref ref-type="fig" rid="F4">4</xref>; results in Figure S2 in Supplementary Material), JLC identified a contiguous combination of superpixels, which together produced the most likely feature histogram most likely to belong to a cluster. Superpixels resulted from an over-segmentation of the image; most objects are assumed to be comprised of numerous adjacent superpixels. Readers may refer to Chandrashekar et al. (<xref ref-type="bibr" rid="B13">2014</xref>) for more details. While the degree of segmentation in this study was hand-tuned to over-segment desert scenes, interesting future work would be to apply an adaptive segmentation technique, such as fuzzy binarization applied in Santosh et al. (<xref ref-type="bibr" rid="B46">2016</xref>) to automatically identify the optimal degree of over-segmentation based on weather conditions.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p><bold>Localization results for three example satellite image patches</bold>. Top row: superpixel-localized regions. Middle row: bounding box-localized regions. Bottom row: heat maps of the same images wherein brighter colors indicate the location of features more indicative of the subimage&#x02019;s similarity cluster. Red circles mark human-localized looting pits. Examples <bold>(A)</bold> and <bold>(B)</bold> contain looting pits, while example <bold>(C)</bold> contains structural walls, with no graves or looting pits present.</p></caption>
<graphic xlink:href="fict-04-00004-g004.tif"/>
</fig>
<p>CSLTrain was implemented with a forest of 100 CSL trees wherein each node of each tree attended to a randomly selected 20% of the 40 available feature histogram dimensions. We chose branching factors of 2 and 6 from an exploration of the relative accuracies of many branching factors. In this case, accuracy peaks just above a branching factor of 2, and falls off if the branching factor is increased further. Practically, many branching factors performed acceptably. Leaf nodes claiming less than seven data points did not include splitting and, instead, had labeling as looting pits based on whether the majority of their data points contained one or more pits. We recombined the final results for CSLTest using a simple voting scheme. Our team performed each validation trial by randomly selecting 2,300 unique images (300 with looting and 2,000 without) with replacement between trials. Within each trial, we selected unique random initializations for JLC and CSLTrain, each of which functions <italic>via</italic> error reduction from an initial state. Results for each trial are the mean across 3 different random initializations, each of which is scored as the mean results of CSLTrain across 10 bootstrapped samples. Two algorithms are compared; CSL on the outputs of JLC and SVM (linear, quadratic, and cubic kernels) on bags of visual words representing the original images. In Figure S3 in Supplementary Material, we also compare CSL on the outputs of JLC against SVM (linear, quadratic, and cubic kernels), LDA, and Na&#x000EF;ve Bayes on the outputs of JLC in place of CSLTrain and CSLTest.</p>
</sec>
<sec id="S5">
<title>Results</title>
<p>We trained HCAL and comparison algorithms on random subsamplings of the data for cross-validation (see &#x0201C;<xref ref-type="sec" rid="S4">Materials and Methods</xref>&#x0201D;). Results of identification and localization are shown for a Dashur satellite image (Figure S1 in Supplementary Material). The HCAL algorithm can operate in either bounding box localization mode or superpixel localization mode. Localization of looting pits within images appeared reasonable in both localization modes. It is important to re-note that HCAL always localizes the most likely object in the image given the other images, regardless of its supervised label.</p>
<p>Figure <xref ref-type="fig" rid="F5">5</xref> illustrates the statistical findings using SIFT features and bounding boxes for localization. Figure <xref ref-type="fig" rid="F5">5</xref>A shows the accuracy of predictions made by the HCAL method introduced here and by SVM trained on a simple bag of visual words run on the original images using the same visual vocabulary. In the SVM case, linear, quadratic, and cubic kernels were evaluated (see &#x0201C;<xref ref-type="sec" rid="S4">Materials and Methods</xref>&#x0201D;). The HCAL method outperforms the SVMs. Shown in Figure <xref ref-type="fig" rid="F5">5</xref>B are accuracy measures given a varying percentage of the data used for training (from 20 to 90%). It can be seen that the HCAL method reliably outperforms linear SVMs by 2&#x02013;4 accuracy percentage points for all sizes of training data used. A receiver operating characteristic plot (Figure <xref ref-type="fig" rid="F6">6</xref>) shows that HCAL results (again, SIFT features and bounding box localization) are competitive with the results of linear SVMs not only in terms of accuracy of prediction but also in terms of low false-positive rates (14.62&#x02009;&#x000B1;&#x02009;0.18% for HCAL with a branching factor of 2; 14.96&#x02009;&#x000B1;&#x02009;0.17% for HCAL with a branching factor of 6; 17.75&#x02009;&#x000B1;&#x02009;0.06% for linear SVM). Using SIFT features, HCAL with a branching factor of 2 scores an accuracy of 85.33&#x02009;&#x000B1;&#x02009;0.16 against linear SVM&#x02019;s 82.11&#x02009;&#x000B1;&#x02009;0.06. Results are similar using SURF features. Accuracy is reduced across the board using the less descriptive but more concise HOG features, although HCAL still outperforms SVM in this case (Figure S2A in Supplementary Material). We also tested superpixel-based localization, which generally performed more poorly; see Figure S2B in Supplementary Material for details. Examples of typical localization results are available in Figure <xref ref-type="fig" rid="F4">4</xref>. Resultant tree structures are available in Figure S4 in Supplementary Material. See Figure S6 in Supplementary Material for a large number of localization examples for the bounding box localized SIFT feature condition.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p><bold>(A)</bold> Predictive accuracy using a bounding box localizer with scale-invariant feature transform (SIFT) features for the frontend, <italic>N</italic>&#x02009;&#x0003D;&#x02009;100 per SE bar; <bold>(B)</bold> predictive accuracy for support vector machine (SVM) and hierarchical categorization and localization (HCAL) as a function of number of training examples, using identical parameters (<italic>N</italic>&#x02009;&#x0003D;&#x02009;50 per SE bar).</p></caption>
<graphic xlink:href="fict-04-00004-g005.tif"/>
</fig>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p><bold>Disrupted burial site (DBS) correct identification rate vs false-alarm rate</bold>. Results generated using scale-invariant feature transform features with bounding box localization from one complete trial (tested using 3 random initializations, 10 bootstrapped train/test splits each). Correct identification rate, also known as sensitivity or true positive (TP) rate, is measured as the number of positive &#x0201C;DBS is present&#x0201D; responses from the algorithm divided by the total number of images of DBSs. False-alarm rate, or false positive (FP) rate, equals the number of negative &#x0201C;no DBS present&#x0201D; responses from the algorithm divided by the total number of images without DBSs. A position in the top left of a receiver operating characteristic (ROC) plot such as this is desirable because it indicates that all DBSs were located, while no distractors were mistaken for DBSs. For comparison, co-displayed are 30 runs on the same images for SVM (linear kernel) and the SVM receiver operating characteristic curve (calculated using 10-fold cross-validation).</p></caption>
<graphic xlink:href="fict-04-00004-g006.tif"/>
</fig>
</sec>
<sec id="S6" sec-type="discussion">
<title>Discussion</title>
<p>The HCAL algorithm outperforms SVM regardless of branching factor used. This is a reflection of the importance of localizing repeated features, further enhanced by HCAL&#x02019;s ability to model the hierarchical substructure of the data in order to create more intelligent decision boundaries. Meanwhile, the algorithm can run on a commodity workstation, making it affordable for wide-scale implementation.</p>
<p>As our data show both SVM and HCAL perform better on SIFT- and SURF-transformed images than on HOG. This was found to be true anecdotally for several other feature transforms as well. Two factors could underlie this difference. First, subpar localizations during the JLC stage could forestall class separability during CSLTrain by excluding regions or including distractor pixels. However, this would not explain the reduced performance in SVM results, where localization was not performed. Alternately, the DMap histograms (<inline-formula><mml:math id="M25"><mml:mrow><mml:msub><mml:mrow><mml:mover accent='true'><mml:mtext>H</mml:mtext><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>FG</mml:mtext></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula>) generated from those localizations could be less informative due to the impoverished feature information encoded in HOG descriptors. This explanation seems more likely.</p>
<p>One of the primary driving factors of algorithm selection for this problem domain is the ability to function on a minimal set of supervised labels. HCAL accomplishes this by modeling the data generatively and not requiring labels for the various and plentiful non-target classes. Multiple subclasses of a target class can be modeled separately, and each can differentiate from each non-target class with unique decision boundaries. These boundaries are not limited to linear functions due to the hierarchical nature of the CSL step. This same feature also reduces overfitting. With each hierarchical split of the data during CSLTrain, a very simple classifier is created (<italic>K</italic>-means and 1NN). Since purely disrupted burial site (DBS) and purely non-DBS clusters are never split again, the model avoids unnecessary complexity.</p>
<p>As can be seen in Section &#x0201C;<xref ref-type="sec" rid="S5">Results</xref>,&#x0201D; HCAL remained superior to SVM as the training dataset shrunk. In the 20% training case, HCAL was able to build a forest of differentiating trees superior in accuracy to SVM using as few as 32 example images. The false-alarm rate for the HCAL algorithm is quite low, as is the FP rate for SVM utilizing the JLC prestep. This is in contrast to competing algorithms for this problem domain, which often encounter many FPs, an understandable issue given the ratio of target to non-target images. Although unnecessary, additional reductions to the false-alarm rate could be achieved by training the same algorithm on some of the most common non-target categories and removing instances of those categories from the dataset. The low false-positive rate of the HCAL method is of particular interest when attempting to minimize human intervention in the automated monitoring of archaeological sites over extended time periods.</p>
<p>We performed an additional evaluation (Figure S3 in Supplementary Material) of predictive accuracy using SVM (linear, quadratic, and cubic kernels), LDA, and Na&#x000EF;ve Bayes in place of the CSL step. Direct comparison of these algorithms with the CSL step is of minor interest, since only CSL learns the hierarchical generative structure of the data, enabling it to model multiple DBS and distractor appearance classes. Results indicate that, while CSL is often the top performer, accuracy is not as strongly tied to this choice&#x02014;what matters most is the JLC localization phase.</p>
<p>To our knowledge, the localizations generated by HCAL are unique in this problem domain. All existing segmentation approaches identify extents of an image based not on structural information but based on edges, keypoints, texture, or spectral content (e.g., Kass et al., <xref ref-type="bibr" rid="B29">1988</xref>; Agouris et al., <xref ref-type="bibr" rid="B2">2001</xref>; Hofmann, <xref ref-type="bibr" rid="B23">2001</xref>; Wei et al., <xref ref-type="bibr" rid="B54">2004</xref>; Mayunga et al., <xref ref-type="bibr" rid="B34">2005</xref>; Peng et al., <xref ref-type="bibr" rid="B42">2005</xref>; Hofmann et al., <xref ref-type="bibr" rid="B24">2008</xref>; Opitz and Blundell, <xref ref-type="bibr" rid="B37">2008</xref>; Sirmacek and Unsalan, <xref ref-type="bibr" rid="B48">2009</xref>). By contrast, HCAL operates on remote sensing data by searching for structural configurations (as described in Section &#x0201C;<xref ref-type="sec" rid="S2">Algorithm Description</xref>&#x0201D;) that have been identified. Of equal interest is the subsequent ability of the method, once it has found a structural object, to then successively identify more fine-grained localizations (see &#x0201C;<xref ref-type="sec" rid="S2-1">Training Steps</xref>&#x0201D; and &#x0201C;<xref ref-type="sec" rid="S2-2">Testing Step</xref>&#x0201D;), greatly facilitating identification of regions of experimenter interest.</p>
<p>Under certain conditions, HCAL can miss looting pits. In images where multiple instances of the same object are present, the algorithm will often localize a subset of the instances. This is likely due to the algorithm&#x02019;s requirement that localizations take the form of a contiguous superpixel or box. Adding, to the histogram, the desert sand between DBSs would be detrimental to forming a good match, but averaging another DBS into an existing DBS histogram changes it little. In practice, the task of identifying looted regions with ongoing additional looting can be accomplished by masking previously identified looting pits.</p>
<p>A variant of the algorithm using superpixel-based localizations rather than bounding boxes was also analyzed (Figure S2B in Supplementary Material). Under such a scheme, the HCAL algorithm still performed reasonably well but found difficulty surpassing SVMs in the SIFT feature case. Superpixel-based localizations can form complex concave regions. These regions can wrap around a disrupted burial sites, capturing much of the pixels indicative of disruption, without covering the human-marked location of disruption. This could lead to DBS-negative exemplars containing DBS features. This problem is perhaps specific to a point-based supervisory labeling system, so further work is required to ascertain the impacts of localization boundaries on HCAL. The superpixel segmentation itself, performed as preprocessing step prior to all other tasks, could easily be supplemented to expand superpixel boundaries into convex regions, helping to guide superpixels toward selection of the entire looting pit.</p>
<p>Minimal parameter tuning is required to reach near-optimal results with HCAL, in contrast with existing methods. The most important parameters in this algorithm are the vocabulary size from which to generate <inline-formula><mml:math id="M26"><mml:mrow><mml:mover accent='true'><mml:mrow><mml:mtext>DMap</mml:mtext></mml:mrow><mml:mo stretchy='true'>&#x021C0;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>, the number of generative classes to expect when localizing, and the branching factor of the CSL step. We chose a vocabulary size of 40, since it is slightly superior to other explored sizes such as 20 or 80. Future work might find improved performance by fine-tuning this parameter. We explored the use of 16, 32, or 64 generative classes for localization. The difference in accuracy was minimal, so we report results for 32 classes here. Branching factor makes a substantial difference, and numbers for two branching factors, 2 and 6, are reported side-by-side. Figure S5 in Supplementary Material supplies accuracies for each of the mentioned parameter choices.</p>
<p>Presented here is a complete solution for target category identification in remote sensing imagery. However, the broader goal is to enable fast response to archaeological looting by academic institutions across the globe. The authors propose a system by which satellite images of each site of interest are downloaded at regular intervals (weekly or monthly depending on the risk of looting to the site and ongoing damage assessment), and each image is analyzed for looting pits. Each new image is compared with the category labels (DBS/non-DBS) of previous views of the same coordinates, and new DBSs are forwarded to humans for confirmation and action. Confirmed looting pits might be added to a supervised dataset for future retraining of the algorithm. This solution also has applications across entire countries where looting patterns are similar. While our error rates might be greater due to an increased number of land classifications units, our process would pick out looting hotspots for individual site inspection. This would effectively cut out the most time-consuming process of looting mapping for an entire country: scanning each site individually and over time. Our proposed solution would also pick out looting hotspots at unknown sites (i.e., not in national archaeological site databases or GISs), facilitating both new site registration and site protection.</p>
</sec>
<sec id="S7">
<title>Conclusion</title>
<p>The novel HCAL algorithm presented here is an ideal candidate for VHD satellite image analysis when the goal is to identify numerous instances of target categories among many distractors using minimal human intervention and training data. We first acquire numerous grayscale VHD satellite images from an archaeological site of interest, some of which contain looting pits and many which do not. These images are grouped by structural content using unsupervised clustering while simultaneously localizing the regions in each image which contain structural content associated with the image&#x02019;s cluster. We then train a hierarchical model of the data consisting of multiple categories, some of which are labeled as containing pits based on human labeling. Numerous future images can be labeled as containing looting pits based upon the same system, and results can be compared with previous data. Changes in labels (not at the pixel level, but at the image level) can be flagged for human confirmation and intervention. These new images can be combined with archival images to increase the power of the model for future predictions. We have shown that this system works well throughout multiple tested subregions of an archaeological site in Egypt, producing advantageous false-positive rates despite the large number of distractors. HCAL is the first algorithm of its type to be applied in the defense of archaeological sites from human intervention. It has minimal parameters, making it easy to apply to new regions. We speculate that it will have high value as a tool across a broad range of unforested sites at risk for looting. While the methodology presented here does not have an accuracy of 100%, it could be applied across satellite datasets of an entire country, finding the majority of total sites affected by looting (both known sites and previously unknown sites). Finding the actual sites affected by looting is the most time consuming process of a countrywide analysis. We simply do not have a good sense of the total numbers of sites affected by looting in the Middle East, nor do we have an efficient way of monitoring countrywide satellite data to alert authorities. This automated approach is at present the best tool for rapid analysis of large-scale site looting.</p>
</sec>
<sec id="S8" sec-type="author-contributor">
<title>Author Contributions</title>
<p>Substantial contributions to conception or design of the work; drafting the work or revising it critically for important intellectual content; final approval of the version to be published; and agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved: RG, EB, BT, and SP.</p>
</sec>
<sec id="S9">
<title>Conflict of Interest Statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<sec id="S10">
<title>Funding</title>
<p>This project was funded by the National Geographic Society, the National Science Foundation (OISE-1103878), D. Lehr and the Antiquities Coalition, and the University of Alabama at Birmingham, and supported in part by grant N00014-15-1-2132 from the Office of Naval Research and grant N000140-15-1-2823 from the Defense Advanced Research Projects Agency. SP would like to thank the Ministry of Antiquities in Egypt for their support, including M. Damaty, H. el-Azm, K. Waheed, M. Ali, and A. Ahmed.</p>
</sec>
<sec id="S11" sec-type="supplementary-material">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at <uri xlink:href="http://journal.frontiersin.org/article/10.3389/fict.2017.00004/full&#x00023;supplementary-material">http://journal.frontiersin.org/article/10.3389/fict.2017.00004/full&#x00023;supplementary-material</uri>.</p>
<supplementary-material xlink:href="Data_Sheet_1.PDF" id="SM1" mimetype="applicationn/PDF" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<sec id="S12">
<title>Abbreviations</title>
<p>BF, branching factor; DBS, disrupted burial site; HCAL, hierarchical categorization and localization (this algorithm); 1NN, one nearest neighbor classification; GIS, geographic information system; PCA, principal component analysis; ROC, receiver operating characteristic; SVM, support vector machine.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="web"><collab>AAAS</collab>. (<year>2014</year>). <source>Ancient History, Modern Destruction: Assessing the Current Status of Syria&#x02019;s World Heritage Sites Using High-Resolution Satellite Imagery</source>. Available at: <uri xlink:href="http://www.aaas.org/page/ancient-history-modern-destruction-assessing-current-status-syria%E2%80%99s-world-heritage-sites-using">http://www.aaas.org/page/ancient-history-modern-destruction-assessing-current-status-syria%E2%80%99s-world-heritage-sites-using</uri></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Agouris</surname> <given-names>P.</given-names></name> <name><surname>Stefanidis</surname> <given-names>A.</given-names></name> <name><surname>Gyftakis</surname> <given-names>S.</given-names></name></person-group> (<year>2001</year>). <article-title>Differential snakes for change detection in road segments</article-title>. <source>Photogramm. Eng. Remote Sensing</source> <volume>67</volume>, <fpage>1391</fpage>&#x02013;<lpage>1400</lpage>.</citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ballabio</surname> <given-names>C.</given-names></name> <name><surname>Sterlacchini</surname> <given-names>S.</given-names></name></person-group> (<year>2012</year>). <article-title>Support vector machines for landslide susceptibility mapping: the Staffora River Basin case study, Italy</article-title>. <source>Math. Geosci.</source> <volume>44</volume>, <fpage>47</fpage>&#x02013;<lpage>70</lpage>.<pub-id pub-id-type="doi">10.1007/s11004-011-9379-9</pub-id></citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bay</surname> <given-names>H.</given-names></name> <name><surname>Ess</surname> <given-names>A.</given-names></name> <name><surname>Tuytelaars</surname> <given-names>T.</given-names></name> <name><surname>van Gool</surname> <given-names>L.</given-names></name></person-group> (<year>2008</year>). <article-title>Speeded-up robust features (SURF)</article-title>. <source>Comput. Vision Image Understand.</source> <volume>110</volume>, <fpage>346</fpage>&#x02013;<lpage>359</lpage>.<pub-id pub-id-type="doi">10.1016/j.cviu.2007.09.014</pub-id></citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bruzzone</surname> <given-names>L.</given-names></name> <name><surname>Bovolo</surname> <given-names>F.</given-names></name></person-group> (<year>2013</year>). <article-title>A novel framework for the design of change-detection systems for very-high-resolution remote sensing images</article-title>. <source>Proc. IEEE</source> <volume>101</volume>, <fpage>609</fpage>&#x02013;<lpage>630</lpage>.<pub-id pub-id-type="doi">10.1109/JPROC.2012.2197169</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bruzzone</surname> <given-names>L.</given-names></name> <name><surname>Prieto</surname> <given-names>D. F.</given-names></name></person-group> (<year>2000</year>). <article-title>Automatic analysis of the difference image for unsupervised change detection</article-title>. <source>IEEE Trans. Geosci. Remote Sensing</source> <volume>38</volume>, <fpage>1171</fpage>&#x02013;<lpage>1182</lpage>.<pub-id pub-id-type="doi">10.1109/36.843009</pub-id></citation></ref>
<ref id="B7"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Camps-Valls</surname> <given-names>G.</given-names></name> <name><surname>Bruzzone</surname> <given-names>L.</given-names></name></person-group> (<year>2009</year>). <source>Kernel Methods for Remote Sensing Data Analysis</source>, Vol. <volume>2</volume>. <publisher-loc>West Sussex</publisher-loc>: <publisher-name>John Wiley and Sons</publisher-name>.</citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Casana</surname> <given-names>J.</given-names></name> <name><surname>Panahipour</surname> <given-names>M.</given-names></name></person-group> (<year>2014</year>). <article-title>Satellite-based monitoring of looting and damage to archaeological sites in Syria</article-title>. <source>J East. Mediterr. Archaeol. Heritage Stud.</source> <volume>2</volume>, <fpage>128</fpage>&#x02013;<lpage>151</lpage>.<pub-id pub-id-type="doi">10.5325/jeasmedarcherstu.2.2.0128</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Celik</surname> <given-names>T.</given-names></name></person-group> (<year>2009a</year>). <article-title>Unsupervised change detection in satellite images using principal component analysis and-means clustering</article-title>. <source>IEEE Geosci. Remote Sensing Lett.</source> <volume>6</volume>, <fpage>772</fpage>&#x02013;<lpage>776</lpage>.<pub-id pub-id-type="doi">10.1109/LGRS.2009.2025059</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Celik</surname> <given-names>T.</given-names></name></person-group> (<year>2009b</year>). <article-title>Multiscale change detection in multitemporal satellite images</article-title>. <source>IEEE Geosci. Remote Sensing Lett.</source> <volume>6</volume>, <fpage>820</fpage>&#x02013;<lpage>824</lpage>.<pub-id pub-id-type="doi">10.1109/LGRS.2009.2026188</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Celik</surname> <given-names>T.</given-names></name></person-group> (<year>2010</year>). <article-title>Change detection in satellite images using a genetic algorithm approach</article-title>. <source>IEEE Geosci. Remote Sensing Lett.</source> <volume>7</volume>, <fpage>386</fpage>&#x02013;<lpage>390</lpage>.<pub-id pub-id-type="doi">10.1109/LGRS.2009.2037024</pub-id></citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chandrashekar</surname> <given-names>A.</given-names></name> <name><surname>Granger</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). <article-title>Derivation of a novel efficient supervised learning algorithm from cortical-subcortical loops</article-title>. <source>Front. Comput. Neurosci.</source> <volume>5</volume>:<fpage>50</fpage>.<pub-id pub-id-type="doi">10.3389/fncom.2011.00050</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chandrashekar</surname> <given-names>A.</given-names></name> <name><surname>Torresani</surname> <given-names>L.</given-names></name> <name><surname>Granger</surname> <given-names>R.</given-names></name></person-group> (<year>2014</year>). <article-title>Learning what is where from unlabeled images: joint localization and clustering of foreground objects</article-title>. <source>Mach. Learn.</source> <volume>94</volume>, <fpage>261</fpage>&#x02013;<lpage>279</lpage>.<pub-id pub-id-type="doi">10.1007/s10994-013-5330-2</pub-id></citation></ref>
<ref id="B14"><citation citation-type="web"><person-group person-group-type="author"><name><surname>Chulov</surname> <given-names>M.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;How an arrest in Iraq revealed Isis&#x02019;s &#x00024;2 bn jihadist network,&#x0201D;</article-title> in <source>The Guardian</source>, Vol. <volume>15</volume>. Available at: <uri xlink:href="http://www.theguardian.com/world/2014/jun/15/iraq-isis-arrest-jihadists-wealth-power">http://www.theguardian.com/world/2014/jun/15/iraq-isis-arrest-jihadists-wealth-power</uri></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Contreras</surname> <given-names>D. A.</given-names></name> <name><surname>Brodie</surname> <given-names>N.</given-names></name></person-group> (<year>2013</year>). <article-title>The utility of publicly-available satellite imagery for investigating looting of archaeological sites in Jordan</article-title>. <source>J. Field Archaeol.</source> <volume>35</volume>, <fpage>101</fpage>&#x02013;<lpage>114</lpage>.<pub-id pub-id-type="doi">10.1179/009346910X12707320296838</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coppin</surname> <given-names>P.</given-names></name> <name><surname>Jonckheere</surname> <given-names>I.</given-names></name> <name><surname>Nackaerts</surname> <given-names>K.</given-names></name> <name><surname>Muys</surname> <given-names>B.</given-names></name> <name><surname>Lambin</surname> <given-names>E.</given-names></name></person-group> (<year>2004</year>). <article-title>Digital change detection methods in natural ecosystem monitoring: a review</article-title>. <source>Int. J. Remote Sens.</source> <volume>25</volume>, <fpage>1565</fpage>&#x02013;<lpage>1596</lpage>.<pub-id pub-id-type="doi">10.1080/0143116031000101675</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dai</surname> <given-names>X.</given-names></name> <name><surname>Khorram</surname> <given-names>S.</given-names></name></person-group> (<year>1998</year>). <article-title>The effects of image misregistration on the accuracy of remotely sensed change detection</article-title>. <source>IEEE Trans. Geosci. Remote Sensing</source> <volume>36</volume>, <fpage>1566</fpage>&#x02013;<lpage>1577</lpage>.<pub-id pub-id-type="doi">10.1109/36.718860</pub-id></citation></ref>
<ref id="B18"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Dalal</surname> <given-names>N.</given-names></name> <name><surname>Triggs</surname> <given-names>B.</given-names></name></person-group> (<year>2005</year>). <article-title>&#x0201C;Histograms of oriented gradients for human detection,&#x0201D;</article-title> in <conf-name>2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR&#x02019;05)</conf-name>, Vol. <volume>1</volume> (<conf-loc>San Diego</conf-loc>), <fpage>886</fpage>&#x02013;<lpage>893</lpage>.</citation></ref>
<ref id="B19"><citation citation-type="web"><person-group person-group-type="author"><name><surname>Felch</surname> <given-names>J.</given-names></name></person-group> (<year>2014</year>). <source>Dante&#x02019;s Inference</source>. Available at: <uri xlink:href="http://chasingaphrodite.com/2014/11/18/dantis-inference-the-known-unknowns-of-isis-and-antiquities-looting/">http://chasingaphrodite.com/2014/11/18/dantis-inference-the-known-unknowns-of-isis-and-antiquities-looting/</uri></citation></ref>
<ref id="B20"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Galton</surname> <given-names>F.</given-names></name></person-group> (<year>1894</year>). <source>Natural Inheritance</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Macmillan</publisher-name>.</citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gill</surname> <given-names>D. W. J.</given-names></name></person-group> (<year>2016</year>). <article-title>Context matters: the auction market and due diligence: the need for action</article-title>. <source>J. Art Crime</source> <volume>15</volume>, <fpage>73</fpage>&#x02013;<lpage>77</lpage>.</citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hartigan</surname> <given-names>J. A.</given-names></name> <name><surname>Wong</surname> <given-names>M. A.</given-names></name></person-group> (<year>1979</year>). <article-title>Algorithm AS 136: a k-means clustering algorithm</article-title>. <source>J. R. Stat. Soc. Ser. C Appl. Stat.</source> <volume>28</volume>, <fpage>100</fpage>&#x02013;<lpage>108</lpage>.</citation></ref>
<ref id="B23"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Hofmann</surname> <given-names>P.</given-names></name></person-group> (<year>2001</year>). <article-title>&#x0201C;Detecting informal settlements from IKONOS image data using methods of object oriented image analysis-an example from Cape Town (South Africa),&#x0201D;</article-title> in <source>Remote Sensing of Urban Areas/Fernerkundung in urbanen R&#x000E4;umen</source>, ed. <person-group person-group-type="editor"><name><surname>J&#x000FC;rgens</surname> <given-names>C.</given-names></name></person-group> (<publisher-loc>Regensburg</publisher-loc>), <fpage>41</fpage>&#x02013;<lpage>42</lpage>. Available at: <uri xlink:href="http://www.dsr.inpe.br/pagina_dsr_ingles/pessoal/hermann/RGS35_home.pdf">http://www.dsr.inpe.br/pagina_dsr_ingles/pessoal/hermann/RGS35_home.pdf</uri></citation></ref>
<ref id="B24"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Hofmann</surname> <given-names>P.</given-names></name> <name><surname>Strobl</surname> <given-names>J.</given-names></name> <name><surname>Blaschke</surname> <given-names>T.</given-names></name> <name><surname>Kux</surname> <given-names>H.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Detecting informal settlements from QuickBird data in Rio de Janeiro using an object based approach,&#x0201D;</article-title> in <source>Object-Based Image Analysis</source>, eds <person-group person-group-type="editor"><name><surname>Blaschke</surname> <given-names>T.</given-names></name> <name><surname>Lang</surname> <given-names>S.</given-names></name> <name><surname>Hay</surname> <given-names>G. J.</given-names></name></person-group> (<publisher-loc>Berlin &#x00026; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>531</fpage>&#x02013;<lpage>553</lpage>.</citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hritz</surname> <given-names>C.</given-names></name></person-group> (<year>2008</year>). <article-title>Remote sensing of cultural heritage in Iraq: a case study of Isin</article-title>. <source>TAARII Newslett.</source> <volume>3</volume>, <fpage>1</fpage>&#x02013;<lpage>8</lpage>.</citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ikram</surname> <given-names>S.</given-names></name></person-group> (<year>2013</year>). <article-title>Cultural heritage in times of crisis: the view from Egypt</article-title>. <source>J. East. Mediterr. Archaeol. Heritage Stud.</source> <volume>1</volume>, <fpage>366</fpage>&#x02013;<lpage>371</lpage>.<pub-id pub-id-type="doi">10.5325/jeasmedarcherstu.1.4.0366</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ikram</surname> <given-names>S.</given-names></name> <name><surname>Hanna</surname> <given-names>M.</given-names></name></person-group> (<year>2013</year>). <article-title>Looting and land grabbing: the current situation in Egypt</article-title>. <source>Bull. Am. Res. Center Egypt</source> <volume>202</volume>, <fpage>34</fpage>&#x02013;<lpage>39</lpage>.</citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ioannidis</surname> <given-names>C.</given-names></name> <name><surname>Psaltis</surname> <given-names>C.</given-names></name> <name><surname>Potsiou</surname> <given-names>C.</given-names></name></person-group> (<year>2009</year>). <article-title>Towards a strategy for control of suburban informal buildings through automatic change detection</article-title>. <source>Comput. Environ. Urban Syst.</source> <volume>33</volume>, <fpage>64</fpage>&#x02013;<lpage>74</lpage>.<pub-id pub-id-type="doi">10.1016/j.compenvurbsys.2008.09.010</pub-id></citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kass</surname> <given-names>M.</given-names></name> <name><surname>Witkin</surname> <given-names>A.</given-names></name> <name><surname>Terzopoulos</surname> <given-names>D.</given-names></name></person-group> (<year>1988</year>). <article-title>Snakes: active contour models</article-title>. <source>Int. J. Comput. Vis.</source> <volume>1</volume>, <fpage>321</fpage>&#x02013;<lpage>331</lpage>.<pub-id pub-id-type="doi">10.1007/BF00133570</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>R.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Baba</surname> <given-names>T.</given-names></name> <name><surname>Masumoto</surname> <given-names>D.</given-names></name> <name><surname>Nagata</surname> <given-names>S.</given-names></name></person-group> (<year>2008</year>). <article-title>SVM-based active feedback in image retrieval using clustering and unlabeled data</article-title>. <source>Pattern Recognit.</source> <volume>41</volume>, <fpage>2645</fpage>&#x02013;<lpage>2655</lpage>.<pub-id pub-id-type="doi">10.1016/j.patcog.2008.01.023</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lowe</surname> <given-names>D. G.</given-names></name></person-group> (<year>2004</year>). <article-title>Distinctive image features from scale-invariant keypoints</article-title>. <source>Int. J. Comput. Vis.</source> <volume>60</volume>, <fpage>91</fpage>&#x02013;<lpage>110</lpage>.<pub-id pub-id-type="doi">10.1023/B:VISI.0000029664.99615.94</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lu</surname> <given-names>D.</given-names></name> <name><surname>Mausel</surname> <given-names>P.</given-names></name> <name><surname>Brondizio</surname> <given-names>E.</given-names></name> <name><surname>Moran</surname> <given-names>E.</given-names></name></person-group> (<year>2004</year>). <article-title>Change detection techniques</article-title>. <source>Int. J. Remote Sens.</source> <volume>25</volume>, <fpage>2365</fpage>&#x02013;<lpage>2401</lpage>.<pub-id pub-id-type="doi">10.1080/0143116031000139863</pub-id></citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mas</surname> <given-names>J.-F.</given-names></name></person-group> (<year>1999</year>). <article-title>Monitoring land-cover changes: a comparison of change detection techniques</article-title>. <source>Int. J. Remote Sens.</source> <volume>20</volume>, <fpage>139</fpage>&#x02013;<lpage>152</lpage>.<pub-id pub-id-type="doi">10.1080/014311699213659</pub-id></citation></ref>
<ref id="B34"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Mayunga</surname> <given-names>S. D.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Coleman</surname> <given-names>D. J.</given-names></name></person-group> (<year>2005</year>). <article-title>&#x0201C;Semi-automatic building extraction utilizing Quickbird imagery,&#x0201D;</article-title> in <source>Proc. ISPRS Workshop CMRT</source>, Vol. <volume>13</volume> (<publisher-loc>Vienna</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>136</lpage>.</citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mountrakis</surname> <given-names>G.</given-names></name> <name><surname>Im</surname> <given-names>J.</given-names></name> <name><surname>Ogole</surname> <given-names>C.</given-names></name></person-group> (<year>2011</year>). <article-title>Support vector machines in remote sensing: a review</article-title>. <source>ISPRS J. Photogramm. Remote Sensing</source> <volume>66</volume>, <fpage>247</fpage>&#x02013;<lpage>259</lpage>.<pub-id pub-id-type="doi">10.1016/j.isprsjprs.2010.11.001</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Newson</surname> <given-names>P.</given-names></name> <name><surname>Young</surname> <given-names>R.</given-names></name></person-group> (<year>2015</year>). <article-title>The archaeology of conflict-damaged sites: Hosn Niha in the Biqa&#x02019; Valley, Lebanon</article-title>. <source>Antiquity</source> <volume>89</volume>, <fpage>449</fpage>&#x02013;<lpage>463</lpage>.<pub-id pub-id-type="doi">10.15184/aqy.2015.4</pub-id></citation></ref>
<ref id="B37"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Opitz</surname> <given-names>D.</given-names></name> <name><surname>Blundell</surname> <given-names>S.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Object recognition and image segmentation: the Feature Analyst<sup>&#x000AE;</sup> approach,&#x0201D;</article-title> in <source>Object-Based Image Analysis</source>, eds <person-group person-group-type="editor"><name><surname>Blaschke</surname> <given-names>T.</given-names></name> <name><surname>Lang</surname> <given-names>S.</given-names></name> <name><surname>Hay</surname> <given-names>G. J.</given-names></name></person-group> (<publisher-loc>Berlin &#x00026; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>153</fpage>&#x02013;<lpage>167</lpage>.</citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pal</surname> <given-names>M.</given-names></name> <name><surname>Mather</surname> <given-names>P. M.</given-names></name></person-group> (<year>2005</year>). <article-title>Support vector machines for classification in remote sensing</article-title>. <source>Int. J. Remote Sens.</source> <volume>26</volume>, <fpage>1007</fpage>&#x02013;<lpage>1011</lpage>.<pub-id pub-id-type="doi">10.1080/01431160512331314083</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Parcak</surname> <given-names>S.</given-names></name></person-group> (<year>2007</year>). <article-title>Satellite remote sensing methods for monitoring archaeological tells in the Middle East</article-title>. <source>J. Field Archaeol.</source><pub-id pub-id-type="doi">10.1179/009346907791071773</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Parcak</surname> <given-names>S.</given-names></name> <name><surname>Gathings</surname> <given-names>D.</given-names></name> <name><surname>Childs</surname> <given-names>C.</given-names></name> <name><surname>Mumford</surname> <given-names>G.</given-names></name> <name><surname>Cline</surname> <given-names>E.</given-names></name></person-group> (<year>2016</year>). <article-title>Satellite evidence of archaeological site looting in Egypt: 2002&#x02013;2013</article-title>. <source>Antiquity</source> <volume>90</volume>, <fpage>188</fpage>&#x02013;<lpage>205</lpage>.<pub-id pub-id-type="doi">10.15184/aqy.2016.1</pub-id></citation></ref>
<ref id="B41"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Parcak</surname> <given-names>S. H.</given-names></name></person-group> (<year>2009</year>). <source>Satellite Remote Sensing for Archaeology</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Routledge</publisher-name>.</citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Peng</surname> <given-names>J.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name></person-group> (<year>2005</year>). <article-title>An improved snake model for building detection from urban aerial images</article-title>. <source>Pattern Recognit. Lett.</source> <volume>26</volume>, <fpage>587</fpage>&#x02013;<lpage>595</lpage>.<pub-id pub-id-type="doi">10.1016/j.patrec.2004.09.033</pub-id></citation></ref>
<ref id="B43"><citation citation-type="web"><person-group person-group-type="author"><name><surname>Pringle</surname> <given-names>H.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;ISIS cashing in on looted antiquities to fuel Iraq insurgency,&#x0201D;</article-title> in <source>National Geographic</source>, Vol. <volume>27</volume>. Available at: <uri xlink:href="http://news.nationalgeographic.com/news/2014/06/140626-isis-insurgents-syria-iraq-looting-antiquities-archaeology/">http://news.nationalgeographic.com/news/2014/06/140626-isis-insurgents-syria-iraq-looting-antiquities-archaeology/</uri></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Proulx</surname> <given-names>B. B.</given-names></name></person-group> (<year>2013</year>). <article-title>Archaeological site looting in &#x0201C;glocal&#x0201D; perspective: nature, scope, and frequency</article-title>. <source>Am. J. Archaeol.</source> <volume>117</volume>, <fpage>111</fpage>&#x02013;<lpage>125</lpage>.<pub-id pub-id-type="doi">10.3764/aja.117.1.0111</pub-id></citation></ref>
<ref id="B45"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Rekik</surname> <given-names>A.</given-names></name> <name><surname>Zribi</surname> <given-names>M.</given-names></name> <name><surname>Benjelloun</surname> <given-names>M.</given-names></name> <name><surname>ben Hamida</surname> <given-names>A.</given-names></name></person-group> (<year>2006</year>). <article-title>&#x0201C;A k-means clustering algorithm initialization for unsupervised statistical satellite image segmentation,&#x0201D;</article-title> in <conf-name>2006 1st IEEE International Conference on E-Learning in Industrial Electronics</conf-name> (<conf-loc>Hammamet</conf-loc>), <fpage>11</fpage>&#x02013;<lpage>16</lpage>.</citation></ref>
<ref id="B46"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Santosh</surname> <given-names>K. C.</given-names></name> <name><surname>Wendling</surname> <given-names>L.</given-names></name> <name><surname>Antani</surname> <given-names>S.</given-names></name> <name><surname>Thoma</surname> <given-names>G. R.</given-names></name></person-group> (<year>2016</year>). <article-title>Overlaid arrow detection for labeling regions of interest in biomedical images</article-title>. <source>IEEE Intell. Syst.</source> <volume>31</volume>, <fpage>66</fpage>&#x02013;<lpage>75</lpage>.<pub-id pub-id-type="doi">10.1109/MIS.2016.24</pub-id></citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname> <given-names>A.</given-names></name></person-group> (<year>1989</year>). <article-title>Review article digital change detection techniques using remotely-sensed data</article-title>. <source>Int. J. Remote Sens.</source> <volume>10</volume>, <fpage>989</fpage>&#x02013;<lpage>1003</lpage>.<pub-id pub-id-type="doi">10.1080/01431168908903939</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sirmacek</surname> <given-names>B.</given-names></name> <name><surname>Unsalan</surname> <given-names>C.</given-names></name></person-group> (<year>2009</year>). <article-title>Urban-area and building detection using SIFT keypoints and graph theory</article-title>. <source>IEEE Trans. Geosci. Remote Sensing</source> <volume>47</volume>, <fpage>1156</fpage>&#x02013;<lpage>1167</lpage>.<pub-id pub-id-type="doi">10.1109/TGRS.2008.2008440</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stone</surname> <given-names>E. C.</given-names></name></person-group> (<year>2008</year>). <article-title>Patterns of looting in southern Iraq</article-title>. <source>Antiquity</source> <volume>82</volume>, <fpage>125</fpage>&#x02013;<lpage>138</lpage>.<pub-id pub-id-type="doi">10.1017/S0003598X00096496</pub-id></citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tapete</surname> <given-names>D.</given-names></name> <name><surname>Cigna</surname> <given-names>F.</given-names></name> <name><surname>Donoghue</surname> <given-names>D. N.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x02018;Looting marks&#x02019; in space-borne SAR imagery: measuring rates of archaeological looting in Apamea (Syria) with TerraSAR-X Staring Spotlight</article-title>. <source>Remote Sensing Environ.</source> <volume>178</volume>, <fpage>42</fpage>&#x02013;<lpage>58</lpage>.<pub-id pub-id-type="doi">10.1016/j.rse.2016.02.055</pub-id></citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Thomas</surname> <given-names>D. C.</given-names></name> <name><surname>Kidd</surname> <given-names>F. J.</given-names></name> <name><surname>Nikolovski</surname> <given-names>S.</given-names></name> <name><surname>Zipfel</surname> <given-names>C.</given-names></name></person-group> (<year>2008</year>). <article-title>The archaeological sites of Afghanistan in google Earth</article-title>. <source>AARGnews</source> <volume>37</volume>, <fpage>22</fpage>&#x02013;<lpage>30</lpage>.</citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Volpi</surname> <given-names>M.</given-names></name> <name><surname>Tuia</surname> <given-names>D.</given-names></name> <name><surname>Kanevski</surname> <given-names>M.</given-names></name></person-group> (<year>2012a</year>). <article-title>Memory-based cluster sampling for remote sensing image classification</article-title>. <source>IEEE Trans. Geosci. Remote Sensing</source> <volume>50</volume>, <fpage>3096</fpage>&#x02013;<lpage>3106</lpage>.<pub-id pub-id-type="doi">10.1109/TGRS.2011.2179661</pub-id></citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Volpi</surname> <given-names>M.</given-names></name> <name><surname>Tuia</surname> <given-names>D.</given-names></name> <name><surname>Camps-Valls</surname> <given-names>G.</given-names></name> <name><surname>Kanevski</surname> <given-names>M.</given-names></name></person-group> (<year>2012b</year>). <article-title>Unsupervised change detection with kernels</article-title>. <source>IEEE Geosci. Remote Sensing Lett.</source> <volume>9</volume>, <fpage>1026</fpage>&#x02013;<lpage>1030</lpage>.<pub-id pub-id-type="doi">10.1109/LGRS.2012.2189092</pub-id></citation></ref>
<ref id="B54"><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>Y.</given-names></name> <name><surname>Zhao</surname> <given-names>Z.</given-names></name> <name><surname>Song</surname> <given-names>J.</given-names></name></person-group> (<year>2004</year>). <article-title>&#x0201C;Urban building extraction from high-resolution satellite panchromatic image using clustering and edge detection,&#x0201D;</article-title> in <conf-name>Geoscience and Remote Sensing Symposium, 2004. IGARSS&#x02019;04. Proceedings. 2004 IEEE International</conf-name>, Vol. <volume>3</volume> (<conf-loc>Anchorage</conf-loc>), <fpage>2008</fpage>&#x02013;<lpage>2010</lpage>.</citation></ref>
<ref id="B55"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Clayton</surname> <given-names>M. K.</given-names></name> <name><surname>Townsend</surname> <given-names>P. A.</given-names></name></person-group> (<year>2011</year>). <article-title>Functional concurrent linear regression model for spatial images</article-title>. <source>J. Agric. Biol. Environ. Stat.</source> <volume>16</volume>, <fpage>105</fpage>&#x02013;<lpage>130</lpage>.<pub-id pub-id-type="doi">10.1007/s13253-010-0047-1</pub-id></citation></ref>
</ref-list>
</back>
</article>