Our objective is to lower the barrier of executing spatial image computations in a computer cluster/cloud environment. We research two related problems encountered during an execution of spatial computations over terabyte-sized (TB) images using Apache Hadoop running on distributed computing resources. The two problems encountered were:
The first problem is solved by designing an iterative estimation methodology. The second problem is formulated as an optimization over three partitioning schemas (physical, logical without overlap and logical with overlap), and evaluated over several system configuration parameters. Our experimental results for the two problems demonstrate 100% accuracy in detecting spatial computations in the Java Advanced Imaging and ImageJ libraries. Also, it resulted in a speed-up of 5.36 between the default Hadoop physical partitioning and developed logical image partitioning with overlap; and 3.14 times faster execution of logical partitioning with overlap than without overlap. The novelty of our work is in designing an extension to Apache Hadoop to run a class of spatial image processing operations efficiently on a distributed computing resource.
The paper: Peter Bajcsy, Phuong Nguyen, Antoine Vandecreme, and Mary Brady, “Spatial Computations over Terabyte-Sized Images on Hadoop Platforms,“ 2014 IEEE International Conference on Big Data (IEEE BigData 2014), October 27-30, 2014, Washington DC, USA (submitted). (download pdf)
The paper: Peter Bajcsy, Phuong Nguyen, Antoine Vandecreme, and Mary Brady, “Spatial Computations over Terabyte-Sized Images on Hadoop Platforms,“ 2014 IEEE International Conference on Big Data (IEEE BigData 2014), October 27-30, 2014, Washington DC, USA (submitted). (download pdf)