Skip to content
This repository has been archived by the owner on Nov 10, 2020. It is now read-only.

Moving to https:/github.com/uacic. Transforms RGB bin files to geo-referenced tiff (based upon TERRA REF extractors-stereo-rgb/bin2tif)

License

Notifications You must be signed in to change notification settings

AgPipeline/transformer-bin2tif

Transformer Bin2Tif

Stereo RGB Image Bin to GeoTIFF Converter.

Sample Docker Command line

Below is a sample command line that shows how the bin2tif Docker image could be run. An explanation of the command line options used follows. Be sure to read up on the docker run command line for more information.

docker run --rm --mount "src=/home/test,target=/mnt,type=bind" agpipeline/bin2tif:2.0 --working_space "/mnt" --metadata "/mnt/08f445ef-b8f9-421a-acf1-8b8c206c1bb8_metadata_cleaned.json" "/mnt/08f445ef-b8f9-421a-acf1-8b8c206c1bb8_left.bin"

This example command line assumes the source files are located in the /home/test folder of the local machine. The name of the image to run is agpipeline/bin2tif:2.0.

We are using the same folder for the source files and the output files. By using multiple --mount options, the source and output files can be located in separate folders.

Docker commands
Everything between 'docker' and the name of the image are docker commands.

  • run indicates we want to run an image
  • --rm automatically delete the image instance after it's run
  • --mount "src=/home/test,target=/mnt,type=bind" mounts the /home/test folder to the /mnt folder of the running image

We mount the /home/test folder to the running image to make files available to the software in the image.

Image's commands
The command line parameters after the image name are passed to the software inside the image. Note that the paths provided are relative to the running image (see the --mount option specified above).

  • --working_space "/mnt" specifies the folder to use as a workspace
  • --metadata "/mnt/08f445ef-b8f9-421a-acf1-8b8c206c1bb8_metadata.cleaned.json" is the name of the source metadata to be cleaned
  • "/mnt/08f445ef-b8f9-421a-acf1-8b8c206c1bb8_left.bin" is the name of the raw image to convert

Original README

Below are the contents of the original README.md for the TERRA REF project. It is included here as reference material for the history of this transformer.

Authors:

  • Zongyang Li, Donald Danforth Plant Science Center, St. Louis, MO
  • Maxwell Burnette, National Supercomputing Applications, Urbana, Il
  • Robert Pless, George Washington University, Washington, DC

Overview

This extractor processes binary stereo images using metadata and outputs JPG and TIFF images.

The extractor execution is triggered when required input files are added to a dataset.

Input

  • Following data must be found
    • _left.bin image
    • _right.bin image
    • dataset metadata for the left+right capture dataset; can be attached as Clowder metadata or included as a metadata.json file

Output

  • The dataset containing the left/right binary images will get left/right JPG and geoTIFF images.
  • Metadata, including geospatial information will be added to the geoTIFF file as well as to the geostreams database.

Algorithm

Algorithm Description

  1. Convert raw data to 3 channels color image

    Stereo RGB camera uses a single charge-coupled device (CCD) sensor, with the CCD pixels preceded in the optical path by a color filter array(CFA) in a Bayer mosaic pattern. For each 2x2 set of pixels, two diagonally opposed pixels have green filters, and the other two have red and blue filters. We assume these pixels are in a GBRG ordering, and use bilinear interpolation to do the demosaicing. This means that three color planes are independently interpolated using symmetric bilinear interpolation from the nearest neighbors of the same color.

Reference: Malvar, H.S., L. He, and R. Cutler, High quality linear interpolation for demosaicing of Bayer-patterned color images. ICASPP, Volume 34, Issue 11, pp. 2274-2282, May 2004.

  1. Steps for geo-referencing bounding box to each image.

    a. Get image shape from metadata,

    b. Get camera center position from metadata

    c. Compute field of view for the image:

     i. The JSON data reports the camera field of view as "the field of view for a scene 2 meters away is: “0.749m x 1.015m"
    
     ii. Predict fov for each image should be:
    
         * fix_fov = fov_in_2_meter*(camera_height/2)
    
     iii. In implementing the stitching process, we required two magic numbers that are computed experimentally to get the correct geometric alignment.  Our experimentally determined values are:
    
          * HEIGHT_MAGIC_NUMBER = 1.64
    
          * PREDICT_MAGIC_SLOPE = 0.574
    
     iv. These numbers are used in the following equations.
    
           * predict_plant_height = PREDICT_MAGIC_SLOPE * camHeight
    
         * camH_fix = camHeight + HEIGHT_MAGIC_NUMBER - predict_plant_height
    
           * fix_fov_x = fov_x*(camH_fix/2)
    
           * fix_fov_y = fov_y*(camH_fix/2)
    

    e. Steps for computing the Magic Numbers:

     i. Make an assumption that the real FOV can be derived as a linear function base on the given FOV.
     ii. Randomly pick several different days stereoTop RGB raw data in different camera height, try to get a best stitched full field map for each day empirically(by personal judgement), and record all the FOV for each day.
     iii. Apply a linear regression to output HEIGHT_MAGIC_NUMBER and PREDICT_MAGIC_SLOPE
    

    d. Compute geo-reference bounding box

       i. Convert coordinates from Scanalyzer to MAC coordinates using formula from [https://terraref.gitbooks.io/terraref-documentation/content/user/geospatial-information.html](https://terraref.gitbooks.io/terraref-documentation/content/user/geospatial-information.html)
    
      ii. Use utm tools to convert coordinates from MAC to lat/lon
    

    e. Using osgeo.gdal, associate image with geospatial bounding box, and create geoTIFF.

Parameters

HEIGHT_MAGIC_NUMBER and PREDICT_MAGIC_SLOPE were applied when we estimate the field of view, these are empirically derived values as described in the above algorithm.

The geo-reference bounding box is based on the assumption that image is aligned to geographic coordinates, so that moving up in the image corresponds to moving exactly north. We expect that error in the relative location of pixels within an image introduced by this assumption is much less than the pixel resolution.

Limitations

  1. Any stitched image introduces new artifacts into the image data; it always introduces edges at the boundary of where one image turns into another --- either an explicitly black line boundary or an implicit boundary that is there because you can't exactly stitch images of a complicated 3D world (without making a full 3D model). Even if you could stitch them, the same bit of the world is usually a different brightness when viewed from different directions.
  2. The stitched full field image may have artifacts that arise from harsh shadows in some imaging conditions.
  3. One of the artifacts is duplication of area, this is unavoidable without a much more complex stitching algorithm that implicitly infers the 3D structure of the ground. The justification for not going for such a rich representation is that:
  4. for the plants, since they move, it would be impossible not to have artifacts at the edges of the image, and
  5. for the ground, we judged that small stitching errors were not worth the (substantial) additional effort to build the more complete model.

Failure Conditions

  • If the camera is moved in the gantry box, then the magic numbers may have to be recalculated or experimentally determined.

  • If the camera is not aligned north-south, then the geo-bounding box may not be accurate.

Application

Files

  • Dockerfile: defines docker image with all dependencies
  • batch_launcher.sh: used to submit jobs on ROGER cluster
  • bin_to_geotiff.py:
  • terra_bin2tif.py: extractor (wrapper for bin_to_geotiff.py)
  • entrypoint.sh, extractor_info.json, terra.bin2tif.service: Clowder utilities

Docker

The Dockerfile included in this directory can be used to launch this extractor in a docker container.

Building the Docker image

docker build -f Dockerfile -t terra-ext-bin2tif .

Running the image locally

docker run \
  -p 5672 -p 9000 --add-host="localhost:{LOCAL_IP}" \
  -e RABBITMQ_URI=amqp://{RMQ_USER}:{RMQ_PASSWORD}@localhost:5672/%2f \
  -e RABBITMQ_EXCHANGE=clowder \
  -e REGISTRATION_ENDPOINTS=http://localhost:9000/clowder/api/extractors?key={SECRET_KEY} \
  terra-ext-bin2tif

Note: by default RabbitMQ will not allow "guest:guest" access to non-local addresses, which includes Docker. You may need to create an additional local RabbitMQ user for testing.

Running the image remotely

docker run \
  -e RABBITMQ_URI=amqp://{RMQ_USER}:{RMQ_PASSWORD}@rabbitmq.ncsa.illinois.edu/clowder \
  -e RABBITMQ_EXCHANGE=terra \
  -e REGISTRATION_ENDPOINTS=http://terraref.ncsa.illinosi.edu/clowder//api/extractors?key={SECRET_KEY} \
  terra-ext-bin2tif

Cluster Computing with TORQUE/PBS

The extractor can also be run on a compute cluster via the TORQUE/PBS batch system. These instructions are for use on ROGER.

This process assumes that you are using the existing Python virtualenv under:

/projects/arpae/terraref/shared/extractors/pyenv/

This also uses a shared environment file for common settings:

/projects/arpae/terraref/shared/extractors/env.sh

The following default batch jobs will start 20 extractors on a single 20-core node:

qsub /projects/arpae/terraref/shared/extractors/extractors-stereo-rgb/bin2tif/batch_launcher.sh

Dependencies

  • All of the Python scripts syntactically support Python >= 2.7. Please make sure that the Python in the running environment is in appropriate version.

  • All the Python scripts also rely on the third-party library including: PIL, scipy, numpy and osgeo.

About

Moving to https:/github.com/uacic. Transforms RGB bin files to geo-referenced tiff (based upon TERRA REF extractors-stereo-rgb/bin2tif)

Topics

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published