Tutorial: Create drone imagery products in ArcGIS Reality for ArcGIS Pro

Available with Standard or Advanced license.

Available for an ArcGIS organization with the ArcGIS Reality license.

In ArcGIS Reality for ArcGIS Pro, you can use photogrammetry to correct drone imagery to remove geometric distortions caused by the sensor, platform, and terrain displacement. After removing these distortions, you can generate Reality mapping products.

First, you will set up a Reality mapping workspace to manage the drone imagery collection. Next, you will perform a block adjustment, followed by a refined adjustment using ground control points. Finally, you'll generate a digital surface model (DSM), true ortho, and 2D DSM mesh.

ArcGIS Reality for ArcGIS Pro requires information about the camera, including focal length and sensor size, as well as the location at which each image was captured. This information is commonly stored as metadata in the image files, typically in the EXIF header. It is also helpful to know the GPS accuracy. In the case of drone imagery, this should be provided by the drone manufacturer. The GPS accuracy for the sample dataset in this tutorial is better than 5 meters.

License:

The following are required to complete this tutorial:

  • ArcGIS Reality Studio
  • ArcGIS Pro 3.1, or later
  • ArcGIS Reality for ArcGIS Pro Setup

Create a Reality mapping workspace

A Reality mapping workspace is an ArcGIS Pro subproject that is dedicated to Reality mapping workflows. It is a container in an ArcGIS Pro project folder that stores the resources and derived files that belong to a single image collection in a Reality mapping task.

A collection of 195 drone images is provided for this tutorial. The GCP folder contains the Tut_Ground_Control_UTM11N_EGM96.csv file, which contains ground control points (GCPs), and a file geodatabase (GCP_desc.gdb) with a Ground_Control_NAD832011_NAVD88 point feature class identifying the GCP locations at the site.

To create a Reality mapping workspace, complete the following steps:

  1. Download the tutorial dataset. unzip it, and save contents to C:\SampleData\RM_Drone_tutorial.
    1. Unzip the package in the C:\SampleData\Aerial Imagery directory.
  2. In ArcGIS Pro, create a project using the Map template and sign in to your ArcGIS Online account if necessary.
  3. Increase the processing speed of block adjustment.
    1. On the Analysis tab, click Environments.
    2. Locate the Parallel Processing parameter in the Environments window and change the value to 90%.
  4. On the Imagery tab, in the Reality Mapping group, click the New Workspace drop-down menu and choose New Workspace.
  5. In the Workspace Configuration window, type a name for the workspace.
  6. Ensure that Workspace Type is set to Reality Mapping.
  7. Click the Sensor Data Type drop-down menu, and choose Drone.
  8. Click the Basemap drop-down menu, and choose Topographic.
  9. Accept all other default values and click Next.

    The Image Collection window appears.

  10. In the Image Collection window, click the Sensor Type drop-down menu, and choose Generic.

    This option is used because the imagery was collected with an RGB camera.

  11. Click Add, browse to the tutorial data location, and select the Images folder.
  12. Note:
    Most modern drones store GPS information in the EXIF header. This will be used to automatically populate the table shown below. However, some older systems or custom-built drones may store GPS data in an external file. In this case, you can use the Import button Import next to the Geolocation parameter to import external GPS files.

    The Spatial Reference and Camera Model values are automatically populated. The workspace requires a projected coordinate system. The autogenerated, default Workspace Spatial Reference is based on the matching WGS84 UTM coordinates for the images. This projection determines the spatial reference for the ArcGIS Reality for ArcGIS Pro products.

    Drone Image Collection window

  13. Click Next.
  14. In the Data Loader Options page, for Elevation Source, select from the drop-down list, Average Elevation from DEM. The World Elevation Service will be used as the Elevation Source.

    Data Loader Options window

    • If you have access to the internet, the Elevation Source value is derived from the World Elevation Service. This provides an initial estimate of the flight height for each image.
    • If you do not have access to the internet or a DEM, choose Constant Elevation from the Elevation Source drop-down menu and enter an elevation value of 414 meters.
    • The parameter, Flight Height Above Terrain (m), refers to the image exclusion height, where images having a flight height above terrain that is less than this value will not be included in the workspace.

  15. Accept all the default settings in the Data Loader Options window and click Finish.

    Once the workspace has been created, the images, drone path, and image footprints are displayed. A Reality mapping container is added to the Contents pane, where the source imagery data and derived Reality mapping products will be stored.

    Reality mapping workspace

    The initial display of imagery in the workspace confirms that all images and necessary metadata were provided to initiate the workspace. The images have not been aligned or adjusted, so the mosaic will not look correct.

Perform a block adjustment

After the Reality mapping workspace has been created, the next step is to perform a block adjustment using the tools in the Adjust and Refine groups. The block adjustment first calculates tie points, which are common points in areas of image overlap. The tie points are then used to calculate the orientation of each image, known as exterior orientation in photogrammetry. The block adjustment process may take a few hours depending on your computer setup and resources.

To perform a block adjustment, complete the following steps:

  1. On the Reality Mapping tab, in the Adjust group, click Adjust Adjust.
  2. In the Adjust window, ensure all the Perform Camera Calibration options are checked.

    This indicates that the input focal length is approximate and that the lens distortion parameters will be calculated during adjustment. For drone imagery, these options are checked by default, since most drone cameras have not been calibrated. This option should not be checked for high-quality cameras with known calibration.

    The camera self-calibration requires that an image collection has in-strip overlap of more than 60 percent and cross-strip overlap of more than 30 percent.

  3. Expand Advanced Options.
  4. Ensure Use Orientations from Metadata option is checked.

    This reduces the adjustment duration by using the exterior orientation information embedded in the image EXIF as initial values.

  5. Uncheck Fix Image Location for High Accuracy GPS.

    This option is used only for imagery acquired with differential GPS, such as Real Time Kinematic (RTK) or Post Processing Kinematic (PPK).

  6. Optionally, uncheck Compute Posterior Standard Deviation for Images.
  7. For Image Location Accuracy, choose High from the drop-down menu.

    GPS location accuracy indicates the accuracy level of the GPS data collected concurrently with the imagery and listed in the corresponding EXIF data file. This is used in the tie point calculation algorithm to determine the number of images in the neighborhood to use. The High value is used for GPS accuracy from 0 to 10 meters.

    Adjust tool options

  8. Accept all other defaults and click Run.

    After the adjustment has been performed, the Logs file displays statistical information such as the mean reprojection error (in pixels), which signifies the accuracy of adjustment, the number of images processed, and the number of tie points generated.

    The relative accuracy of the images is also improved, and derived products can be generated using the options in the Product category. To improve the absolute accuracy of generated products, GCPs must be added to the block.

Add GCPs

GCPs are points with known x,y,z ground coordinates, which are often obtained from ground survey, that are used to ensure that the photogrammetric process has reference points on the ground. A block adjustment can be applied without GCPs and still ensure relative accuracy, although adding GCPs increases the absolute accuracy of the adjusted imagery.

Import GCPs

To import GCPs, complete the following steps:

  1. On the Reality Mapping tab, in the Refine group, click Manage GCPs.

    The GCP Manager window appears.

  2. In the GCP Manager window, click the Import GCPs button Import GCPs.
  3. In the Import GCPs window, under GCP File, browse to and select the Tut_Ground_Control_UTM11N_EGM96.csv file, and click OK.
  4. Under Set GCP Spatial Reference, click the Spatial Reference button Spatial Reference and do the following in the Spatial Reference window:
    1. For Current XY, expand Projected, UTM, North America, and NAD83, and choose NAD 1983 UTM Zone 11N.
    2. To define the vertical coordinate system, click in the box below Current Z, expand Vertical Coordinate System, Gravity-related, and World, and choose EGM96 Geoid.
  5. Click OK to accept the changes and close the Spatial Reference window.
  6. Under Geographic Transformations, click the Horizontal tab and choose WGS 1984 (ITRF00) to NAD83 from the drop-down list.
  7. Click the Vertical tab and choose None from the drop-down menu.
  8. Map the GCP fields accordingly under Field Mappings.

    Import GCPs window

  9. Click OK to import the GCPs.

    The imported GCPs are listed in the GCP Manager window, and the relative locations are displayed in the 2D Map view with a red position indicator GCP.

    Imported GCPs on map

Add tie points for selected GCPs

To add tie points, complete the following steps:

  1. Add the control point file from the file geodatabase in the tutorial data location to the 2D map view.
    1. In the Catalog pane, browse to the file by clicking SampleData > RM_Drone_Tutorial > GCP > GCP_desc.gdb, right-click the file Ground_Control_NAD832011_NAVD88, and choose Add To Current Map.
  2. In the Contents pane, right-click Ground_Control_NAD832011_NAVD88 and click Label. If the points are not labeled with the GCP control point name, do the following:
    1. With the feature class selected, on the ArcGIS Pro main menu, click Labeling.
    2. In the Label Class category, for Field, choose Control Point Name from the drop-down list.

    Once the GCPs are imported, the table in the GCP Manager becomes populated.

    GCP Manager window with a photo of the GCP location
  3. In the GCP Manager window, click the BldgE-19 GCP, and do the following to see the exact location of the GCP:
    1. Click the point labeled with the same name in the 2D map view, and scroll to the bottom of the pop-up window.
    2. If the pop-up window is covering the GCP Manager window, drag the pop-up to a different location in the map window.
    3. To view a larger scale GCP image, clicking the image.
    4. If the enlarged GCP image covers the entire window, click the Restore Down button.

  4. Click the Add GCP or Tie Point button Add GCP or Tie Point in the GCP Manager window to accurately measure GCPs in the image viewer for each image.

    The tie points for other images are automatically calculated by the image matching algorithm when possible, indicated by a blue cross next to the image ID. Review each tie point for accuracy. If the tie point is not automatically identified, add the tie point manually by selecting the appropriate location in the image.

  5. Click a single GCP listed in the upper panel of the GCP manager, do the measurement steps described below, and repeat the process for all GCPs.

    A list of images containing the GCP will be displayed in the lower panel of the GCP Manager pane.

    1. Click on each image listed, and reposition the GCP displayed in the image viewer on the exact location of the surveyed image panel target visible in the image.

    Zoom and pan in the image window by right clicking in the image window while scrolling the mouse roller.

  6. After each GCP has been added and measured with tie points, select BldgE-19, right-click, and change it to a check point.

    This provides a measure of the absolute accuracy of the adjustment, as check points are not used in the adjustment process.

  7. After adding GCPs and check points, click Adjust to run the adjustment again to incorporate these points.

Tie point selection in the GCP Manager window

Review adjustment results

Adjustment quality results can be viewed in the GCP Manager window by analyzing the residuals for each GCP. Residuals represents the difference between the measured position and the computed position of a point. They are measured in the units of the project's spatial referencing system. After completing adjustment with GCPs, three new fields—dX, dY, and dZ—are added to the GCP Manager table and display the residuals for each GCP. The quality of the fit between the adjusted block and the map coordinate system can be evaluated using these values. The root mean square error (RMSE) of the residuals can be viewed by expanding the Residual Overview section of the GCP Manager window.

GCP residuals in the GCP Manager window

Additional adjustment statistics are provided in the adjustment report. To generate the report, on the Reality Mapping tab, in the Review group, click Adjustment Report.

Generate Reality mapping products

Once the block adjustment is complete, 2D and 3D imagery products can be generated using the tools in the Product group on the Reality Mapping tab. Multiple products can be generated simultaneously using the Reality Mapping Products wizard or individually by selecting the applicable product tool from the Product group. The types of products to be generated depends on various factors including the sensor, data flight configuration, and scenario type. The flight configuration of the sample dataset is nadir, which is ideal for 2D products such as DSM, true ortho, and DSM mesh.

Note:

In this tutorial, two approaches to generate derived products are outlined below. One approach uses the Multiple Products wizard and the second uses the individually named product wizards listed in the Product group. It is recommended that you follow one or the other workflow approach, since doing both workflows is not required for this tutorial.

Generate multiple products using the Reality Mapping Products wizard

The Reality Mapping Products wizard guides you through the workflow to create one or multiple Reality mapping products in a single process. The products that can be generated using the Reality Mapping Products wizard are DSM, true ortho, DSM mesh, point cloud, and 3D mesh. All generated products are stored in product folders of the same name under the Reality Mapping category in the Catalog pane.

To generate products using the Reality Mapping Products wizard, complete the following steps:

  1. On the Reality Mapping tab, click the Multiple Products button in the Product group.

    The Reality Mapping Products Wizard window appears.

  2. In the Product Generation Settings pane, uncheck the 3D check box.

    In this tutorial, only 2D products are generated.

  3. In the 2D products category, uncheck Digital Terrain Model (DTM) then check Digital Surface Model (DSM).
  4. Click the Shared Advanced Settings button.

    The Advanced Products Settings dialog box appears, where you can define parameters that will impact the Reality mapping products to be generated. For details of the advanced product settings, see Reality Advanced Settings in the wizard.

    The Quality value is automatically set and should not be changed to ensure optimum performance and product quality.

  5. Accept the Pixel Size default settings to generate products at source image resolution.
  6. For Product Boundary, click the Browse button Open folder connection, browse to the file geodatabase in the tutorial data location, select Project_bdry, and click OK.

    It is recommended that you provide a product boundary for the following reasons:

    • Define the proper output extent—When you do not define a product boundary, the application automatically defines an extent based on various dataset parameters that may not match the project extent.
    • Reduce processing time—If the required product extent is smaller than the image collection extent, defining a product boundary reduces the processing duration and automatically clips the output to the boundary extent.

    Advanced settings for multiple products

  7. For Processing Folder, click the Browse button Open folder connection, navigate to a location on disk that has a minimum available storage of 10 times the total size of the images being processed. In this case, 11 GB or more available storage.

    The processing folder stores temporary files generated during Reality processing. It is recommended that the processing folder be located on a fast drive with large available storage space.

  8. Accept all the other default values and click OK.

    The Advanced Products Settings dialog box closes, and you return to the Product Generation Settings pane in the Reality Mapping Products wizard.

  9. Click Next to go to the DSM Settings pane, and ensure that the parameter values match the following:
    • Output TypeMosaic
    • FormatCloud Raster Format
    • CompressionNone
    • ResamplingBilinear
  10. Click Next to go to the True Ortho Settings pane, and ensure that the parameter values match the following:
    • Output TypeMosaic
    • FormatCloud Raster Format
    • CompressionNone
    • ResamplingBilinear
  11. Click Next to go to the DSM Mesh Settings pane, and ensure that the parameter values match the following:
    • FormatSLPK
  12. Click Finish to start the product generation process.

    Once product generation is complete, the DSM and true ortho products are automatically added to the 2D map view. In the Catalog pane, in the Reality Mapping container, DSM is added to the DEM folder, the DSM mesh products are added to the Meshes folder, and the true ortho products are added to the Ortho folder.

You have completed the generation of 2D products including a DSM, true ortho and DSM mesh using the Multiple Products wizard. You can also generate these 2D products using wizards for each specific product, as described below.

Generate a DSM

To generate a DSM using the Reality Mapping Products wizard, complete the steps below.

Note:
Elevation values can be derived when the image collection has a good amount of overlap to form the stereo pairs. Typical image overlap necessary to produce point clouds is 80 percent forward overlap along a flight line and 60 percent overlap between flight lines.

  1. On the Reality Mapping tab, click the DSM button DSM in the Product group.

    The Reality Mapping Products Wizard window appears.

  2. Click Shared Advanced Settings.

    The Advanced Product Settings dialog box appears, where you can define parameters that will impact the Reality mapping products to be generated. For details of the advanced product settings, see Shared Advanced Settings in the wizard.

    The Quality value is automatically set and should not be changed to ensure optimum performance and product quality. However, if you want to generate a reduced resolution product, the Quality value can be lowered. See Shared Advanced Settings for more information about the impact of various quality settings on product generation.

  3. Accept the Pixel Size default values to generate products at source image resolution.
  4. For Product Boundary, select a feature class identifying the output product extent from the drop-down list, or click the Browse button Open folder connection, browse to the file geodatabase in the tutorial data location, select Project_bdry, and click OK.

    It is recommended that you provide a product boundary for the following reasons:

    • Define the proper output extent—When you do not define a product boundary, the application automatically defines an extent based on various dataset parameters that may not match the project extent.
    • Reduce processing time—If the required product extent is smaller than the image collection extent, defining a product boundary reduces the processing duration and automatically clips the output to the boundary extent.

  5. For Processing Folder, click the Browse button , navigate to a location on disk that has a minimum available storage of 10 times the total size of the images being processed. In this case, 11 GB or more available storage.
  6. Accept all the other default values and click OK.

    The Advanced Products Settings dialog box closes and you return to the Products Generation Settings pane in the Reality Mapping Products wizard.

  7. Click Next to go to the DSM Settings pane, and ensure that the parameter values match the following;
    • Output TypeMosaic
    • FormatCloud Raster Format
    • CompressionNone
    • ResamplingBilinear
  8. Click Finish to start the product generation process.

    Once processing is complete, the DSM product is added to the Contents pane, the Data Products category, and the 2D map view. It is also added to the Catalog pane, the Reality Mapping container, and the DEMs folder.

    Generated DSM

Generate a true ortho

One of the options in the Reality mapping Product group is True Ortho. A true ortho is an orthorectified image with no perspective distortion so that above-ground features do not lean and obscure other features. To create a true ortho, a DSM derived from the adjusted block of overlapping images is required. As a result, a DSM will be generated as a part of the true ortho process regardless of whether a DSM was previously selected as a product. The generated true ortho image will be stored in the Orthos folder under the Reality Mapping category in the Catalog pane.

To generate a true ortho using the Reality Mapping Products wizard, complete the following steps:

  1. On the Reality Mapping tab, click the True Ortho button True Ortho in the Product group.

    The Reality Mapping Products Wizard window appears.

  2. Click Shared Advanced Settings.

    The Advanced Product Settings dialog box appears, where you can define parameters that will impact the Reality mapping products that you generate. For a detailed description of the advanced product settings, see Shared Advanced Settings in the wizard.

  3. Complete steps 4 through 6 in the Generate DSM section above.
  4. Click Next to go to the True Ortho Settings pane, and ensure that the parameter values match the following:
    • Output TypeMosaic
    • FormatCloud Raster Format
    • CompressionNone
    • ResamplingBilinear
  5. Click Finish to start the product generation process.

    Once processing is complete, the True Ortho product is added to the Contents pane, the Data Products category, and the 2D map view. It is also added to the Catalog pane, the Reality Mapping container, and the Ortho folder.

    Generated True Ortho

Generate a DSM mesh

A DSM mesh is a 2.5D textured model of the project area where the adjusted images are draped on a triangulated irregular network (TIN) version of the DSM extracted from the overlapping images in the adjusted block. The Reality Mapping Products wizard simplifies the creation of the DSM mesh product by providing a streamlined workflow with preconfigured parameters. To create a DSM mesh product, a DSM derived from the adjusted block of overlapping images is required. As a result, a DSM is generated as a part of the DSM mesh generation process whether a DSM was selected as a separate product.

To generate a DSM mesh using Reality Mapping Products wizard, complete the following steps:

  1. On the Reality Mapping tab, click the DSM Mesh button DSM Mesh in the Product group.

    The Reality Mapping Products Wizard window appears.

  2. Click Shared Advanced Settings.

    The Advanced Product Settings dialog box appears, where you can define parameters that will impact the Reality mapping products that you generate. For a detailed description of the advanced product settings, see Shared Advanced Settings in the wizard.

  3. Complete steps 4 through 7 in the Generate a DSM section above.
  4. Click Next to go to the DSM Mesh Settings pane, and ensure that the parameter values match the following:
    • FormatSLPK
  5. Click Finish to start the product generation process.

    Once processing is complete, the DSM mesh product is added to the Catalog pane, the Reality Mapping container, and the Meshes folder.

  6. To visualize the generated DSM mesh product, right-click the 3D_Mesh.slpk file in the Meshes folder, and from the drop-down menu, click Add to New > Local Scene.

    Generated DSM Mesh product

Summary

In this tutorial, you created a Reality mapping workspace for drone imagery and used tools on the Reality Mapping tab to apply a photogrammetric adjustment with ground control points. You then used tools in the Products group to generate DSM, true ortho, and DSM mesh products. For more information about Reality mapping, see the following topics:

The imagery used in this tutorial was acquired and provided by Esri, Inc.