Label | Explanation | Data Type |
Target Point Cloud
| The point cloud that will be classified. | LAS Dataset Layer |
Input Model Definition
| The input Esri model definition file (*.emd) or deep learning package (*.dlpk) that will be used to classify the point cloud. A URL for a deep learning package that is published on ArcGIS Online or ArcGIS Living Atlas can also be used. | File; String |
Target Classification
| The class codes from the trained model that will be used to classify the input point cloud. All classes from the input model will be used by default unless a subset is specified. | String |
Existing Class Code Handling
(Optional) | Specifies how the editable points from the input point cloud will be defined.
| String |
Existing Class Codes
(Optional) | The classes for which points will be edited or have their original class code designation preserved based on the Existing Class Code Handling parameter value. | Long |
Compute statistics (Optional) | Specifies whether statistics will be computed for the .las files referenced by the LAS dataset. Computing statistics provides a spatial index for each .las file, which improves analysis and display performance. Statistics also enhance the filtering and symbology experience by limiting the display of LAS attributes, such as classification codes and return information, to values that are present in the .las file.
| Boolean |
Processing Boundary
| The polygon boundary that defines the subset of points to be processed from the input point cloud. Points outside the boundary features will not be evaluated. | Feature Layer |
Update pyramid
(Optional) | Specifies whether the LAS dataset pyramid will be updated after the class codes are modified.
| Boolean |
Reference Surface
(Optional) | The raster surface that will be used to provide relative height values for each point in the point cloud data. Points that do not overlap with the raster will be omitted from the analysis. | Raster Layer |
Excluded Class Codes
(Optional) | The class codes that will be excluded from processing. Any value in the range of 0 to 255 can be specified. | Long |
Batch Size
(Optional) | The number of point cloud data blocks that will be simultaneously processed during the inferencing operation. Generally, a larger batch size will cause faster processing of the data, but avoid using a batch size that is too large for the resources of the computer. When using the GPU, the available GPU memory is the most common constraint on the batch size the computer can handle. The memory used by a given block depends on the model's block point limit and required point attributes. To find the available GPU memory and for more information about evaluating GPU memory usage, use the NVIDIA SMI command line utility described in the Usages section. For certain architectures, an optimal batch size will be calculated if the batch size is unspecified. When the GPU is used, the optimal batch size will be based on how much memory is consumed by a given block of data and how much GPU memory is freely available when the tool is run. When the CPU is used for inferencing, each block is processed on a CPU thread, and the optimal batch size is calculated to be half of the available, unused CPU threads. | Long |
Derived Output
Label | Explanation | Data Type |
Output Point Cloud | The point cloud that was classified by the deep learning model. | Feature Layer |