Basic parameters
- Spark job ID. The unique ID for the Spark job that references this job in the API. This is the
idfield in the configuration file. Required field. - Collection. The Solr collection to transfer or copy to cloud storage. This is the
inputCollectionfield in the configuration file. Required field. - Output location. The name or location (URI) where the Solr collection is being transferred or copied. This is the
outputLocationfield in the configuration file. Required field. - Overwrite output. If this checkbox is selected (set to
true), overwrite any information that currently exists in the Output location with the data in the Collection being transferred or copied. If this checkbox is not selected and data exists in the output collection, the collection is not copied to the output location and the system generates an error. If this checkbox is not selected and data does not exist in the output collection, the collection is copied to the output location. This is theoverwriteOutputfield in the configuration file. Optional field. - Output format. The format for the output transferred or copied to the cloud. Values include
parquet,json, andcsv. This is theoutputFormatfield in the configuration file. Optional field.
Advanced parameters
If you click the Advanced toggle, the following optional fields are displayed in the UI.- Spark Settings. This section lets you enter
parameter name:parameter valueoptions to use for Spark configuration. This is thesparkConfigfield in the configuration file. - Set minimum Spark partitions for input. The number of partitions that Spark sets for the input. For greater parallelism, increase the value in this field. This is the
sparkPartitionsfield in the configuration file. - Read Options. This section lets you enter
parameter name:parameter valueoptions to use when reading input from Solr. This is thereadOptionsfield in the configuration file.