v0.9.6
GeoWave Command Line Controls 
Commands
Helpful Commands & Flags
GeoWave supports a few extra commands that can be used for informational purposes to debug or explore command usage.
Debug Flag (--debug)
Use the debug flag to increase the debug logging output by GeoWave on the console to DEBUG. By default, it is set to WARN. This flag must come right after 'geowave' and before any subcommand:
geowave --debug <command> <subcommand> <options...>
Version Flag (--version)
The version flag will output the build arguments that were used to build GeoWave, as well as the version of the GeoWave tools jar you’re using:
geowave --version
Help Command
The help command will show arguments and their defaults. It can be prepended to any GeoWave command. If you use it while also specifying a sub-command and its arguments, that command’s help information will be displayed:
geowave help <command> <subcommand>
Explain Command
The explain command will show a simplified tabular view of the arguments and their current values. Use this to determine what values are being passed to GeoWave. It also shows hidden parameters and their values, if there are any. An example would be additional Accumulo options:
geowave explain config addstore -t accumulo
Config Commands
Commands that affect local configuration only (Required options are designated with an *)
geowave config addindex
NAME
geowave config addindex - Configure an index for usage in GeoWave
SYNOPSIS
geowave config addindex [options] <name>
DESCRIPTION
The geowave config addindex operator will create a local index configuration that can be reused but is not associated with a store until data is ingested.
OPTIONS
-
--d, --default
-
Make this the default index creating stores
-
-
--indexName
-
A custom name can be given to this index. Default name will be the based on configuration parameters.
-
-
-np, --numPartitions
-
The number of partitions. Default partitions will be 1.
-
Default: 1
-
-
-ps, --partitionStrategy
-
The partition strategy to use. Default will be none.
-
Default: NONE
-
Possible Values: [NONE, HASH, ROUND_ROBIN]
-
-
* -t, --type
-
The type of index, such as spatial, or spatial_temporal
-
geowave config addindexgrp
NAME
geowave config addindexgrp - Create an index group for usage in GeoWave
SYNOPSIS
geowave config addindexgrp [options] <name> <comma separated list of indexes>
DESCRIPTION
The geowave config addindexgrp operator will group multiple index configurations together given a name. This acts as a convenience for re-using multiple indices together on ingest.
OPTIONS
There are currently no options for this command
geowave config addstore
NAME
geowave config addstore - Create a store within Geowave
SYNOPSIS
geowave config addstore [options] <name>
DESCRIPTION
The geowave config addstore operator will create a new store in GeoWave.
OPTIONS
-
-d, --default
-
Make this the default store in all operations
-
-
*-t, --type <arg>
-
The type of store, such as accumulo, memory, etc.
-
Required!
-
When -t accummulo option is used, additional options are:
-
--gwNamespace
-
The geowave namespace
-
Default is no namespace
-
-
-
*-i, --instance
-
The Accumulo instance ID
-
Required!
-
-
*-p, --password
-
The password for the user
-
Required!
-
-
*-u, --user
-
A valid Accumulo user ID
-
Required!
-
-
*-z, --zookeeper
-
A comma-separated list of zookeeper servers that an Accumulo instance is using
-
Required!
-
-
-
When -t, --type hbase option is used, additional options are:
-
-d, --default
-
Make this the default index creating stores
-
-
-np, --numPartitions
-
The number of partitions.
-
Default: 1
-
-
-ps, --partitionStrategy
-
The partition strategy to use.
-
Default: NONE
-
Possible Values: [NONE, HASH, ROUND_ROBIN]
-
-
-
geowave config cpindex
NAME
geowave config cpindex - Copy and modify existing index configuration
SYNOPSIS
geowave config cpindex [options] <name> <new name>
DESCRIPTION
The geowave config cpindex operator will copy and modify an existing index configuration. It is possible to override values as you copy, such as cpindex old new --gwNamespace new_namespace.
OPTIONS
-
-d, --default
-
Make this the default index creating stores
-
-
-np, numPartitions
-
The number of partitions.
-
Default: 1
-
-
-ps, --partitionStrategy
-
The partition strategy to use.
-
Default: NONE
-
Possible Values: [NONE, HASH, ROUND ROBIN]
-
geowave config cpstore
NAME
geowave config cpstore - Copy and modify existing store configuration
SYNOPSIS
geowave config cpstore [options] <name> <new name>
DESCRIPTION
The geowave cpstore operator will copy and modify an existing GeoWave store. It is possible to override values as you copy, such as cpstore old new --gwNamespace new_namespace.
OPTIONS
-
-d, --default
-
Makes this the default store in all operations
-
geowave config geoserver
NAME
geowave config geoserver - Create a local configuration for GeoServer
SYNOPSIS
geowave config geoserver [options] <GeoServer URL (for example http://localhost:8080/geoserver or https://localhost:8443/geoserver), or simply host:port and appropriate assumptions are made>
DESCRIPTION
This command will create a local configuration for connecting to GeoServer.
OPTIONS
-
-p, --password
-
GeoServer Password - Can be specified as 'pass:<password>', 'file:<local file containing the password>', 'propfile:<local properties file containing the password>:<property file key>', 'env:<variable containing the pass>', or stdin
-
-
-u, --username
-
GeoServer User
-
-
-ws, --workspace
-
GeoServer Default Workspace
-
SSL CONFIGURATION OPTIONS
-
--sslKeyManagerAlgorithm
-
Specify the algorithm to use for the keystore.
-
-
--sslKeyManagerProvider
-
Specify the key manager factory provider.
-
-
--sslKeyPassword
-
Specify the password to be used to access the server certificate from the specified keystore file.
-
Can be specified as 'pass:<password>', 'file:<local file containing the password>', 'propfile:<local properties file containing the password>:<property file key>', 'env:<variable containing the pass>', or stdin
-
-
--sslKeyStorePassword
-
Specify the password to use to access the keystore file.
-
Can be specified as 'pass:<password>', 'file:<local file containing the password>', 'propfile:<local properties file containing the password>:<property file key>', 'env:<variable containing the pass>', or stdin
-
-
--sslKeyStorePath
-
Specify the absolute path to where the keystore file is located on system. The keystore contains the server certificate to be loaded.
-
-
--sslKeyStoreProvider
-
Specify the name of the keystore provider to be used for the server certificate.
-
-
--sslKeyStoreType
-
The type of keystore file to be used for the server certificate.
-
-
--sslSecurityProtocol
-
Specify the Transport Layer Security (TLS) protocol to use when connecting to the server. By default, the system will use TLS.
-
-
--sslTrustManagerAlgorithm
-
Specify the algorithm to use for the truststore.
-
-
--sslTrustManagerProvider
-
Specify the trust manager factory provider.
-
-
--sslTrustStorePassword
-
Specify the password to use to access the truststore file.
-
Can be specified as 'pass:<password>', 'file:<local file containing the password>', 'propfile:<local properties file containing the password>:<property file key>', 'env:<variable containing the pass>', or stdin
-
-
--sslTrustStorePath
-
Specify the absolute path to where truststore file is located on system.
-
The truststore file is used to validate client certificates.
-
-
--sslTrustStoreProvider
-
Specify the name of the truststore provider to be used for the server certificate.
-
-
--sslTrustStoreType
-
Specify the type of key store used for the truststore, i.e. JKS (Java KeyStore).
-
geowave config list
NAME
geowave config list - List property name within cache
SYNOPSIS
geowave config list [options]
DESCRIPTION
The geowave config list operator will list all properties in local config. -f or --filter will allow you to give a regex to filter the list by (useful regexes may be 'store' or 'index' to isolate properties for one or the other or a particular store/index name to further isolate the list).
OPTIONS
-
-f or --filter <arg>
-
Filter list by a regex
-
geowave config newcryptokey
NAME
geowave config newcryptokey - Generate a new security cryptography key for use with configuration properties
SYNOPSIS
geowave config newcryptokey
DESCRIPTION
This command will generate a new security cryptography key for use with configuration properties. This is primarily used if there is a need to re-encrypt the local configurations based on a new security token, should the old one have been compromised.
OPTIONS
There are currently no options available for this command
geowave config rmindex
NAME
geowave config rmindex - Remove index configuration from Geowave configuration
SYNOPSIS
geowave config rmindex [options] <name>
DESCRIPTION
The geowave config rmindex operator will remove an index configuration from GeoWave configuration
OPTIONS
There are currently no options for this command
geowave config rmindexgrp
NAME
geowave config rmindexgrp - removes an index group from GeoWave configuration
SYNOPSIS
geowave config rmindexgrp [options] <name>
DESCRIPTION
The geowave config rmindexgrp operator will remove an index group from GeoWave configuration
OPTIONS
There are currently no options for this command
geowave config rmstore
NAME
geowave config rmstore - removes an existing store from GeoWave configuration
SYNOPSIS
geowave config rmstore [options] <name>
DESCRIPTION
The geowave config rmstore operator will remove an existing store from GeoWave configuration GeoWave store
OPTIONS
There are currently no options for this command
geowave config set
NAME
geowave config set - will set a valid property name within the cache
SYNOPSIS
geowave config set [options]
DESCRIPTION
The geowave config set operator will set a valid property name within the cache. This can be useful if you want to update a particular property of a index or store.
OPTIONS
There are currently no options for this command
Ingest Commands
Commands that ingest data directly into GeoWave or stage data to be ingested into GeoWave (Required options are designated with an *)
geowave ingest kafkaToGW
NAME
geowave ingest kafkaToGW - Subscribe to a Kafka topic and ingest into GeoWave
SYNOPSIS
geowave ingest kafkaToGW [options] <store name> <comma delimited index/group list>
DESCRIPTION
The geowave ingest kafkaToGW operator will ingest supported files that already exist in HDFS
OPTIONS
-
--autoOffsetReset
-
What to do when there is no initial offset in ZooKeeper or if an offset is out of range:
-
smallest : automatically reset the offset to the smallest offset
-
largest : automatically reset the offset to the largest offset
-
anything else: throw exception to the consumer
-
-
-
--avro.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--avro.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--avro.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--batchSize
-
The data will automatically flush after this number of entries
-
Default: 10000
-
-
--consumerTimeoutMs
-
By default, this value is -1 and a consumer blocks indefinitely if no new message is available for consumption. By setting the value to a positive integer,a timeout exception is thrown to the consumer if no message is available for consumption after the specified timeout value.
-
-
-x, --extension
-
individual or comma-delimited set of file extensions to accept (optional)
-
-
--fetchMessageMaxBytes
-
The number of bytes of messages to attempt to fetch for each topic-partition in each fetch request. These bytes will be read into memory for each partition, so this helps control the memory used by the consumer. The fetch request size must be at least as large as the maximum message size the server allows or else it is possible for the producer to send messages larger than the consumer can fetch.
-
-
-f, --formats
-
Explicitly set the ingest formats by name (or multiple comma-delimited formats), if not set all available ingest formats will be used
-
-
--gdelt.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gdelt.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gdelt.extended
-
A flag to indicate whether extended data format should be used
-
Default: false
-
-
--gdelt.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geolife.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--geolife.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geolife.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geotools-raster.coverage
-
Optional parameter to set the coverage name (default is the file name)
-
-
--geotools-raster.crs
-
A CRS override for the provided raster file
-
-
--geotools-raster.histogram
-
Build a histogram of samples per band on ingest for performing band equalization
-
Default: false
-
-
--geotools-raster.mergeStrategy
-
Optional parameter to choose a tile merge strategy used for mosaic.
-
Default behavior will be
none
. -
Alternatively 'no-data' will mosaic the most recent tile over previous tiles, except where there are no data values.
-
-
--geotools-raster.nodata
-
Optional parameter to set 'no data' values, if 1 value is giving it is applied for each band, if multiple are given then the first totalNoDataValues/totalBands are applied to the first band and so on, so each band can have multiple differing 'no data' values if needed
-
Default: []
-
-
--geotools-raster.pyramid
-
Build an image pyramid on ingest for quick reduced resolution query
-
Default: false
-
-
--geotools-raster.separateBands
-
Optional parameter to separate each band into its own coverage name. By default the coverage name will have '_Bn' appended to it where
n
is the band’s index. -
Default: false
-
-
--geotools-raster.tileSize
-
Optional parameter to set the tile size stored (default is 256)
-
Default: 256
-
-
--geotools-vector.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geotools-vector.data
-
A map of date field names to the date format of the file. Use commas to separate each entry, then the first ':' character will separate the field name from the format. Use '\,' to include a comma in the format. For example: "time:MM:dd:YYYY,time2:YYYY/MM/dd hh:mm:ss" configures fields 'time' and 'time2' as dates with different formats
-
-
--geotools-vector.type
-
Optional parameter that specifies specific type name(s) from the source file
-
Default: []
-
-
--gpx.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gpx.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gpx.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--groupId
-
A string that uniquely identifies the group of consumer processes to which this consumer belongs. By setting the same group id multiple processes indicate that they are all part of the same consumer group.
-
-
* --kafkaprops
-
Properties file containing Kafka properties
-
-
--reconnectOnTimeout
-
This flag will flush when the consumer timeout occurs (based on kafka property 'consumer.timeout.ms') and immediately reconnect
-
Default: false
-
-
--tdrive.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--tdrive.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--tdrive.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--twitter.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--twitter.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--twitter.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-v, --visibility
-
The visibility of the data ingested (optional; default is 'public')
-
-
--zookeeperConnect
-
Specifies the ZooKeeper connection string in the form hostname:port where host and port are the host and port of a ZooKeeper server. To allow connecting through other ZooKeeper nodes when that ZooKeeper machine is down you can also specify multiple hosts in the form hostname1:port1,hostname2:port2,hostname3:port3.
-
geowave ingest listplugins
NAME
geowave ingest listplugins - List supported data store types, index types, and ingest formats
SYNOPSIS
geowave ingest listplugins [options]
DESCRIPTION
This command will list all the data store types, index types, and ingest formats supported by the version of GeoWave being run.
OPTIONS
There are currently no options for this command
geowave ingest localToGW
NAME
geowave ingest localToGW - Ingest supported files in local file system directly, without using HDFS
SYNOPSIS
geowave ingest localToGW [options] <file or directory> <storename> <comma delimited index/group list>
DESCRIPTION
The geowave ingest localToGW operator will run the ingest code (parse to features, load features to geowave) against local file system content.
OPTIONS
-
--avro.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--avro.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--avro.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-x, --extension
-
individual or comma-delimited set of file extensions to accept (optional)
-
-
-f, --formats
-
Explicitly set the ingest formats by name (or multiple comma-delimited formats), if not set all available ingest formats will be used
-
-
--gdelt.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gdelt.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gdelt.extended
-
A flag to indicate whether extended data format should be used
-
Default: false
-
-
--gdelt.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geolife.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--geolife.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geolife.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geotools-raster.coverage
-
Optional parameter to set the coverage name (default is the file name)
-
-
--geotools-raster.crs
-
A CRS override for the provided raster file
-
-
--geotools-raster.histogram
-
Build a histogram of samples per band on ingest for performing band equalization
-
Default: false
-
-
--geotools-raster.mergeStrategy
-
Optional parameter to choose a tile merge strategy used for mosaic.
-
Default behavior will be
none
. Alternatively 'no-data' will mosaic the most recent tile over previous tiles, except where there are no data values. -
Default: none
-
-
--geotools-raster.nodata
-
Optional parameter to set 'no data' values, if 1 value is giving it is applied for each band, if multiple are given then the first totalNoDataValues/totalBands are applied to the first band and so on, so each band can have multiple differing 'no data' values if needed
-
Default: []
-
-
--geotools-raster.pyramid
-
Build an image pyramid on ingest for quick reduced resolution query
-
Default: false
-
-
--geotools-raster.separateBands
-
Optional parameter to separate each band into its own coverage name. By default the coverage name will have '_Bn' appended to it where
n
is the band’s index. -
Default: false
-
-
--geotools-raster.tileSize
-
Optional parameter to set the tile size stored (default is 256)
-
Default: 256
-
-
--geotools-vector.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geotools-vector.data
-
A map of date field names to the date format of the file. Use commas to separate each entry, then the first ':' character will separate the field name from the format. Use '\,' to include a comma in the format. For example: "time:MM:dd:YYYY,time2:YYYY/MM/dd hh:mm:ss" configures fields 'time' and 'time2' as dates with different formats
-
-
--geotools-vector.type
-
Optional parameter that specifies specific type name(s) from the source file
-
Default: []
-
-
--gpx.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gpx.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gpx.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--tdrive.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--tdrive.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--tdrive.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-t, --threads
-
number of threads to use for ingest, default to 1 (optional)
-
Default: 1
-
-
--twitter.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--twitter.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--twitter.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-v, --visibility
-
The visibility of the data ingested (optional; default is 'public')
-
geowave ingest localToHdfs
NAME
geowave ingest localToHdfs - Stage supported files in local file system to HDFS
SYNOPSIS
geowave ingest localToHdfs [options] <file or directory> <hdfs host:port> <path to base directory to write to>
DESCRIPTION
The geowave ingest localToHdfs operator will stage supported files in the local file system to HDFS
OPTIONS
-
--avro.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--avro.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--avro.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-x, --extension
-
individual or comma-delimited set of file extensions to accept (optional)
-
-
-f, --formats
-
Explicitly set the ingest formats by name (or multiple comma-delimited formats), if not set all available ingest formats will be used
-
-
--gdelt.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gdelt.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gdelt.extended
-
A flag to indicate whether extended data format should be used
-
Default: false
-
-
--gdelt.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geolife.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--geolife.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geolife.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geotools-raster.coverage
-
Optional parameter to set the coverage name (default is the file name)
-
-
--geotools-raster.crs
-
A CRS override for the provided raster file
-
-
--geotools-raster.histogram
-
Build a histogram of samples per band on ingest for performing band equalization
-
Default: false
-
-
--geotools-raster.mergeStrategy
-
Optional parameter to choose a tile merge strategy used for mosaic.
-
Default behavior will be
none
. Alternatively 'no-data' will mosaic the most recent tile over previous tiles, except where there are no data values. -
Default: none
-
-
--geotools-raster.nodata
-
Optional parameter to set 'no data' values, if 1 value is giving it is applied for each band, if multiple are given then the first totalNoDataValues/totalBands are applied to the first band and so on, so each band can have multiple differing 'no data' values if needed
-
Default: []
-
-
--geotools-raster.pyramid
-
Build an image pyramid on ingest for quick reduced resolution query
-
Default: false
-
-
--geotools-raster.separateBands
-
Optional parameter to separate each band into its own coverage name. By default the coverage name will have '_Bn' appended to it where
n
is the band’s index. -
Default: false
-
-
--geotools-raster.tileSize
-
Optional parameter to set the tile size stored (default is 256)
-
Default: 256
-
-
--geotools-vector.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geotools-vector.data
-
A map of date field names to the date format of the file. Use commas to separate each entry, then the first ':' character will separate the field name from the format. Use '\,' to include a comma in the format. For example: "time:MM:dd:YYYY,time2:YYYY/MM/dd hh:mm:ss" configures fields 'time' and 'time2' as dates with different formats
-
-
--geotools-vector.type
-
Optional parameter that specifies specific type name(s) from the source file
-
Default: []
-
-
--gpx.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gpx.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gpx.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--tdrive.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--tdrive.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--tdrive.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--twitter.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--twitter.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--twitter.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
geowave ingest localToKafka
NAME
geowave ingest localToKafka - Stage supported files in local file system to a Kafka topic
SYNOPSIS
geowave ingest localToKafka [options] <file or directory>
DESCRIPTION
The geowave ingest localToKafka operator will stage supported files in the local file system to a Kafka topic
OPTIONS
-
--avro.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--avro.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--avro.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-x, --extension
-
individual or comma-delimited set of file extensions to accept (optional)
-
-
-f, --formats
-
Explicitly set the ingest formats by name (or multiple comma-delimited formats), if not set all available ingest formats will be used
-
-
--gdelt.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gdelt.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gdelt.extended
-
A flag to indicate whether extended data format should be used
-
Default: false
-
-
--gdelt.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geolife.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--geolife.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geolife.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geotools-raster.coverage
-
Optional parameter to set the coverage name (default is the file name)
-
-
--geotools-raster.crs
-
A CRS override for the provided raster file
-
-
--geotools-raster.histogram
-
Build a histogram of samples per band on ingest for performing band equalization
-
Default: false
-
-
--geotools-raster.mergeStrategy
-
Optional parameter to choose a tile merge strategy used for mosaic.
-
Default behavior will be
none
. Alternatively 'no-data' will mosaic the most recent tile over previous tiles, except where there are no data values. -
Default: none
-
-
--geotools-raster.nodata
-
Optional parameter to set 'no data' values, if 1 value is giving it is applied for each band, if multiple are given then the first totalNoDataValues/totalBands are applied to the first band and so on, so each band can have multiple differing 'no data' values if needed
-
Default: []
-
-
--geotools-raster.pyramid
-
Build an image pyramid on ingest for quick reduced resolution query
-
Default: false
-
-
--geotools-raster.separateBands
-
Optional parameter to separate each band into its own coverage name. By default the coverage name will have '_Bn' appended to it where
n
is the band’s index. -
Default: false
-
-
--geotools-raster.tileSize
-
Optional parameter to set the tile size stored (default is 256)
-
Default: 256
-
-
--geotools-vector.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geotools-vector.data
-
A map of date field names to the date format of the file. Use commas to separate each entry, then the first ':' character will separate the field name from the format. Use '\,' to include a comma in the format. For example: "time:MM:dd:YYYY,time2:YYYY/MM/dd hh:mm:ss" configures fields 'time' and 'time2' as dates with different formats
-
-
--geotools-vector.type
-
Optional parameter that specifies specific type name(s) from the source file
-
Default: []
-
-
--gpx.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gpx.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gpx.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
* --kafkaprops
-
Properties file containing Kafka properties
-
-
--metadataBrokerList
-
This is for bootstrapping and the producer will only use it for getting metadata (topics, partitions and replicas). The socket connections for sending the actual data will be established based on the broker information returned in the metadata. The format is host1:port1,host2:port2, and the list can be a subset of brokers or a VIP pointing to a subset of brokers.
-
-
--producerType
-
This parameter specifies whether the messages are sent asynchronously in a background thread. Valid values are (1) async for asynchronous send and (2) sync for synchronous send. By setting the producer to async we allow batching together of requests (which is great for throughput) but open the possibility of a failure of the client machine dropping unsent data.
-
-
--requestRequiredAcks
-
This value controls when a produce request is considered completed. Specifically, how many other brokers must have committed the data to their log and acknowledged this to the leader?
-
-
--retryBackoffMs
-
The amount of time to wait before attempting to retry a failed produce request to a given topic partition. This avoids repeated sending-and-failing in a tight loop.
-
-
--serializerClass
-
The serializer class for messages. The default encoder takes a byte[] and returns the same byte[].
-
-
--tdrive.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--tdrive.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--tdrive.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--twitter.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--twitter.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--twitter.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
geowave ingest localToMrGW
NAME
geowave ingest localToMrGW - Copy supported files from local file system to HDFS and ingest from HDFS
SYNOPSIS
geowave ingest localToMrGW [options] <file or directory> <hdfs host:port> <path to base directory to write to> <store name> <comma delimited index/group list>
DESCRIPTION
The geowave ingest localToMrGW operator will Copy supported files from local file system to HDFS and ingest from HDFS
OPTIONS
-
--avro.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--avro.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--avro.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-x, --extension
-
individual or comma-delimited set of file extensions to accept (optional)
-
-
-f, --formats
-
Explicitly set the ingest formats by name (or multiple comma-delimited formats), if not set all available ingest formats will be used
-
-
--gdelt.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gdelt.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gdelt.extended
-
A flag to indicate whether extended data format should be used
-
Default: false
-
-
--gdelt.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geolife.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--geolife.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geolife.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geotools-raster.coverage
-
Optional parameter to set the coverage name (default is the file name)
-
-
--geotools-raster.crs
-
A CRS override for the provided raster file
-
-
--geotools-raster.histogram
-
Build a histogram of samples per band on ingest for performing band equalization
-
Default: false
-
-
--geotools-raster.mergeStrategy
-
Optional parameter to choose a tile merge strategy used for mosaic.
-
Default behavior will be
none
. Alternatively 'no-data' will mosaic the most recent tile over previous tiles, except where there are no data values. -
Default: none
-
-
--geotools-raster.nodata
-
Optional parameter to set 'no data' values, if 1 value is giving it is applied for each band, if multiple are given then the first totalNoDataValues/totalBands are applied to the first band and so on, so each band can have multiple differing 'no data' values if needed
-
Default: []
-
-
--geotools-raster.pyramid
-
Build an image pyramid on ingest for quick reduced resolution query
-
Default: false
-
-
--geotools-raster.separateBands
-
Optional parameter to separate each band into its own coverage name. By default the coverage name will have '_Bn' appended to it where
n
is the band’s index. -
Default: false
-
-
--geotools-raster.tileSize
-
Optional parameter to set the tile size stored (default is 256)
-
Default: 256
-
-
--geotools-vector.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geotools-vector.data
-
A map of date field names to the date format of the file. Use commas to separate each entry, then the first ':' character will separate the field name from the format. Use '\,' to include a comma in the format. For example: "time:MM:dd:YYYY,time2:YYYY/MM/dd hh:mm:ss" configures fields 'time' and 'time2' as dates with different formats
-
-
--geotools-vector.type
-
Optional parameter that specifies specific type name(s) from the source file
-
Default: []
-
-
--gpx.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gpx.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gpx.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--jobtracker
-
Hadoop job tracker hostname and port in the format hostname:port
-
-
--resourceman
-
Yarn resource manager hostname and port in the format hostname:port
-
-
--tdrive.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--tdrive.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--tdrive.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--twitter.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--twitter.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--twitter.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-v, --visibility
-
The visibility of the data ingested (optional; default is 'public')
-
geowave ingest mrToGW
NAME
geowave ingest mrToGW - Ingest supported files that already exist in HDFS
SYNOPSIS
geowave ingest mrToGW [options] <hdfs host:port> <path to base directory to write to> <store name> <comma delimited index/group list>
DESCRIPTION
The geowave ingest mrToGW operator will ingest supported files that already exist in HDFS
OPTIONS
-
--avro.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--avro.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--avro.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-x, --extension
-
individual or comma-delimited set of file extensions to accept (optional)
-
-
-f, --formats
-
Explicitly set the ingest formats by name (or multiple comma-delimited formats), if not set all available ingest formats will be used
-
-
--gdelt.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gdelt.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gdelt.extended
-
A flag to indicate whether extended data format should be used
-
Default: false
-
-
--gdelt.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geolife.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--geolife.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geolife.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--geotools-raster.coverage
-
Optional parameter to set the coverage name (default is the file name)
-
-
--geotools-raster.crs
-
A CRS override for the provided raster file
-
-
--geotools-raster.histogram
-
Build a histogram of samples per band on ingest for performing band equalization
-
Default: false
-
-
--geotools-raster.mergeStrategy
-
Optional parameter to choose a tile merge strategy used for mosaic.
-
Default behavior will be
none
. Alternatively 'no-data' will mosaic the most recent tile over previous tiles, except where there are no data values. -
Default: none
-
-
--geotools-raster.nodata
-
Optional parameter to set 'no data' values, if 1 value is giving it is applied for each band, if multiple are given then the first totalNoDataValues/totalBands are applied to the first band and so on, so each band can have multiple differing 'no data' values if needed
-
Default: []
-
-
--geotools-raster.pyramid
-
Build an image pyramid on ingest for quick reduced resolution query
-
Default: false
-
-
--geotools-raster.separateBands
-
Optional parameter to separate each band into its own coverage name. By default the coverage name will have '_Bn' appended to it where
n
is the band’s index. -
Default: false
-
-
--geotools-raster.tileSize
-
Optional parameter to set the tile size stored (default is 256)
-
Default: 256
-
-
--geotools-vector.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--geotools-vector.data
-
A map of date field names to the date format of the file. Use commas to separate each entry, then the first ':' character will separate the field name from the format. Use '\,' to include a comma in the format. For example: "time:MM:dd:YYYY,time2:YYYY/MM/dd hh:mm:ss" configures fields 'time' and 'time2' as dates with different formats
-
-
--geotools-vector.type
-
Optional parameter that specifies specific type name(s) from the source file
-
Default: []
-
-
--gpx.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--gpx.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--gpx.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--jobtracker
-
Hadoop job tracker hostname and port in the format hostname:port
-
-
--resourceman
-
Yarn resource manager hostname and port in the format hostname:port
-
-
--tdrive.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--tdrive.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--tdrive.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
--twitter.avro
-
A flag to indicate whether avro feature serialization should be used
-
Default: false
-
-
--twitter.cql
-
A CQL filter, only data matching this filter will be ingested
-
Default: <empty string>
-
-
--twitter.typename
-
A comma-delimitted set of typenames to ingest, feature types matching the specified typenames will be ingested (optional, by default all types will be ingested)
-
-
-v, --visibility
-
The visibility of the data ingested (optional; default is 'public')
-
Remote Commands
Operations to manage a remote store (Required options are designated with an *)
geowave remote calcstat
NAME
geowave remote calcstat - Calculate a specific statistic in the remote store, given adapter ID and statistic ID
SYNOPSIS
geowave remote calcstat [options] <store name> <adapterId> <statId>
DESCRIPTION
The geowave remote calcstat operator will calculate a specific statistic in the remote store, given adapter ID and statistic ID.
OPTIONS
-
--auth
-
The authorizations used for the statistics calculation as a subset of the accumulo user authorization; by default all authorizations are used.
-
-
--json
-
Output in JSON format.
-
Default: false
-
geowave remote clear
NAME
geowave remote clear - Clear ALL data from a GeoWave store and delete tables
SYNOPSIS
geowave remote clear [options] <store name>
DESCRIPTION
The geowave remote clear operator will clear ALL data from a GeoWave store and delete tables
OPTIONS
There are currently no options for this command
geowave remote copy
NAME
geowave remote copy - Copy a data store
SYNOPSIS
geowave remote copy [options] <input store name> <output store name>
DESCRIPTION
This command will allow a user to copy a data store
OPTIONS
-
* --hdfsHostPort
-
The hdfs host port
-
-
* --jobSubmissionHostPort
-
The job submission tracker
-
-
--maxSplits
-
The max partitions for the input data
-
-
--minSplits
-
The min partitions for the input data
-
-
--numReducers
-
Number of threads writing at a time (default: 8)
-
Default: 8
-
geowave remote listadapter
NAME
geowave remote listadapter - display all adapters in this remote store
SYNOPSIS
geowave remote listadapter [options] <store name>
DESCRIPTION
The geowave remote listadapter operator will display all adapters in this remote store.
OPTIONS
There are currently no options for this command
geowave remote listindex
NAME
geowave remote listindex - Display all indices in this remote store
SYNOPSIS
geowave remote listindex [options] <store name>
DESCRIPTION
The geowave remote listindex operator will display all indices in a specific remote store
OPTIONS
There are currently no options for this command
geowave remote liststats
NAME
geowave remote liststats - Print statistics of an existing GeoWave dataset to standard output
SYNOPSIS
geowave remote liststats [options] <store name> [<adapter name>]
DESCRIPTION
The geowave remote liststats operator will print statistics of an existing GeoWave dataset to standard output
OPTIONS
-
--auth
-
The authorizations used for the statistics calculation as a subset of the accumulo user authorization; by default all authorizations are used.
-
-
--json
-
Output in JSON format.
-
Default: false
-
geowave remote mergedata
NAME
geowave remote mergedata - Merge all rows for a given adapter and index
SYNOPSIS
geowave remote mergedata [options] <storename> <indexname>
DESCRIPTION
This command will allow a user to merge all rows for a given adapter and index
OPTIONS
There are currently no options for this command
geowave remote recalcstats
NAME
geowave remote recalcstats - Calculate the statistics of an existing GeoWave dataset
SYNOPSIS
geowave remote recalcstats [options] <store name> [<adapter name>]
DESCRIPTION
The geowave remote recalcstats operator will calculate the statistics of an existing GeoWave dataset
OPTIONS
-
--auth
-
The authorizations used for the statistics calculation as a subset of the accumulo user authorization; by default all authorizations are used.
-
-
--json
-
Output in JSON format.
-
Default: false
-
geowave remote rmstat
NAME
geowave remote rmstat - Remove a statistic from the remote store
SYNOPSIS
geowave remote rmstat [options] <store name> <adapterId> <statId>
DESCRIPTION
The geowave remote rmstat operator will Remove a statistic from the remote store. You will be prompted with "Are you sure?"
OPTIONS
-
--auth
-
The authorizations used for the statistics calculation as a subset of the accumulo user authorization; by default all authorizations are used.
-
-
--json
-
Output in JSON format.
-
Default: false
-
geowave remote version
NAME
geowave remote version - Get the version of GeoWave running on the instance of a remote datastore
SYNOPSIS
geowave remote version [options] <storename>
DESCRIPTION
This command will return the version of GeoWave running on the instance of a remote datastore.
OPTIONS
There are currently no options for this command
Analytic Commands
Commands that run mapreduce or spark processing to enhance an existing GeoWave dataset (Required options are designated with an *)
The commands below can also be run as a yarn or hadoop api commands (i.e. mapreduce) For instance, if running the analytic using yarn:
|
geowave analytic dbscan
NAME
geowave analytic dbscan - Density Based Scanner
SYNOPSIS
geowave analytic dbscan [options] <storename>
DESCRIPTION
The geowave analytic dbscan operator will run a density based scanner analytic on GeoWave data
EXAMPLE
yarn jar geowave-tools.jar analytic dbscan -cmi 5 -cms 10 -emn 2 -emx 6 -pmd 1000 -orc 4 -hdfs localhost:53000 -jobtracker localhost:8032 -hdfsbase /user/rwgdrummer --query.adapters gpxpoint my_store
Run through 5 max iterations (cmi), with max distance between points as 10 meters (cms), min hdfs input split is 2 (emn), max hdfs input split is 6 (emx), max search distance is 1000 meters (pmd), reducer count is 4 (orc), the hdfs ipc port is localhost:53000 (hdfs), the yarn job tracker is at localhost:8032 (-jobtracker), the temporary files needed by this job are stored in hdfs:/host:port//user/rwgdrummer (hdfsbase), and the data executed against DBSCAN is 'gpxpoint' (query.adapters). The accumulo connection parameters are loaded from my_store.
EXECUTION
DBSCAN uses GeoWaveInputFormat to load data from GeoWave into HDFS. You can use the extract query parameter to limit the records used in the analytic.
It iteratively calls Nearest Neighbor to execute a sequence of concave hulls. The hulls are saved into sequence files written to a temporary HDFS directory, and then read in again for the next DBSCAN iteration.
After completion, the data is written back from HDFS to Accumulo using a job called the "input load runner".
OPTIONS
-
* -cmi, --clusteringMaxIterations
-
Maximum number of iterations when finding optimal clusters
-
-
* -cms, --clusteringMinimumSize
-
Minimum Cluster Size
-
-
--cdf, --commonDistanceFunctionClass
-
Distance Function Class implements mil.nga.giat.geowave.analytics.distance.DistanceFn
-
-
* -emx, --extractMaxInputSplit
-
Maximum hdfs input split size
-
-
* -emn, --extractMinInputSplit
-
Minimum hdfs input split size
-
-
-eq, --extractQuery
-
Query
-
-
-b, --globalBatchId
-
Batch ID
-
-
-hdt, --hullDataTypeId
-
Data Type ID for a centroid item
-
-
-hpe, --hullProjectionClass
-
Class to project on to 2D space. Implements mil.nga.giat.geowave.analytics.tools.Projection
-
-
-ifc, --inputFormatClass
-
Input Format Class
-
-
-conf, --mapReduceConfigFile
-
MapReduce Configuration
-
-
* -hdfsbase, --mapReduceHdfsBaseDir
-
Fully qualified path to the base directory in hdfs
-
-
* -hdfs, --mapReduceHdfsHostPort
-
HDFS hostname and port in the format hostname:port
-
-
-jobtracker, --mapReduceJobtrackerHostPort
-
[REQUIRED (or resourceman)] Hadoop job tracker hostname and port in the format hostname:port
-
-
-resourceman, --mapReduceYarnResourceManager
-
[REQUIRED (or jobtracker)] Yarn resource manager hostname and port in the format hostname:port
-
-
-ons, --outputDataNamespaceUri
-
Output namespace for objects that will be written to GeoWave
-
-
-odt, --outputDataTypeId
-
Output Data ID assigned to objects that will be written to GeoWave
-
-
-oop, --outputHdfsOutputPath
-
Output HDFS File Path
-
-
-oid, --outputIndexId
-
Output Index ID for objects that will be written to GeoWave
-
-
-ofc, --outputOutputFormat
-
Output Format Class
-
-
-orc, --outputReducerCount
-
Number of Reducers For Output
-
-
-pdt, --partitionDistanceThresholds
-
Comma separated list of distance thresholds, per dimension
-
-
-pdu, --partitionGeometricDistanceUnit
-
Geometric distance unit (m=meters,km=kilometers, see symbols for javax.units.BaseUnit)
-
-
* -pmd, --partitionMaxDistance
-
Maximum Partition Distance
-
-
-pms, --partitionMaxMemberSelection
-
Maximum number of members selected from a partition
-
-
-pdr, --partitionPartitionDecreaseRate
-
Rate of decrease for precision(within (0,1])
-
-
-pp, --partitionPartitionPrecision
-
Partition Precision
-
-
-pc, --partitionPartitionerClass
-
Index Identifier for Centroids
-
-
-psp, --partitionSecondaryPartitionerClass
-
Perform secondary partitioning with the provided class
-
-
* --query.adapters
-
The comma-separated list of data adapters to query; by default all adapters are used.
-
-
--query.auth
-
The comma-separated list of authorizations used during extract; by default all authorizations are used.
-
-
--query.index
-
The specific index to query; by default one is chosen for each adapter.
-
geowave analytic kde
NAME
geowave analytic kde - Kernel Density Estimate
SYNOPSIS
geowave analytic kde [options] <input storename> <output storename>
DESCRIPTION
The geowave analytic kde operator will run a Kernel Density Estimate analytic on GeoWave data
OPTIONS
-
* --coverageName
-
The coverage name
-
-
--cqlFilter
-
An optional CQL filter applied to the input data
-
-
* --featureType
-
The name of the feature type to run a KDE on
-
-
* --hdfsHostPort
-
The hdfs host port
-
-
--indexId
-
An optional index ID to filter the input data
-
-
* --jobSubmissionHostPort
-
The job submission tracker
-
-
* --maxLevel
-
The max level to run a KDE at
-
-
—maxSplits
-
The max partitions for the input data
-
-
* --minLevel
-
The min level to run a KDE at
-
-
--minSplits
-
The min partitions for the input data
-
-
--tileSize
-
The tile size
-
Default: 1
-
geowave analytic kmeansjump
NAME
geowave analytic kmeansjump - KMeans Clustering using Jump Method
SYNOPSIS
geowave analytic kmeansjump [options] <storename>
DESCRIPTION
The geowave analytic kmeansjump operator will execute a KMeans Clustering analytic using a Jump Method
EXAMPLE
yarn jar geowave-tools.jar analytic kmeansjump -cmi 15 -zl 1 -emx 4000 -emn 100 -hdfsbase /usr/rwgdrummer/temp_dir_kmeans -hdfs localhost:53000 -jobtracker localhost:8032 --query.adapters hail -jkp 3 -jrc 4,8 my_store
The min clustering iterations is 15 (cmi), the zoomlevel is 1 (zl), the max hdfs input split is 4000 (emx), the min hdfs input split is 100 (emn), the temporary files needed by this job are stored in hdfs:/host:port/user/rwgdrummer/temp_dir_kmeans (hdfsbase), the hdfs ipc port is localhost:53000 (hdfs), the yarn job tracker is at localhost:8032 (-jobtracker), the data executed against is 'hail' (query.adapters), the min k for kmeans parallel sampling is 3 (jkp), the comma separated range of centroids is 4,8 (jrc). The accumulo connection parameters are loaded from my_store.
EXECUTION
KMeansJump uses most of the same parameters from KMeansParallel. It tries every k value given (-jrc) to find the value with least entropy. The other value, jkp, will specify which k values should use kmeans parallel for sampling versus a single sampler (which uses a random sample). For instance, if you specify 4,8 for jrc and 6 for jkp, then k=4,5 will use the kmeansparallel sampler, while 6,7,8 will use the single sampler.
KMeansJump executes by executing several iterations, running the sampler (described above, which also calls the normal k-means algorithm to determine centroids) and then executing a KMeans distortion job, which calculates the entropy of the calculated centroids.
Look at the "EXECUTION" documentation for kmeansparallel operation for discussion of output, tolerance and performance variables.
OPTIONS
-
-cce, --centroidExtractorClass
-
Centroid Exractor Class implements mil.nga.giat.geowave.analytics.extract.CentroidExtractor
-
-
-cid, --centroidIndexId
-
Index Identifier for Centroids
-
-
-cfc, --centroidWrapperFactoryClass
-
A factory class that implements mil.nga.giat.geowave.analytics.tools.AnalyticItemWrapperFactory
-
-
-czl, --centroidZoomLevel
-
Zoom Level Number
-
-
-cct, --clusteringConverganceTolerance
-
Convergence Tolerance
-
-
* -cmi, --clusteringMaxIterations
-
Maximum number of iterations when finding optimal clusters
-
-
-crc, --clusteringMaxReducerCount
-
Maximum Clustering Reducer Count
-
-
* -zl, --clusteringZoomLevels
-
Number of Zoom Levels to Process
-
-
-dde, --commonDimensionExtractClass
-
Dimension Extractor Class implements mil.nga.giat.geowave.analytics.extract.DimensionExtractor
-
-
-cdf, --commonDistanceFunctionClass
-
Distance Function Class implements mil.nga.giat.geowave.analytics.distance.DistanceFn
-
-
-ens, --extractDataNamespaceUri
-
Output Data Namespace URI
-
-
-ede, --extractDimensionExtractClass
-
Class to extract dimensions into a simple feature output
-
-
* -emx, --extractMaxInputSplit
-
Maximum hdfs input split size
-
-
* -emn, --extractMinInputSplit
-
Minimum hdfs input split size
-
-
-eot, --extractOutputDataTypeId
-
Output Data Type ID
-
-
-eq, --extractQuery
-
Query
-
-
-erc, --extractReducerCount
-
Number of Reducers For initial data extraction and de-duplication
-
-
-b, --globalBatchId
-
Batch ID
-
-
-pb, --globalParentBatchId
-
Batch ID
-
-
-hns, --hullDataNamespaceUri
-
Data Type Namespace for a centroid item
-
-
-hdt, --hullDataTypeId
-
Data Type ID for a centroid item
-
-
-hid, --hullIndexId
-
Index Identifier for Centroids
-
-
-hpe, --hullProjectionClass
-
Class to project on to 2D space. Implements mil.nga.giat.geowave.analytics.tools.Projection
-
-
-hrc, --hullReducerCount
-
Centroid Reducer Count
-
-
-hfc, --hullWrapperFactoryClass
-
Class to create analytic item to capture hulls. Implements mil.nga.giat.geowave.analytics.tools.AnalyticItemWrapperFactory
-
-
-ifc, --inputFormatClass
-
Input Format Class
-
-
* -jkp, --jumpKplusplusMin
-
The minimum k when K means ++ takes over sampling.
-
-
* -jrc, --jumpRangeOfCentroids
-
Comma-separated range of centroids (e.g. 2,100)
-
-
-conf, --mapReduceConfigFile
-
MapReduce Configuration
-
-
* -hdfsbase, --mapReduceHdfsBaseDir
-
Fully qualified path to the base directory in hdfs
-
-
* -hdfs, --mapReduceHdfsHostPort
-
HDFS hostname and port in the format hostname:port
-
-
-jobtracker, --mapReduceJobtrackerHostPort
-
[REQUIRED (or resourceman)] Hadoop job tracker hostname and port in the format hostname:port
-
-
-resourceman, --mapReduceYarnResourceManager
-
[REQUIRED (or jobtracker)] Yarn resource manager hostname and port in the format hostname:port
-
-
-ofc, --outputOutputFormat
-
Output Format Class
-
-
-orc, --outputReducerCount
-
Number of Reducers For Output
-
-
* --query.adapters
-
The comma-separated list of data adapters to query; by default all adapters are used.
-
-
--query.auth
-
The comma-separated list of authorizations used during extract; by default all authorizations are used.
-
-
--query.index
-
The specific index to query; by default one is chosen for each adapter.
-
geowave analytic kmeansparallel
NAME
geowave analytic kmeansparallel - KMeans Parallel Clustering
SYNOPSIS
geowave analytic kmeansparallel [options] <storename>
DESCRIPTION
The geowave analytic kmeansparallel operator will execute a KMeans Parallel Clustering analytic
EXAMPLE
yarn jar geowave-tools.jar analytic kmeansparallel -cmi 15 -zl 1 -emx 4000 -emn 100 -hdfsbase /usr/rwgdrummer/temp_dir_kmeans
-hdfs localhost:53000 -jobtracker localhost:8032 --query.adapters hail -sms 4 -sxs 8 -ssi 10 my_store
The min clustering iterations is 15 (cmi), the zoomlevel is 1 (zl), the max hdfs input split is 4000 (emx), the min hdfs input split is 100 (emn), the temporary files needed by this job are stored in hdfs:/host:port/user/rwgdrummer/temp_dir_kmeans (hdfsbase), the hdfs ipc port is localhost:53000 (hdfs), the yarn job tracker is at localhost:8032 (-jobtracker), the data executed against is 'hail' (query.adapters), the min sample size is 4 (sms, which is kmin), the max sample size is 8 (which is kmax), the minimum number of sampling iterations is 10. The accumulo connection parameters are loaded from my_store.
EXECUTION
KMeansParallel tries to identify the optimal k (sms, smx) for a set of zoom levels (1 → zl). When zoom level is 1, it will perform a normal kmeans and find k clusters. If zoomlevel is 2 or higher, it will take each cluster found, and then try to create sub-clusters (bounded by that cluster), identifying a new optimal k for that sub-cluster. As such, without powerful infrastucture, this approach could take a significant amount of time to complete with zoom levels higher than 1.
KMeansParallel executes by first executing an extraction and de-duplication on data received via GeoWaveInputFormat. The data is copied to HDFS for faster processing. The K-Sampler job is used to pick sample centroid points. These centroids are then assigned a cost, and then weak centroids are stripped before the K-Sampler is executed again. This process iterates several times, before the best centroid locations are found, which are fed into the real K-Means algorithm as initial guesses. K-Means iterates until the tolerance is reached (-cct, which defaults to 0.0001) or the max iterations is met (-cmi).
After execution, KMeansParallel writes the centroids to an output data type (-eot, defaults to centroid), and then creates an informational set of convex hulls which you can plot in GeoServer to visually identify cluster groups (-hdt, defaults to convex_hull).
For tuning performance, you can set the number of reducers used in each step. Extraction/Dedupe reducer count is -crc, Clustering reducer count is -erc, Convex Hull reducer count is -hrc, and Output reducer count is -orc).
If you would like to run the algorithm multiple times, it may be useful to set the batch id (-b), which can be used to distinguish between multiple batches (runs).
OPTIONS
-
-cce, --centroidExtractorClass
-
Centroid Exractor Class implements mil.nga.giat.geowave.analytics.extract.CentroidExtractor
-
-
-cid, --centroidIndexId
-
Index Identifier for Centroids
-
-
-cfc, --centroidWrapperFactoryClass
-
A factory class that implements mil.nga.giat.geowave.analytics.tools.AnalyticItemWrapperFactory
-
-
-czl, --centroidZoomLevel
-
Zoom Level Number
-
-
-cct, --clusteringConverganceTolerance
-
Convergence Tolerance
-
-
* -cmi, --clusteringMaxIterations
-
Maximum number of iterations when finding optimal clusters
-
-
-crc, --clusteringMaxReducerCount
-
Maximum Clustering Reducer Count
-
-
* -zl, --clusteringZoomLevels
-
Number of Zoom Levels to Process
-
-
-dde, --commonDimensionExtractClass
-
Dimension Extractor Class implements mil.nga.giat.geowave.analytics.extract.DimensionExtractor
-
-
-cdf, --commonDistanceFunctionClass
-
Distance Function Class implements mil.nga.giat.geowave.analytics.distance.DistanceFn
-
-
-ens, --extractDataNamespaceUri
-
Output Data Namespace URI
-
-
-ede, --extractDimensionExtractClass
-
Class to extract dimensions into a simple feature output
-
-
* -emx, --extractMaxInputSplit
-
Maximum hdfs input split size
-
-
* -emn, --extractMinInputSplit
-
Minimum hdfs input split size
-
-
-eot, --extractOutputDataTypeId
-
Output Data Type ID
-
-
-eq, --extractQuery
-
Query
-
-
-erc, --extractReducerCount
-
Number of Reducers For initial data extraction and de-duplication
-
-
-b, --globalBatchId
-
Batch ID
-
-
-pb, --globalParentBatchId
-
Batch ID
-
-
-hns, --hullDataNamespaceUri
-
Data Type Namespace for a centroid item
-
-
-hdt, --hullDataTypeId
-
Data Type ID for a centroid item
-
-
-hid, --hullIndexId
-
Index Identifier for Centroids
-
-
-hpe, --hullProjectionClass
-
Class to project on to 2D space. Implements mil.nga.giat.geowave.analytics.tools.Projection
-
-
-hrc, --hullReducerCount
-
Centroid Reducer Count
-
-
-hfc, --hullWrapperFactoryClass
-
Class to create analytic item to capture hulls. Implements mil.nga.giat.geowave.analytics.tools.AnalyticItemWrapperFactory
-
-
-ifc, --inputFormatClass
-
Input Format Class
-
-
-conf, --mapReduceConfigFile
-
MapReduce Configuration
-
-
* -hdfsbase, --mapReduceHdfsBaseDir
-
Fully qualified path to the base directory in hdfs
-
-
* -hdfs, --mapReduceHdfsHostPort
-
HDFS hostname and port in the format hostname:port
-
-
-jobtracker, --mapReduceJobtrackerHostPort
-
[REQUIRED (or resourceman)] Hadoop job tracker hostname and port in the format hostname:port
-
-
-resourceman, --mapReduceYarnResourceManager
-
[REQUIRED (or jobtracker)] Yarn resource manager hostname and port in the format hostname:port
-
-
-ofc, --outputOutputFormat
-
Output Format Class
-
-
-orc, --outputReducerCount
-
Number of Reducers For Output
-
-
* --query.adapters
-
The comma-separated list of data adapters to query; by default all adapters are used.
-
-
--query.auth
-
The comma-separated list of authorizations used during extract; by default all authorizations are used.
-
-
--query.index
-
The specific index to query; by default one is chosen for each adapter.
-
-
* -sxs, --sampleMaxSampleSize
-
Max Sample Size
-
-
* -sms, --sampleMinSampleSize
-
Minimum Sample Size
-
-
* -ssi, --sampleSampleIterations
-
Minimum number of sample iterations
-
geowave analytic nn
NAME
geowave analytic nn - Nearest Neighbors
SYNOPSIS
geowave analytic nn [options] <storename>
DESCRIPTION
The geowave analytic nn operator will execute a Nearest Neighbors analytic. Analytic 'nn' is similar to DBScan, with less arguments. 'nn' just dumps all near neighbors for every feature id in a list of pairs. Most developers will want to extend the framework to add their own extensions.
EXAMPLE
yarn jar geowave-tools.jar analytic nn -emn 2 -emx 6 -pmd 1000 -oop /user/rwgdrummer_out -orc 4 -hdfs localhost:53000 -jobtracker localhost:8032 -hdfsbase /user/rwgdrummer --query.adapters gpxpoint my_store
The min hdfs input split is 2 (emn), max hdfs input split is 6 (emx), max search distance is 1000 meters (pmd), the sequence file output directory is hdfs://host:port/user/rwgdrummer_out, reducer count is 4 (orc), the hdfs ipc port is localhost:53000 (hdfs), the yarn job tracker is at localhost:8032 (-jobtracker), the temporary files needed by this job are stored in hdfs:/host:port//user/rwgdrummer (hdfsbase), and the data executed against is 'gpxpoint' (query.adapters). The accumulo connection parameters are loaded from my_store.
EXECUTION
To execute nearest neighbor search in GeoWave, we use the concept of a "partitioner" to partition all data on the hilbert curve into square segments for the purposes of parallelizing the search.
The default partitioner will multiply this value by 2 and use that for the actual partition sizes. Because of this, the terminology is a bit confusing, but the "pmd" option is actually the most important variable here, describing the max distance for a point to be considered a neighbor to another point.
OPTIONS
-
-cdf, --commonDistanceFunctionClass
-
Distance Function Class implements mil.nga.giat.geowave.analytics.distance.DistanceFn
-
-
* -emx, --extractMaxInputSplit
-
Maximum hdfs input split size
-
-
* -emn, --extractMinInputSplit
-
Minimum hdfs input split size
-
-
-eq, --extractQuery
-
Query
-
-
-ifc, --inputFormatClass
-
Input Format Class
-
-
-conf, --mapReduceConfigFile
-
MapReduce Configuration
-
-
* -hdfsbase, --mapReduceHdfsBaseDir
-
Fully qualified path to the base directory in hdfs
-
-
* -hdfs, --mapReduceHdfsHostPort
-
HDFS hostname and port in the format hostname:port
-
-
-jobtracker, --mapReduceJobtrackerHostPort
-
[REQUIRED (or resourceman)] Hadoop job tracker hostname and port in the format hostname:port
-
-
-resourceman, --mapReduceYarnResourceManager
-
[REQUIRED (or jobtracker)] Yarn resource manager hostname and port in the format hostname:port
-
-
* -oop, --outputHdfsOutputPath
-
Output HDFS File Path
-
-
-ofc, --outputOutputFormat
-
Output Format Class
-
-
-orc, --outputReducerCount
-
Number of Reducers For Output
-
-
-pdt, --partitionDistanceThresholds
-
Comma separated list of distance thresholds, per dimension
-
-
-pdu, --partitionGeometricDistanceUnit
-
Geometric distance unit (m=meters,km=kilometers, see symbols for javax.units.BaseUnit)
-
-
* -pmd, --partitionMaxDistance
-
Maximum Partition Distance
-
-
-pms, --partitionMaxMemberSelection
-
Maximum number of members selected from a partition
-
-
-pp, --partitionPartitionPrecision
-
Partition Precision
-
-
-pc, --partitionPartitionerClass
-
Index Identifier for Centroids
-
-
-psp, --partitionSecondaryPartitionerClass
-
Perform secondary partitioning with the provided class
-
-
* --query.adapters
-
The comma-separated list of data adapters to query; by default all adapters are used.
-
-
--query.auth
-
The comma-separated list of authorizations used during extract; by default all authorizations are used.
-
-
--query.index
-
The specific index to query; by default one is chosen for each adapter.
-
geowave analytic sql
NAME
geowave analytic sql - SparkSQL queries
SYNOPSIS
geowave analytic sql [options] <sql query> - e.g. 'select * from storename[|adaptername] where condition…'
DESCRIPTION
The geowave analytic sql operator will execute a SparkSQL query
OPTIONS
-
--csv
-
The output CSV file name
-
-
--out
-
The output datastore name
-
-
--outtype
-
The output feature type (adapter) name
-
-
-s, --show
-
Number of result rows to display
-
Default: 20
-
Accumulo Commands
Utility operations to set accumulo splits and run a test server (Required options are designated with an *)
geowave accumulo dbscan
NAME
geowave accumulo presplitpartitionid - Pre-split Accumulo table by providing the number of partition IDs
SYNOPSIS
geowave accumulo presplitpartitionid [options] <storename>
DESCRIPTION
This command will pre-split an accumulo table by providing the number of partition IDs
OPTIONS
-
--indexId
-
The geowave index ID (optional; default is all indices)
-
-
--num
-
The number of partitions (or entries)
-
Default: 0
-
geowave accumulo runserver
NAME
geowave accumulo runserver - Runs a standalone mini Accumulo server for test and debug with GeoWave
SYNOPSIS
geowave accumulo runserver [options]
DESCRIPTION
This command will run a standalone mini single-node accumulo server, which can be used locally for testing and debugging GeoWave, without needing to stand up an entire cluster.
OPTIONS
There are currently no options available for this command
geowave accumulo splitequalinterval
NAME
geowave accumulo splitequalinterval - Set Accumulo splits by providing the number of partitions based on an equal interval strategy
SYNOPSIS
geowave accumulo splitequalinterval [options] <storename>
DESCRIPTION
This command will allow a user to set the accumulated splits through providing the number of partitions based on an equal interval strategy.
OPTIONS
-
--indexId
-
The geowave index ID (optional; default is all indices)
-
-
--num
-
The number of partitions (or entries)
-
Default: 0
-
geowave accumulo splitnumrecords
NAME
geowave accumulo splitnumrecords - Set Accumulo splits by providing the number of entries per split
SYNOPSIS
geowave accumulo splitnumrecords [options] <storename>
DESCRIPTION
This command will set the accumulo datastore splits by providing the number of entries per split.
OPTIONS
-
--indexId
-
The geowave index ID (optional; default is all indices)
-
-
--num
-
The number of partitions (or entries)
-
Default: 0
-
geowave accumulo splitquantile
NAME
geowave accumulo splitquantile - Set Accumulo splits by providing the number of partitions based on a quantile distribution strategy
SYNOPSIS
geowave accumulo splitquantile [options] <storename>
DESCRIPTION
This command will allow a user to set the accumulo datastore splits by providing the number of partitions based on a quantile distribution strategy.
OPTIONS
-
--indexId
-
The geowave index ID (optional; default is all indices)
-
-
--num
-
The number of partitions (or entries)
-
Default: 0
-
GeoServer Commands
Commands that manage geoserver data stores and layers (Required options are designated with an *)
geowave gs addcs
NAME
geowave gs addcs - Add a GeoServer coverage store
SYNOPSIS
geowave gs addcs [options] <GeoWave store name>
DESCRIPTION
This command will add a GeoServer coverage store
OPTIONS
-
-cs, --coverageStore
-
<coverage store name>
-
-
-histo, --equalizeHistogramOverride
-
This parameter will override the behavior to always perform histogram equalization if a histogram exists.
-
Valid values are true and false.
-
-
-interp, --interpolationOverride
-
This will override the default interpolation stored for each layer.
-
Valid values are 0, 1, 2, 3 for NearestNeighbor, Bilinear, Bicubic, and Bicubic (polynomial variant) resepctively.
-
-
-scale, --scaleTo8Bit
-
By default, integer values will automatically be scaled to 8-bit and floating point values will not. This can be overridden setting this value to true or false.
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs addcvs
NAME
geowave gs addcv - Add a GeoServer coverage
SYNOPSIS
geowave gs addcv [options] <coverage name>
DESCRIPTION
This command will add a GeoServer coverage
OPTIONS
-
* -cs, --cvgstore
-
<coverage store name>
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs addds
NAME
geowave gs addds - Add a GeoServer datastore
SYNOPSIS
geowave gs addds [options] <GeoWave store name>
DESCRIPTION
This command will add a GeoServer datastore
OPTIONS
-
-ds, --datastore
-
<datastore name>
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs addfl
NAME
geowave gs addfl - Add a GeoServer feature layer
SYNOPSIS
geowave gs addfl [options] <layer name>
DESCRIPTION
This command will add a GeoServer feature layer
OPTIONS
-
* -ds, --datastore
-
<datastore name>
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs addlayer
NAME
geowave gs addlayer - Add a GeoServer layer from the given GeoWave store
SYNOPSIS
geowave gs addlayer [options] <GeoWave store name>
DESCRIPTION
This command will add a GeoServer layer from the given GeoWave store
OPTIONS
-
-id, --adapterId
-
select just <adapter id> from the store
-
-
-a, --add
-
For multiple layers, add (all | raster | vector)
-
Possible Values: [ALL, RASTER, VECTOR]
-
-
-sld, --setStyle
-
<default style sld>
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs addstyle
NAME
geowave gs addstyle - Add a GeoServer style
SYNOPSIS
geowave gs addstyle [options] <GeoWave style name>
DESCRIPTION
This command will add a GeoServer style
OPTIONS
-
* -sld, --stylesld
-
<style sld file>
-
geowave gs addws
NAME
geowave gs addws - Add GeoServer workspace
SYNOPSIS
geowave gs addws [options] <workspace name>
DESCRIPTION
This command will add a GeoServer workspace
OPTIONS
There are currently no options available for this command
geowave gs getcs
NAME
geowave gs getcs - Get GeoServer CoverageStore info
SYNOPSIS
geowave gs getcs [options] <coverage store name>
DESCRIPTION
This command will return GeoServer CoverageStore info
OPTIONS
-
-ws, --workspace
-
<workspace name>
-
geowave gs getcv
NAME
geowave gs getcv - Get a GeoServer coverage’s info
SYNOPSIS
geowave gs getcv [options] <coverage name>
DESCRIPTION
This command will return a GeoServer coverage’s info
OPTIONS
-
-cs, --coverageStore
-
<coverage store name>
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs getds
NAME
geowave gs getds - Get GeoServer DataStore info
SYNOPSIS
geowave gs getds [options] <datastore name>
DESCRIPTION
This command will return a GeoServer DataStore info
OPTIONS
-
-ws, --workspace
-
<workspace name>
-
geowave gs getfl
NAME
geowave gs getfl - Get GeoServer feature layer info
SYNOPSIS
geowave gs getfl [options] <layer name>
DESCRIPTION
This command will return GeoServer feature layer info
OPTIONS
There are currently no options available for this command
geowave gs getsa
NAME
geowave config getsa - Get GeoWave store adapters
SYNOPSIS
geowave gs getsa [options] <store name>
DESCRIPTION
This command will return GeoWave store adapters
OPTIONS
There are currently no options available for this command
geowave gs getstyle
NAME
geowave gs getstyle - Get GeoServer Style info
SYNOPSIS
geowave gs getstyle [options] <style name>
DESCRIPTION
This command will return GeoServer Style info
OPTIONS
There are currently no options available for this command
geowave gs listcs
NAME
geowave gs listcs - List GeoServer coverage stores
SYNOPSIS
geowave gs listcs [options]
DESCRIPTION
This command will list all GeoServer coverage stores
OPTIONS
-
-ws, --workspace
-
<workspace name>
-
geowave gs listcv
NAME
geowave gs listcv - List GeoServer Coverages
SYNOPSIS
geowave gs listcv [options] <coverage store name>
DESCRIPTION
This command will list all GeoServer Coverages
OPTIONS
-
-ws, --workspace
-
<workspace name>
-
geowave gs listds
NAME
geowave gs listds - List GeoServer datastores
SYNOPSIS
geowave gs listds [options]
DESCRIPTION
This command will list all GeoServer datastores
OPTIONS
-
-ws, --workspace
-
<workspace name>
-
geowave gs listfl
NAME
geowave gs listfl - List GeoServer feature layers
SYNOPSIS
geowave gs listfl [options]
DESCRIPTION
This command will list all GeoServer feature layers
OPTIONS
-
-ds, --datastore
-
Datastore Name
-
-
-g, --geowaveOnly
-
Show only GeoWave feature layers (default: false)
-
Default: false
-
-
-ws, --workspace
-
Workspace Name
-
geowave gs liststyles
NAME
geowave gs liststyles - List GeoServer styles
SYNOPSIS
geowave gs liststyles [options]
DESCRIPTION
This command will list all GeoServer styles
OPTIONS
There are currently no options available for this command
geowave gs listws
NAME
geowave gs listws - List GeoServer workspaces
SYNOPSIS
geowave gs listws [options]
DESCRIPTION
This command will list all GeoServer workspaces
OPTIONS
There are currently no options available for this command
geowave gs rmcs
NAME
geowave gs rmcs - Remove GeoServer Coverage Store
SYNOPSIS
geowave gs rmcs [options] <coverage store name>
DESCRIPTION
This command will remove a GeoServer Coverage Store
OPTIONS
-
-ws, --workspace
-
Workspace Name
-
geowave gs rmcv
NAME
geowave gs rmcv - Remove a GeoServer coverage
SYNOPSIS
geowave gs rmcv [options] <coverage name>
DESCRIPTION
This command will remove a GeoServer coverage
OPTIONS
-
* -cs, --cvgstore
-
<coverage store name>
-
-
-ws, --workspace
-
<workspace name>
-
geowave gs rmds
NAME
geowave gs rmds - Remove GeoServer DataStore
SYNOPSIS
geowave gs rmds [options] <datastore name>
DESCRIPTION
This command will remove a GeoServer DataStore
OPTIONS
-
-ws, --workspace
-
<workspace name>
-
geowave gs rmfl
NAME
geowave gs rmfl - Remove GeoServer feature Layer
SYNOPSIS
geowave gs rmfl [options] <layer name>
DESCRIPTION
This command will remove a GeoServer feature Layer
OPTIONS
There are currently no options available for this command
geowave gs rmstyle
NAME
geowave gs rmstyle - Remove GeoServer Style
SYNOPSIS
geowave gs rmstyle [options] <style name>
DESCRIPTION
This command will remove a GeoServer Style
OPTIONS
There are currently no options available for this command
HBase Commands
Utility operations to combine statistics in hbase (Required options are designated with an *)
geowave hbase combinestats
NAME
geowave hbase combinestats - Combine all statistics in HBase namespace
SYNOPSIS
geowave hbase combinestats [options] <storename> <adapter id>
DESCRIPTION
This command will combine all statistics in an HBase namespace
OPTIONS
There are currently no options for this command
geowave hbase runserver
NAME
geowave hbase runserver - Runs a standalone mini HBase server for test and debug with GeoWave
SYNOPSIS
geowave hbase runserver [options]
DESCRIPTION
This command will run a standalone mini single-node HBase server, which can be used locally for testing and debugging GeoWave, without needing to stand up an entire cluster.
OPTIONS
There are currently no options for this command
Landsat8 Commands
Operations to analyze, download, and ingest Landsat 8 imagery publicly available on AWS (Required options are designated with an *)
geowave landsat analyze
NAME
geowave landsat analyze - Print out basic aggregate statistics for available Landsat 8 imagery
SYNOPSIS
geowave landsat analyze [options]
DESCRIPTION
This command will print out basic aggregate statistics that are for available Landsat 8 imagery
OPTIONS
-
--cql
-
An optional CQL expression to filter the ingested imagery. The feature type for the expression has the following attributes: shape (Geometry), acquisitionDate (Date), cloudCover (double), processingLevel (String), path (int), row (int) and the feature ID is entityId for the scene. Additionally attributes of the individuals band can be used such as band (String), sizeMB (double), and bandDownloadUrl (String)
-
Default: <empty string>
-
-
--nbestbands
-
An option to identify and only use a set number of bands with the best cloud cover
-
Default: 0
-
-
--nbestperspatial
-
A boolean flag, when applied with --nbestscenes or --nbestbands will aggregate scenes and/or bands by path/row
-
Default: false
-
-
--nbestscenes
-
An option to identify and only use a set number of scenes with the best cloud cover
-
Default: 0
-
-
--sincelastrun
-
An option to check the scenes list from the workspace and if it exists, to only ingest data since the last scene.
-
Default: false
-
-
--usecachedscenes
-
An option to run against the existing scenes catalog in the workspace directory if it exists.
-
Default: false
-
-
-ws, --workspaceDir
-
A local directory to write temporary files needed for landsat 8 ingest.
-
Default is <TEMP_DIR>/landsat8
-
Default: landsat8
-
geowave landsat download
NAME
geowave landsat download - Download Landsat 8 imagery to a local directory
SYNOPSIS
geowave landsat download [options]
DESCRIPTION
This command will download the Landsat 8 imagery to a local directory
OPTIONS
-
--cql
-
An optional CQL expression to filter the ingested imagery. The feature type for the expression has the following attributes: shape (Geometry), acquisitionDate (Date), cloudCover (double), processingLevel (String), path (int), row (int) and the feature ID is entityId for the scene. Additionally attributes of the individuals band can be used such as band (String), sizeMB (double), and bandDownloadUrl (String)
-
Default: <empty string>
-
-
--nbestbands
-
An option to identify and only use a set number of bands with the best cloud cover
-
Default: 0
-
-
--nbestperspatial
-
A boolean flag, when applied with --nbestscenes or --nbestbands will aggregate scenes and/or bands by path/row
-
Default: false
-
-
--nbestscenes
-
An option to identify and only use a set number of scenes with the best cloud cover
-
Default: 0
-
-
--sincelastrun
-
An option to check the scenes list from the workspace and if it exists, to only ingest data since the last scene.
-
Default: false
-
-
--usecachedscenes
-
An option to run against the existing scenes catalog in the workspace directory if it exists.
-
Default: false
-
-
-ws, --workspaceDir
-
A local directory to write temporary files needed for landsat 8 ingest.
-
Default is <TEMP_DIR>/landsat8
-
Default: landsat8
-
geowave landsat ingest
NAME
geowave landsat ingest - Ingest routine for locally downloading Landsat 8 imagery and ingesting it into GeoWave’s raster store and in parallel ingesting the scene metadata into GeoWave’s vector store. These two stores can actually be the same or they can be different.
SYNOPSIS
geowave landsat ingest [options] <storename> <comma delimited index/group list>
DESCRIPTION
This command provides a standard ingest routine for downloading Landsat 8 imagery and ingesting it into GeoWave’s raster store. In parallel, it allows for ingesting the scene metadata into GeoWave’s vector store.
OPTIONS
-
--converter
-
Prior to ingesting an image, this converter will be used to massage the data. The default is not to convert the data.
-
-
--coverage
-
The name to give to each unique coverage. Freemarker templating can be used for variable substition based on the same attributes used for filtering. The default coverage name is '${entityId}_${band}'. If ${band} is unused in the coverage name, all bands will be merged together into the same coverage.
-
Default: ${entityId}_${band}
-
-
--cql
-
An optional CQL expression to filter the ingested imagery. The feature type for the expression has the following attributes: shape (Geometry), acquisitionDate (Date), cloudCover (double), processingLevel (String), path (int), row (int) and the feature ID is entityId for the scene. Additionally attributes of the individuals band can be used such as band (String), sizeMB (double), and bandDownloadUrl (String)
-
Default: <empty string>
-
-
--crop
-
Use the spatial constraint provided in CQL to crop the image. If no spatial constraint is provided, this will not have an effect.
-
Default: false
-
-
--histogram
-
An option to store the histogram of the values of the coverage so that histogram equalization will be performed
-
Default: false
-
-
--nbestbands
-
An option to identify and only use a set number of bands with the best cloud cover
-
Default: 0
-
-
--nbestperspatial
-
A boolean flag, when applied with --nbestscenes or --nbestbands will aggregate scenes and/or bands by path/row
-
Default: false
-
-
--nbestscenes
-
An option to identify and only use a set number of scenes with the best cloud cover
-
Default: 0
-
-
--overwrite
-
An option to overwrite images that are ingested in the local workspace directory. By default it will keep an existing image rather than downloading it again.
-
Default: false
-
-
--pyramid
-
An option to store an image pyramid for the coverage
-
Default: false
-
-
--retainimages
-
An option to keep the images that are ingested in the local workspace directory. By default it will delete the local file after it is ingested successfully.
-
Default: false
-
-
--sincelastrun
-
An option to check the scenes list from the workspace and if it exists, to only ingest data since the last scene.
-
Default: false
-
-
--skipMerge
-
By default the ingest will automerge overlapping tiles as a post-processing optimization step for efficient retrieval, but this will skip the merge process
-
Default: false
-
-
--subsample
-
Subsample the image prior to ingest by the scale factor provided. The scale factor should be an integer value greater than 1.
-
Default: 1
-
-
--tilesize
-
The option to set the pixel size for each tile stored in GeoWave.
-
Default: 512
-
-
--usecachedscenes
-
An option to run against the existing scenes catalog in the workspace directory if it exists.
-
Default: false
-
-
--vectorindex
-
By ingesting as both vectors and rasters you may want each indexed differently. This will override the index used for vector output.
-
-
--vectorstore
-
By ingesting as both vectors and rasters you may want to ingest into different stores. This will override the store for vector output.
-
-
-ws, --workspaceDir
-
A local directory to write temporary files needed for landsat 8 ingest. Default is <TEMP_DIR>/landsat8
-
Default: landsat8
-
geowave landsat ingestraster
NAME
geowave landsat ingestraster - Ingest routine for locally downloading Landsat 8 imagery and ingesting it into GeoWave
SYNOPSIS
geowave landsat analyze [options]
DESCRIPTION
This command provides an ingest routine for locally downloading Landsat 8 imagery and ingesting it into GeoWave.
OPTIONS
-
--converter
-
Prior to ingesting an image, this converter will be used to massage the data. The default is not to convert the data.
-
-
--coverage
-
The name to give to each unique coverage. Freemarker templating can be used for variable substition based on the same attributes used for filtering. The default coverage name is '${entityId}_${band}'. If ${band} is unused in the coverage name, all bands will be merged together into the same coverage.
-
Default: ${entityId}_${band}
-
-
--cql
-
An optional CQL expression to filter the ingested imagery. The feature type for the expression has the following attributes: shape (Geometry), acquisitionDate (Date), cloudCover (double), processingLevel (String), path (int), row (int) and the feature ID is entityId for the scene. Additionally attributes of the individuals band can be used such as band (String), sizeMB (double), and bandDownloadUrl (String)
-
Default: <empty string>
-
-
--crop
-
Use the spatial constraint provided in CQL to crop the image. If no spatial constraint is provided, this will not have an effect.
-
Default: false
-
-
--histogram
-
An option to store the histogram of the values of the coverage so that histogram equalization will be performed
-
Default: false
-
-
--nbestbands
-
An option to identify and only use a set number of bands with the best cloud cover
-
Default: 0
-
-
--nbestperspatial
-
A boolean flag, when applied with --nbestscenes or --nbestbands will aggregate scenes and/or bands by path/row
-
Default: false
-
-
--nbestscenes
-
An option to identify and only use a set number of scenes with the best cloud cover
-
Default: 0
-
-
--overwrite
-
An option to overwrite images that are ingested in the local workspace directory. By default it will keep an existing image rather than downloading it again.
-
Default: false
-
-
--pyramid
-
An option to store an image pyramid for the coverage
-
Default: false
-
-
--retainimages
-
An option to keep the images that are ingested in the local workspace directory. By default it will delete the local file after it is ingested successfully.
-
Default: false
-
-
--sincelastrun
-
An option to check the scenes list from the workspace and if it exists, to only ingest data since the last scene.
-
Default: false
-
-
--skipMerge
-
By default the ingest will automerge overlapping tiles as a post-processing optimization step for efficient retrieval, but this will skip the merge process
-
Default: false
-
-
--subsample
-
Subsample the image prior to ingest by the scale factor provided. The scale factor should be an integer value greater than 1.
-
Default: 1
-
-
--tilesize
-
The option to set the pixel size for each tile stored in GeoWave. The default is 512
-
Default: 512
-
-
--usecachedscenes
-
An option to run against the existing scenes catalog in the workspace directory if it exists.
-
Default: false
-
-
-ws, --workspaceDir
-
A local directory to write temporary files needed for landsat 8 ingest. Default is <TEMP_DIR>/landsat8
-
Default: landsat8
-
geowave landsat ingestvector
NAME
geowave landsat ingestvector - Ingest routine for searching landsat scenes that match certain criteria and ingesting the scene and band metadata into GeoWave’s vector store.
SYNOPSIS
geowave landsat ingestvector [options] <storename> <comma delimited index/group list>
DESCRIPTION
This command provides an ingest routine for searching landsat scenes that match certain criteria, and ingesting the scene and band metadata into GeoWave’s vector store.
OPTIONS
-
--cql
-
An optional CQL expression to filter the ingested imagery. The feature type for the expression has the following attributes: shape (Geometry), acquisitionDate (Date), cloudCover (double), processingLevel (String), path (int), row (int) and the feature ID is entityId for the scene. Additionally attributes of the individuals band can be used such as band (String), sizeMB (double), and bandDownloadUrl (String)
-
Default: <empty string>
-
-
--nbestbands
-
An option to identify and only use a set number of bands with the best cloud cover
-
Default: 0
-
-
--nbestperspatial
-
A boolean flag, when applied with --nbestscenes or --nbestbands will aggregate scenes and/or bands by path/row
-
Default: false
-
-
--nbestscenes
-
An option to identify and only use a set number of scenes with the best cloud cover
-
Default: 0
-
-
--sincelastrun
-
An option to check the scenes list from the workspace and if it exists, to only ingest data since the last scene.
-
Default: false
-
-
--usecachedscenes
-
An option to run against the existing scenes catalog in the workspace directory if it exists.
-
Default: false
-
-
-ws, --workspaceDir
-
A local directory to write temporary files needed for landsat 8 ingest. Default is <TEMP_DIR>/landsat8
-
Default: landsat8
-
OSM Commands
Operations to ingest Open Street Map (OSM) nodes, ways and relations to GeoWave (Required options are designated with an *)
geowave osm ingest
NAME
geowave osm ingest - Ingest and convert OSM data from HDFS to GeoWave
SYNOPSIS
geowave osm ingest [options] <hdfs host:port> <path to base directory to read from> <store name>
DESCRIPTION
This command will ingest and convert OSM data from HDFS to GeoWave.
OPTIONS
-
-jn, --jobName
-
Name of mapreduce job
-
Default: Ingest (mcarrier)
-
-
-m, --mappingFile
-
Mapping file, imposm3 form
-
-
--table
-
OSM Table name in GeoWave
-
Default: OSM
-
-
* -t, --type
-
Mapper type - one of node, way, or relation
-
-
-v, --visibility
-
The visibility of the data ingested (optional; default is 'public')
-
geowave osm stage
NAME
geowave osm stage - Stage OSM data to HDFS
SYNOPSIS
geowave osm stage [options] <file or directory> <hdfs host:port> <path to base directory to write to>
DESCRIPTION
This command will stage OSM data from a local directory and write it to HDFS
OPTIONS
-
--extension
-
PBF File extension
-
Default: .pbf
-
Raster Commands
Operations to perform transformations on raster data in GeoWave (Required options are designated with an *)
geowave raster resize
NAME
geowave raster resize - Resize Raster Tiles
SYNOPSIS
geowave raster resize [options] <input store name> <output store name>
DESCRIPTION
This command will resize raster tiles that are stored in a GeoWave datastore, and write the resized tiles to a new output store.
OPTIONS
-
* --hdfsHostPort
-
the hdfs host port
-
-
--indexId
-
The index that the input raster is stored in
-
-
* --inputCoverageName
-
The name of the input raster coverage
-
-
* --jobSubmissionHostPort
-
The job submission tracker
-
-
--maxSplits
-
The max partitions for the input data
-
-
--minSplits
-
The min partitions for the input data
-
-
* --outputCoverageName
-
The out output raster coverage name
-
-
* --outputTileSize
-
The tile size to output
-
Vector Commands
Vector data operations (Required options are designated with an *)
geowave vector cqldelete
NAME
geowave vector cqldelete - Delete data that matches a CQL filter
SYNOPSIS
geowave vector cqldelete [options] <storename>
DESCRIPTION
This command will delete all data in a data store that matches a CQL filter.
OPTIONS
-
--adapterId
-
Optional ability to provide an adapter ID
-
-
* --cql
-
CQL Filter for delete
-
-
--debug
-
Print out additional info for debug purposes
-
Default: false
-
-
--indexId
-
The name of the index (optional)
-
geowave vector localexport
NAME
geowave vector localexport - Export data directly
SYNOPSIS
geowave vector localexport [options] <store name>
DESCRIPTION
This command will export data from a data store
OPTIONS
-
--adapterIds
-
Comma separated list of adapter Ids
-
-
--batchSize
-
Records to process at a time
-
Default: 10000
-
-
--cqlFilter
-
Filter exported data based on CQL filter
-
-
--indexId
-
The index to export from
-
-
* --outputFile
geowave vector mrexport
NAME
geowave vector mrexport - Export data using map-reduce
SYNOPSIS
geowave vector mrexport [options] <hdfs host:port> <path to base directory to write to> <store name>
DESCRIPTION
This command will perform a data export for data in a data store, and will use MapReduce to support high-volume data stores.
OPTIONS
-
--adapterIds
-
Comma separated list of adapter Ids
-
-
--batchSize
-
Records to process at a time
-
Default: 10000
-
-
--cqlFilter
-
Filter exported data based on CQL filter
-
-
--indexId
-
The index to export from
-
-
--maxSplits
-
The max partitions for the input data
-
-
--minSplits
-
The min partitions for the input data
-
-
--resourceManagerHostPort