Create harvest job configuration
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string^[0-9]*$ Unique, numeric ID for the job definition. Will be assigned if not provided. | ||
name required | string The name assigned to the harvest configuration. | ||
description | string Free form description of the configuration to support the administration. | ||
openAccess | string Enum: "true" "false" tbd | ||
required | object or object Reference to the storage configuration to use. | ||
Any of
| |||
required | object or object Reference to the transformation pipeline to use. | ||
Any of
| |||
enabled required | string Enum: "true" "false" Indicates if the job is scheduled for running | ||
harvestImmediately required | string Enum: "true" "false" Whether to harvest when the config is persisted. | ||
scheduleString | string Crontab style schedule string (simplified): minute(0-59) hour(0-24) day of month(* or 1-31) month (* or 1-12) day of week (* or 0-6). | ||
dateFormat | string For example yyyy-MM-dd'T'hh:mm:ss'Z'. | ||
url required | string The URL to access the data from. | ||
timeout | string^[0-9]*$ Default: "300" Connection/read timeout in seconds; application depending on the specific protocol used for fetching data. | ||
cacheEnabled | string Enum: "true" "false" Whether or not to store incoming records in Harvester's file system. | ||
diskRun | string Enum: "true" "false" Whether or not to run harvest job from records cached in a previous job run. | ||
storageBatchLimit | string^[0-9]*$ Batch size: Number of records to send to storage at a time. | ||
recordLimit | string^[0-9]*$ Maximum number of records to harvest. | ||
laxParsing | string Default: "false" Enum: "true" "false" When enabled, Harvester will attempt to parse malformed XML (missing closing tags, entities) | ||
constantFields | string Values related to target handling in MasterKey. Otherwise obsolete. | ||
storeOriginal | string Default: "false" Enum: "true" "false" Indicates whether to store incoming original record, if supported by the job type and the storage configuration. | ||
managedBy | string Free-text field for tagging a job with the producer or manager of the resource. Multiple tags may be separated by commas. The tags can be used for filtering status reports by job administrators for example. | ||
usedBy | string Free form administrative information; could be tags for the clients using this harvestable. | ||
serviceProvider | string Free-text field for administrative information about the harvest job. | ||
contactNotes | string Free form text field for administrator's notes. | ||
technicalNotes | string Free-text field for administrative information. | ||
logLevel | string Enum: "ERROR" "WARN" "INFO" "DEBUG" "TRACE" Specifies the logging level for the job with TRACE being the most (extremely) verbose. INFO is the recommended log level in most cases. | ||
failedRecordsLogging | string Default: "CLEAN_DIRECTORY" Enum: "NO_STORE" "CLEAN_DIRECTORY" "CREATE_OVERWRITE" "ADD_ALL" Specify whether or not failed records should be saved as XML files in a designated log directory. Also specifies retention policy for the directory, that is, whether to retain files that were saved in previous runs (CLEAN_DIRECTORY = don't retain) and, if so, whether to overwrite any existing files if the same record fails again (CREATE_OVERWRITE) or rather add a sequence number to the new file name in order not to overwrite (ADD_ALL). | ||
maxSavedFailedRecordsPerRun | string^[0-9]*$ Default: "100" Sets a maximum number of files to save in the failed records directory per run. The job log will tell when the limit is reached. | ||
maxSavedFailedRecordsTotal | string^[0-9]*$ Default: "1000" Sets a maximum number of files to be saved in the failed records directory at any given time - as the sum of previously saved records (that were not cleaned up before this run) plus any new records added during the run.The job log will tell when the limit is reached. | ||
mailAddress | string Comma separated list of e-mail addresses that should receive notification on job completion. | ||
mailLevel | string Enum: "OK" "WARN" "ERROR" The minimum severity of a job's completion status that will trigger email notification. | ||
initiallyHarvested | string Date and time, assigned by Harvester | ||
lastHarvestFinished | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration completed. | ||
lastHarvestStarted | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration began. | ||
lastUpdated | string Assigned by API. The date and time when this definition was last modified. | ||
nextHarvestSchedule | string The date and time when a job with this definition should be run (if job is enabled). | ||
amountHarvested | string^[0-9]*$ Assigned by API. Number of records harvested in last run. It seems this should really be an integer, but string is what the WSAPI gives us. | ||
message | string Assigned by API. Message summarising results of last run. | ||
type required | string Value: "xmlBulk" Indicates bulk XML job. | ||
retryCount | string^[0-9]*$ Default: "2" Obsolete but allowed for XML bulk. | ||
retryWait | string^[0-9]*$ Default: "60" Obsolete but allowed for XML bulk. | ||
allowErrors | string Default: "false" Enum: "true" "false" Whether or not to continue despite harvest record errors. | ||
allowCondReq | string Default: "false" Enum: "true" "false" Whether or not to filter on file date to only harvest new XML files | ||
fromDate | string Initial start date (yyyy-MM-dd) for incremental updates (when allowCondReq is 'true') | ||
csvConfiguration | string Semicolon-separated key-value pairs that specifies parsing of a CSV file into XML for further processing (see Harvester documentation for details). | ||
excludeFilePattern | string Regular expression; setting to skip harvesting of files with names matching the given regular expression (see Harvester documentation for details). | ||
expectedSchema | string Mime-type override (e.g: application/marc; charset=MARC-8). | ||
includeFilePattern | string Regular expression; setting to request harvesting of files with names matching the given regular expression unless those file names are simultaneously excluded by the excludeFilePattern. .zip, .gz, .tar included by default unless explicitly excluded by excludeFilePattern (see Harvester documentation for details). | ||
outputSchema | string MARC XML transformation format (application/marc or application/tmarc). | ||
passiveMode | string Default: "false" Enum: "true" "false" Whether or not to use passive mode for FTP transfers. | ||
recurse | string Default: "false" Enum: "true" "false" Whether or not to recurse into sub-folders in the source directory tree. | ||
splitAt | string^[0-9]*$ Level/depth to split XML files at to extract records. Zero/empty disables split. | ||
splitSize | string^[0-9]*$ Setting to split large XML files into chunks of `splitSize' number of records; to preserve memory during XSLT transformations. | ||
json | object Custom configurations in JSON format (has no current applications). | ||
overwrite | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. Will delete all previously harvested data before beginning the next scheduled (or manually triggered) run, if set to true. | ||
keepPartial | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. When true, partial records harvested during a failed harvest run will be retained in Solr. |
{- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "id": "string"
}, - "transformation": {
- "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "storageBatchLimit": "string",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "initiallyHarvested": "string",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "type": "xmlBulk",
- "retryCount": "2",
- "retryWait": "60",
- "allowErrors": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "json": { },
- "overwrite": "true",
- "keepPartial": "true"
}
{- "type": "xmlBulk",
- "allowErrors": "true",
- "overwrite": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "entityType": "inventoryStorageEntity",
- "bulkSize": "string",
- "currentStatus": "string",
- "customClass": "string",
- "enabled": "true",
- "idAsString": "string",
- "name": "string",
- "url": "string"
}, - "transformation": {
- "entityType": "basicTransformation",
- "acl": "string",
- "description": "string",
- "enabled": "true",
- "name": "string",
- "parallel": "true",
- "stepAssociations": [
- {
- "id": "string",
- "position": "string",
- "step": {
- "entityType": "xmlTransformationStep",
- "acl": "string",
- "description": "string",
- "inputFormat": "string",
- "name": "string",
- "outputFormat": "string",
- "script": "<'script' omitted from nested displays>",
- "id": "string",
- "testData": "<'testData' omitted from nested displays>",
- "testOutput": "<'testOutput' omitted from nested displays>"
}, - "transformation": "string"
}
], - "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "currentStatus": "NEW",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "acl": "string"
}
Get brief harvest job definitions
query | string CQL |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "titles": [
- {
- "id": "string",
- "name": "string",
- "currentStatus": "NEW",
- "enabled": "true",
- "storageUrl": "string",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "jobClass": "XmlBulkResource",
- "amountHarvested": "string",
- "message": "string"
}
]
}
Create bulk XML harvest job configuration
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string^[0-9]*$ Unique, numeric ID for the job definition. Will be assigned if not provided. | ||
name required | string The name assigned to the harvest configuration. | ||
description | string Free form description of the configuration to support the administration. | ||
openAccess | string Enum: "true" "false" tbd | ||
required | object or object Reference to the storage configuration to use. | ||
Any of
| |||
required | object or object Reference to the transformation pipeline to use. | ||
Any of
| |||
enabled required | string Enum: "true" "false" Indicates if the job is scheduled for running | ||
harvestImmediately required | string Enum: "true" "false" Whether to harvest when the config is persisted. | ||
scheduleString | string Crontab style schedule string (simplified): minute(0-59) hour(0-24) day of month(* or 1-31) month (* or 1-12) day of week (* or 0-6). | ||
dateFormat | string For example yyyy-MM-dd'T'hh:mm:ss'Z'. | ||
url required | string The URL to access the data from. | ||
timeout | string^[0-9]*$ Default: "300" Connection/read timeout in seconds; application depending on the specific protocol used for fetching data. | ||
cacheEnabled | string Enum: "true" "false" Whether or not to store incoming records in Harvester's file system. | ||
diskRun | string Enum: "true" "false" Whether or not to run harvest job from records cached in a previous job run. | ||
storageBatchLimit | string^[0-9]*$ Batch size: Number of records to send to storage at a time. | ||
recordLimit | string^[0-9]*$ Maximum number of records to harvest. | ||
laxParsing | string Default: "false" Enum: "true" "false" When enabled, Harvester will attempt to parse malformed XML (missing closing tags, entities) | ||
constantFields | string Values related to target handling in MasterKey. Otherwise obsolete. | ||
storeOriginal | string Default: "false" Enum: "true" "false" Indicates whether to store incoming original record, if supported by the job type and the storage configuration. | ||
managedBy | string Free-text field for tagging a job with the producer or manager of the resource. Multiple tags may be separated by commas. The tags can be used for filtering status reports by job administrators for example. | ||
usedBy | string Free form administrative information; could be tags for the clients using this harvestable. | ||
serviceProvider | string Free-text field for administrative information about the harvest job. | ||
contactNotes | string Free form text field for administrator's notes. | ||
technicalNotes | string Free-text field for administrative information. | ||
logLevel | string Enum: "ERROR" "WARN" "INFO" "DEBUG" "TRACE" Specifies the logging level for the job with TRACE being the most (extremely) verbose. INFO is the recommended log level in most cases. | ||
failedRecordsLogging | string Default: "CLEAN_DIRECTORY" Enum: "NO_STORE" "CLEAN_DIRECTORY" "CREATE_OVERWRITE" "ADD_ALL" Specify whether or not failed records should be saved as XML files in a designated log directory. Also specifies retention policy for the directory, that is, whether to retain files that were saved in previous runs (CLEAN_DIRECTORY = don't retain) and, if so, whether to overwrite any existing files if the same record fails again (CREATE_OVERWRITE) or rather add a sequence number to the new file name in order not to overwrite (ADD_ALL). | ||
maxSavedFailedRecordsPerRun | string^[0-9]*$ Default: "100" Sets a maximum number of files to save in the failed records directory per run. The job log will tell when the limit is reached. | ||
maxSavedFailedRecordsTotal | string^[0-9]*$ Default: "1000" Sets a maximum number of files to be saved in the failed records directory at any given time - as the sum of previously saved records (that were not cleaned up before this run) plus any new records added during the run.The job log will tell when the limit is reached. | ||
mailAddress | string Comma separated list of e-mail addresses that should receive notification on job completion. | ||
mailLevel | string Enum: "OK" "WARN" "ERROR" The minimum severity of a job's completion status that will trigger email notification. | ||
initiallyHarvested | string Date and time, assigned by Harvester | ||
lastHarvestFinished | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration completed. | ||
lastHarvestStarted | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration began. | ||
lastUpdated | string Assigned by API. The date and time when this definition was last modified. | ||
nextHarvestSchedule | string The date and time when a job with this definition should be run (if job is enabled). | ||
amountHarvested | string^[0-9]*$ Assigned by API. Number of records harvested in last run. It seems this should really be an integer, but string is what the WSAPI gives us. | ||
message | string Assigned by API. Message summarising results of last run. | ||
type required | string Value: "xmlBulk" Indicates bulk XML job. | ||
retryCount | string^[0-9]*$ Default: "2" Obsolete but allowed for XML bulk. | ||
retryWait | string^[0-9]*$ Default: "60" Obsolete but allowed for XML bulk. | ||
allowErrors | string Default: "false" Enum: "true" "false" Whether or not to continue despite harvest record errors. | ||
allowCondReq | string Default: "false" Enum: "true" "false" Whether or not to filter on file date to only harvest new XML files | ||
fromDate | string Initial start date (yyyy-MM-dd) for incremental updates (when allowCondReq is 'true') | ||
csvConfiguration | string Semicolon-separated key-value pairs that specifies parsing of a CSV file into XML for further processing (see Harvester documentation for details). | ||
excludeFilePattern | string Regular expression; setting to skip harvesting of files with names matching the given regular expression (see Harvester documentation for details). | ||
expectedSchema | string Mime-type override (e.g: application/marc; charset=MARC-8). | ||
includeFilePattern | string Regular expression; setting to request harvesting of files with names matching the given regular expression unless those file names are simultaneously excluded by the excludeFilePattern. .zip, .gz, .tar included by default unless explicitly excluded by excludeFilePattern (see Harvester documentation for details). | ||
outputSchema | string MARC XML transformation format (application/marc or application/tmarc). | ||
passiveMode | string Default: "false" Enum: "true" "false" Whether or not to use passive mode for FTP transfers. | ||
recurse | string Default: "false" Enum: "true" "false" Whether or not to recurse into sub-folders in the source directory tree. | ||
splitAt | string^[0-9]*$ Level/depth to split XML files at to extract records. Zero/empty disables split. | ||
splitSize | string^[0-9]*$ Setting to split large XML files into chunks of `splitSize' number of records; to preserve memory during XSLT transformations. | ||
json | object Custom configurations in JSON format (has no current applications). | ||
overwrite | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. Will delete all previously harvested data before beginning the next scheduled (or manually triggered) run, if set to true. | ||
keepPartial | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. When true, partial records harvested during a failed harvest run will be retained in Solr. |
{- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "id": "string"
}, - "transformation": {
- "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "storageBatchLimit": "string",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "initiallyHarvested": "string",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "type": "xmlBulk",
- "retryCount": "2",
- "retryWait": "60",
- "allowErrors": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "json": { },
- "overwrite": "true",
- "keepPartial": "true"
}
{- "type": "xmlBulk",
- "allowErrors": "true",
- "overwrite": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "entityType": "inventoryStorageEntity",
- "bulkSize": "string",
- "currentStatus": "string",
- "customClass": "string",
- "enabled": "true",
- "idAsString": "string",
- "name": "string",
- "url": "string"
}, - "transformation": {
- "entityType": "basicTransformation",
- "acl": "string",
- "description": "string",
- "enabled": "true",
- "name": "string",
- "parallel": "true",
- "stepAssociations": [
- {
- "id": "string",
- "position": "string",
- "step": {
- "entityType": "xmlTransformationStep",
- "acl": "string",
- "description": "string",
- "inputFormat": "string",
- "name": "string",
- "outputFormat": "string",
- "script": "<'script' omitted from nested displays>",
- "id": "string",
- "testData": "<'testData' omitted from nested displays>",
- "testOutput": "<'testOutput' omitted from nested displays>"
}, - "transformation": "string"
}
], - "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "currentStatus": "NEW",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "acl": "string"
}
Create OAI-PMH harvest job configuration
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string^[0-9]*$ Unique, numeric ID for the job definition. Will be assigned if not provided. | ||
name required | string The name assigned to the harvest configuration. | ||
description | string Free form description of the configuration to support the administration. | ||
openAccess | string Enum: "true" "false" tbd | ||
required | object or object Reference to the storage configuration to use. | ||
Any of
| |||
required | object or object Reference to the transformation pipeline to use. | ||
Any of
| |||
enabled required | string Enum: "true" "false" Indicates if the job is scheduled for running | ||
harvestImmediately required | string Enum: "true" "false" Whether to harvest when the config is persisted. | ||
scheduleString | string Crontab style schedule string (simplified): minute(0-59) hour(0-24) day of month(* or 1-31) month (* or 1-12) day of week (* or 0-6). | ||
dateFormat | string For example yyyy-MM-dd'T'hh:mm:ss'Z'. | ||
url required | string The URL to access the data from. | ||
timeout | string^[0-9]*$ Default: "300" Connection/read timeout in seconds; application depending on the specific protocol used for fetching data. | ||
cacheEnabled | string Enum: "true" "false" Whether or not to store incoming records in Harvester's file system. | ||
diskRun | string Enum: "true" "false" Whether or not to run harvest job from records cached in a previous job run. | ||
storageBatchLimit | string^[0-9]*$ Batch size: Number of records to send to storage at a time. | ||
recordLimit | string^[0-9]*$ Maximum number of records to harvest. | ||
laxParsing | string Default: "false" Enum: "true" "false" When enabled, Harvester will attempt to parse malformed XML (missing closing tags, entities) | ||
constantFields | string Values related to target handling in MasterKey. Otherwise obsolete. | ||
storeOriginal | string Default: "false" Enum: "true" "false" Indicates whether to store incoming original record, if supported by the job type and the storage configuration. | ||
managedBy | string Free-text field for tagging a job with the producer or manager of the resource. Multiple tags may be separated by commas. The tags can be used for filtering status reports by job administrators for example. | ||
usedBy | string Free form administrative information; could be tags for the clients using this harvestable. | ||
serviceProvider | string Free-text field for administrative information about the harvest job. | ||
contactNotes | string Free form text field for administrator's notes. | ||
technicalNotes | string Free-text field for administrative information. | ||
logLevel | string Enum: "ERROR" "WARN" "INFO" "DEBUG" "TRACE" Specifies the logging level for the job with TRACE being the most (extremely) verbose. INFO is the recommended log level in most cases. | ||
failedRecordsLogging | string Default: "CLEAN_DIRECTORY" Enum: "NO_STORE" "CLEAN_DIRECTORY" "CREATE_OVERWRITE" "ADD_ALL" Specify whether or not failed records should be saved as XML files in a designated log directory. Also specifies retention policy for the directory, that is, whether to retain files that were saved in previous runs (CLEAN_DIRECTORY = don't retain) and, if so, whether to overwrite any existing files if the same record fails again (CREATE_OVERWRITE) or rather add a sequence number to the new file name in order not to overwrite (ADD_ALL). | ||
maxSavedFailedRecordsPerRun | string^[0-9]*$ Default: "100" Sets a maximum number of files to save in the failed records directory per run. The job log will tell when the limit is reached. | ||
maxSavedFailedRecordsTotal | string^[0-9]*$ Default: "1000" Sets a maximum number of files to be saved in the failed records directory at any given time - as the sum of previously saved records (that were not cleaned up before this run) plus any new records added during the run.The job log will tell when the limit is reached. | ||
mailAddress | string Comma separated list of e-mail addresses that should receive notification on job completion. | ||
mailLevel | string Enum: "OK" "WARN" "ERROR" The minimum severity of a job's completion status that will trigger email notification. | ||
lastHarvestFinished | string Assigned by API. The date and time when the most recent harvest job with this configuration completed. | ||
initiallyHarvested | string Date and time, assigned by Harvester | ||
lastHarvestStarted | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration began. | ||
lastUpdated | string Assigned by Harvester. The date and time when this definition was last modified. | ||
nextHarvestSchedule | string The date and time when a job with this definition should be run (if job is enabled). | ||
amountHarvested | string^[0-9]*$ Assigned by API. Number of records harvested in last run. It seems this should really be an integer, but string is what the WSAPI gives us. | ||
message | string Assigned by API. Message summarising results of last run. | ||
type required | string Value: "oaiPmh" Indicates OAI-PMH job. | ||
metadataPrefix required | string OAI-PMH only. The metadata prefix supported by the OAI-PMH service to harvest from. | ||
oaiSetName required | string OAI-PMH only. The name of a record set offered by the OAI-PMH service to harvest from. | ||
resumptionToken | string OAI-PMH only. PMH identifier for fetching the next batch of records. | ||
clearRtOnError | string Default: "false" Enum: "true" "false" Clear the resumption token for harvests that complete in an error state. This is useful when server errors out and the last resumption token is no longer valid. | ||
fromDate | string yyyy-MM-dd. If empty and no resumption token is set, the Harvester will harvest the full data set from the resource. When this field contains a value, upon completion of the job the Harvester will reset the value of this field to the day prior to the current run date, so subsequent runs will harvest only new records. | ||
untilDate | string yyyy-MM-dd. Upper date limit for selective harvesting. On consecutive runs the Harvester will clear this field making the date interval open-ended. | ||
retryCount | string^[0-9]*$ Default: "2" Indicates how many times Harvester should retry a failed OAI-PMH request. | ||
retryWait | string^[0-9]*$ Default: "60" Indicates how many seconds Harvester should wait before retrying a failed OAI-PMH request. | ||
allowErrors | string Default: "false" Enum: "true" "false" NA for OAI-PMH | ||
json | object Custom configurations in JSON format (has no current applications). | ||
overwrite | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. Will delete all previously harvested data before beginning the next scheduled (or manually triggered) run, if set to true. | ||
keepPartial | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. When true, partial records harvested during a failed harvest run will be retained in Solr. |
{- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "id": "string"
}, - "transformation": {
- "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "storageBatchLimit": "string",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "lastHarvestFinished": "string",
- "initiallyHarvested": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "type": "oaiPmh",
- "metadataPrefix": "string",
- "oaiSetName": "string",
- "resumptionToken": "string",
- "clearRtOnError": "true",
- "fromDate": "string",
- "untilDate": "string",
- "retryCount": "2",
- "retryWait": "60",
- "allowErrors": "true",
- "json": { },
- "overwrite": "true",
- "keepPartial": "true"
}
{- "type": "xmlBulk",
- "allowErrors": "true",
- "overwrite": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "entityType": "inventoryStorageEntity",
- "bulkSize": "string",
- "currentStatus": "string",
- "customClass": "string",
- "enabled": "true",
- "idAsString": "string",
- "name": "string",
- "url": "string"
}, - "transformation": {
- "entityType": "basicTransformation",
- "acl": "string",
- "description": "string",
- "enabled": "true",
- "name": "string",
- "parallel": "true",
- "stepAssociations": [
- {
- "id": "string",
- "position": "string",
- "step": {
- "entityType": "xmlTransformationStep",
- "acl": "string",
- "description": "string",
- "inputFormat": "string",
- "name": "string",
- "outputFormat": "string",
- "script": "<'script' omitted from nested displays>",
- "id": "string",
- "testData": "<'testData' omitted from nested displays>",
- "testOutput": "<'testOutput' omitted from nested displays>"
}, - "transformation": "string"
}
], - "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "currentStatus": "NEW",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "acl": "string"
}
Get harvest configuration
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "type": "xmlBulk",
- "allowErrors": "true",
- "overwrite": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "entityType": "inventoryStorageEntity",
- "bulkSize": "string",
- "currentStatus": "string",
- "customClass": "string",
- "enabled": "true",
- "idAsString": "string",
- "name": "string",
- "url": "string"
}, - "transformation": {
- "entityType": "basicTransformation",
- "acl": "string",
- "description": "string",
- "enabled": "true",
- "name": "string",
- "parallel": "true",
- "stepAssociations": [
- {
- "id": "string",
- "position": "string",
- "step": {
- "entityType": "xmlTransformationStep",
- "acl": "string",
- "description": "string",
- "inputFormat": "string",
- "name": "string",
- "outputFormat": "string",
- "script": "<'script' omitted from nested displays>",
- "id": "string",
- "testData": "<'testData' omitted from nested displays>",
- "testOutput": "<'testOutput' omitted from nested displays>"
}, - "transformation": "string"
}
], - "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "currentStatus": "NEW",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "acl": "string"
}
Update harvest configuration
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string^[0-9]*$ Unique, numeric ID for the job definition. Will be assigned if not provided. | ||
name required | string The name assigned to the harvest configuration. | ||
description | string Free form description of the configuration to support the administration. | ||
openAccess | string Enum: "true" "false" tbd | ||
required | object or object Reference to the storage configuration to use. | ||
Any of
| |||
required | object or object Reference to the transformation pipeline to use. | ||
Any of
| |||
enabled required | string Enum: "true" "false" Indicates if the job is scheduled for running | ||
harvestImmediately required | string Enum: "true" "false" Whether to harvest when the config is persisted. | ||
scheduleString | string Crontab style schedule string (simplified): minute(0-59) hour(0-24) day of month(* or 1-31) month (* or 1-12) day of week (* or 0-6). | ||
dateFormat | string For example yyyy-MM-dd'T'hh:mm:ss'Z'. | ||
url required | string The URL to access the data from. | ||
timeout | string^[0-9]*$ Default: "300" Connection/read timeout in seconds; application depending on the specific protocol used for fetching data. | ||
cacheEnabled | string Enum: "true" "false" Whether or not to store incoming records in Harvester's file system. | ||
diskRun | string Enum: "true" "false" Whether or not to run harvest job from records cached in a previous job run. | ||
storageBatchLimit | string^[0-9]*$ Batch size: Number of records to send to storage at a time. | ||
recordLimit | string^[0-9]*$ Maximum number of records to harvest. | ||
laxParsing | string Default: "false" Enum: "true" "false" When enabled, Harvester will attempt to parse malformed XML (missing closing tags, entities) | ||
constantFields | string Values related to target handling in MasterKey. Otherwise obsolete. | ||
storeOriginal | string Default: "false" Enum: "true" "false" Indicates whether to store incoming original record, if supported by the job type and the storage configuration. | ||
managedBy | string Free-text field for tagging a job with the producer or manager of the resource. Multiple tags may be separated by commas. The tags can be used for filtering status reports by job administrators for example. | ||
usedBy | string Free form administrative information; could be tags for the clients using this harvestable. | ||
serviceProvider | string Free-text field for administrative information about the harvest job. | ||
contactNotes | string Free form text field for administrator's notes. | ||
technicalNotes | string Free-text field for administrative information. | ||
logLevel | string Enum: "ERROR" "WARN" "INFO" "DEBUG" "TRACE" Specifies the logging level for the job with TRACE being the most (extremely) verbose. INFO is the recommended log level in most cases. | ||
failedRecordsLogging | string Default: "CLEAN_DIRECTORY" Enum: "NO_STORE" "CLEAN_DIRECTORY" "CREATE_OVERWRITE" "ADD_ALL" Specify whether or not failed records should be saved as XML files in a designated log directory. Also specifies retention policy for the directory, that is, whether to retain files that were saved in previous runs (CLEAN_DIRECTORY = don't retain) and, if so, whether to overwrite any existing files if the same record fails again (CREATE_OVERWRITE) or rather add a sequence number to the new file name in order not to overwrite (ADD_ALL). | ||
maxSavedFailedRecordsPerRun | string^[0-9]*$ Default: "100" Sets a maximum number of files to save in the failed records directory per run. The job log will tell when the limit is reached. | ||
maxSavedFailedRecordsTotal | string^[0-9]*$ Default: "1000" Sets a maximum number of files to be saved in the failed records directory at any given time - as the sum of previously saved records (that were not cleaned up before this run) plus any new records added during the run.The job log will tell when the limit is reached. | ||
mailAddress | string Comma separated list of e-mail addresses that should receive notification on job completion. | ||
mailLevel | string Enum: "OK" "WARN" "ERROR" The minimum severity of a job's completion status that will trigger email notification. | ||
initiallyHarvested | string Date and time, assigned by Harvester | ||
lastHarvestFinished | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration completed. | ||
lastHarvestStarted | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration began. | ||
lastUpdated | string Assigned by API. The date and time when this definition was last modified. | ||
nextHarvestSchedule | string The date and time when a job with this definition should be run (if job is enabled). | ||
amountHarvested | string^[0-9]*$ Assigned by API. Number of records harvested in last run. It seems this should really be an integer, but string is what the WSAPI gives us. | ||
message | string Assigned by API. Message summarising results of last run. | ||
type required | string Value: "xmlBulk" Indicates bulk XML job. | ||
retryCount | string^[0-9]*$ Default: "2" Obsolete but allowed for XML bulk. | ||
retryWait | string^[0-9]*$ Default: "60" Obsolete but allowed for XML bulk. | ||
allowErrors | string Default: "false" Enum: "true" "false" Whether or not to continue despite harvest record errors. | ||
allowCondReq | string Default: "false" Enum: "true" "false" Whether or not to filter on file date to only harvest new XML files | ||
fromDate | string Initial start date (yyyy-MM-dd) for incremental updates (when allowCondReq is 'true') | ||
csvConfiguration | string Semicolon-separated key-value pairs that specifies parsing of a CSV file into XML for further processing (see Harvester documentation for details). | ||
excludeFilePattern | string Regular expression; setting to skip harvesting of files with names matching the given regular expression (see Harvester documentation for details). | ||
expectedSchema | string Mime-type override (e.g: application/marc; charset=MARC-8). | ||
includeFilePattern | string Regular expression; setting to request harvesting of files with names matching the given regular expression unless those file names are simultaneously excluded by the excludeFilePattern. .zip, .gz, .tar included by default unless explicitly excluded by excludeFilePattern (see Harvester documentation for details). | ||
outputSchema | string MARC XML transformation format (application/marc or application/tmarc). | ||
passiveMode | string Default: "false" Enum: "true" "false" Whether or not to use passive mode for FTP transfers. | ||
recurse | string Default: "false" Enum: "true" "false" Whether or not to recurse into sub-folders in the source directory tree. | ||
splitAt | string^[0-9]*$ Level/depth to split XML files at to extract records. Zero/empty disables split. | ||
splitSize | string^[0-9]*$ Setting to split large XML files into chunks of `splitSize' number of records; to preserve memory during XSLT transformations. | ||
json | object Custom configurations in JSON format (has no current applications). | ||
overwrite | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. Will delete all previously harvested data before beginning the next scheduled (or manually triggered) run, if set to true. | ||
keepPartial | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. When true, partial records harvested during a failed harvest run will be retained in Solr. |
{- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "id": "string"
}, - "transformation": {
- "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "storageBatchLimit": "string",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "initiallyHarvested": "string",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "type": "xmlBulk",
- "retryCount": "2",
- "retryWait": "60",
- "allowErrors": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "json": { },
- "overwrite": "true",
- "keepPartial": "true"
}
Update bulk XML harvest configuration
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string^[0-9]*$ Unique, numeric ID for the job definition. Will be assigned if not provided. | ||
name required | string The name assigned to the harvest configuration. | ||
description | string Free form description of the configuration to support the administration. | ||
openAccess | string Enum: "true" "false" tbd | ||
required | object or object Reference to the storage configuration to use. | ||
Any of
| |||
required | object or object Reference to the transformation pipeline to use. | ||
Any of
| |||
enabled required | string Enum: "true" "false" Indicates if the job is scheduled for running | ||
harvestImmediately required | string Enum: "true" "false" Whether to harvest when the config is persisted. | ||
scheduleString | string Crontab style schedule string (simplified): minute(0-59) hour(0-24) day of month(* or 1-31) month (* or 1-12) day of week (* or 0-6). | ||
dateFormat | string For example yyyy-MM-dd'T'hh:mm:ss'Z'. | ||
url required | string The URL to access the data from. | ||
timeout | string^[0-9]*$ Default: "300" Connection/read timeout in seconds; application depending on the specific protocol used for fetching data. | ||
cacheEnabled | string Enum: "true" "false" Whether or not to store incoming records in Harvester's file system. | ||
diskRun | string Enum: "true" "false" Whether or not to run harvest job from records cached in a previous job run. | ||
storageBatchLimit | string^[0-9]*$ Batch size: Number of records to send to storage at a time. | ||
recordLimit | string^[0-9]*$ Maximum number of records to harvest. | ||
laxParsing | string Default: "false" Enum: "true" "false" When enabled, Harvester will attempt to parse malformed XML (missing closing tags, entities) | ||
constantFields | string Values related to target handling in MasterKey. Otherwise obsolete. | ||
storeOriginal | string Default: "false" Enum: "true" "false" Indicates whether to store incoming original record, if supported by the job type and the storage configuration. | ||
managedBy | string Free-text field for tagging a job with the producer or manager of the resource. Multiple tags may be separated by commas. The tags can be used for filtering status reports by job administrators for example. | ||
usedBy | string Free form administrative information; could be tags for the clients using this harvestable. | ||
serviceProvider | string Free-text field for administrative information about the harvest job. | ||
contactNotes | string Free form text field for administrator's notes. | ||
technicalNotes | string Free-text field for administrative information. | ||
logLevel | string Enum: "ERROR" "WARN" "INFO" "DEBUG" "TRACE" Specifies the logging level for the job with TRACE being the most (extremely) verbose. INFO is the recommended log level in most cases. | ||
failedRecordsLogging | string Default: "CLEAN_DIRECTORY" Enum: "NO_STORE" "CLEAN_DIRECTORY" "CREATE_OVERWRITE" "ADD_ALL" Specify whether or not failed records should be saved as XML files in a designated log directory. Also specifies retention policy for the directory, that is, whether to retain files that were saved in previous runs (CLEAN_DIRECTORY = don't retain) and, if so, whether to overwrite any existing files if the same record fails again (CREATE_OVERWRITE) or rather add a sequence number to the new file name in order not to overwrite (ADD_ALL). | ||
maxSavedFailedRecordsPerRun | string^[0-9]*$ Default: "100" Sets a maximum number of files to save in the failed records directory per run. The job log will tell when the limit is reached. | ||
maxSavedFailedRecordsTotal | string^[0-9]*$ Default: "1000" Sets a maximum number of files to be saved in the failed records directory at any given time - as the sum of previously saved records (that were not cleaned up before this run) plus any new records added during the run.The job log will tell when the limit is reached. | ||
mailAddress | string Comma separated list of e-mail addresses that should receive notification on job completion. | ||
mailLevel | string Enum: "OK" "WARN" "ERROR" The minimum severity of a job's completion status that will trigger email notification. | ||
initiallyHarvested | string Date and time, assigned by Harvester | ||
lastHarvestFinished | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration completed. | ||
lastHarvestStarted | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration began. | ||
lastUpdated | string Assigned by API. The date and time when this definition was last modified. | ||
nextHarvestSchedule | string The date and time when a job with this definition should be run (if job is enabled). | ||
amountHarvested | string^[0-9]*$ Assigned by API. Number of records harvested in last run. It seems this should really be an integer, but string is what the WSAPI gives us. | ||
message | string Assigned by API. Message summarising results of last run. | ||
type required | string Value: "xmlBulk" Indicates bulk XML job. | ||
retryCount | string^[0-9]*$ Default: "2" Obsolete but allowed for XML bulk. | ||
retryWait | string^[0-9]*$ Default: "60" Obsolete but allowed for XML bulk. | ||
allowErrors | string Default: "false" Enum: "true" "false" Whether or not to continue despite harvest record errors. | ||
allowCondReq | string Default: "false" Enum: "true" "false" Whether or not to filter on file date to only harvest new XML files | ||
fromDate | string Initial start date (yyyy-MM-dd) for incremental updates (when allowCondReq is 'true') | ||
csvConfiguration | string Semicolon-separated key-value pairs that specifies parsing of a CSV file into XML for further processing (see Harvester documentation for details). | ||
excludeFilePattern | string Regular expression; setting to skip harvesting of files with names matching the given regular expression (see Harvester documentation for details). | ||
expectedSchema | string Mime-type override (e.g: application/marc; charset=MARC-8). | ||
includeFilePattern | string Regular expression; setting to request harvesting of files with names matching the given regular expression unless those file names are simultaneously excluded by the excludeFilePattern. .zip, .gz, .tar included by default unless explicitly excluded by excludeFilePattern (see Harvester documentation for details). | ||
outputSchema | string MARC XML transformation format (application/marc or application/tmarc). | ||
passiveMode | string Default: "false" Enum: "true" "false" Whether or not to use passive mode for FTP transfers. | ||
recurse | string Default: "false" Enum: "true" "false" Whether or not to recurse into sub-folders in the source directory tree. | ||
splitAt | string^[0-9]*$ Level/depth to split XML files at to extract records. Zero/empty disables split. | ||
splitSize | string^[0-9]*$ Setting to split large XML files into chunks of `splitSize' number of records; to preserve memory during XSLT transformations. | ||
json | object Custom configurations in JSON format (has no current applications). | ||
overwrite | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. Will delete all previously harvested data before beginning the next scheduled (or manually triggered) run, if set to true. | ||
keepPartial | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. When true, partial records harvested during a failed harvest run will be retained in Solr. |
{- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "id": "string"
}, - "transformation": {
- "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "storageBatchLimit": "string",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "initiallyHarvested": "string",
- "lastHarvestFinished": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "type": "xmlBulk",
- "retryCount": "2",
- "retryWait": "60",
- "allowErrors": "true",
- "allowCondReq": "true",
- "fromDate": "string",
- "csvConfiguration": "string",
- "excludeFilePattern": "string",
- "expectedSchema": "string",
- "includeFilePattern": "string",
- "outputSchema": "string",
- "passiveMode": "true",
- "recurse": "true",
- "splitAt": "string",
- "splitSize": "string",
- "json": { },
- "overwrite": "true",
- "keepPartial": "true"
}
Update OAI-PMH harvest configuration
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string^[0-9]*$ Unique, numeric ID for the job definition. Will be assigned if not provided. | ||
name required | string The name assigned to the harvest configuration. | ||
description | string Free form description of the configuration to support the administration. | ||
openAccess | string Enum: "true" "false" tbd | ||
required | object or object Reference to the storage configuration to use. | ||
Any of
| |||
required | object or object Reference to the transformation pipeline to use. | ||
Any of
| |||
enabled required | string Enum: "true" "false" Indicates if the job is scheduled for running | ||
harvestImmediately required | string Enum: "true" "false" Whether to harvest when the config is persisted. | ||
scheduleString | string Crontab style schedule string (simplified): minute(0-59) hour(0-24) day of month(* or 1-31) month (* or 1-12) day of week (* or 0-6). | ||
dateFormat | string For example yyyy-MM-dd'T'hh:mm:ss'Z'. | ||
url required | string The URL to access the data from. | ||
timeout | string^[0-9]*$ Default: "300" Connection/read timeout in seconds; application depending on the specific protocol used for fetching data. | ||
cacheEnabled | string Enum: "true" "false" Whether or not to store incoming records in Harvester's file system. | ||
diskRun | string Enum: "true" "false" Whether or not to run harvest job from records cached in a previous job run. | ||
storageBatchLimit | string^[0-9]*$ Batch size: Number of records to send to storage at a time. | ||
recordLimit | string^[0-9]*$ Maximum number of records to harvest. | ||
laxParsing | string Default: "false" Enum: "true" "false" When enabled, Harvester will attempt to parse malformed XML (missing closing tags, entities) | ||
constantFields | string Values related to target handling in MasterKey. Otherwise obsolete. | ||
storeOriginal | string Default: "false" Enum: "true" "false" Indicates whether to store incoming original record, if supported by the job type and the storage configuration. | ||
managedBy | string Free-text field for tagging a job with the producer or manager of the resource. Multiple tags may be separated by commas. The tags can be used for filtering status reports by job administrators for example. | ||
usedBy | string Free form administrative information; could be tags for the clients using this harvestable. | ||
serviceProvider | string Free-text field for administrative information about the harvest job. | ||
contactNotes | string Free form text field for administrator's notes. | ||
technicalNotes | string Free-text field for administrative information. | ||
logLevel | string Enum: "ERROR" "WARN" "INFO" "DEBUG" "TRACE" Specifies the logging level for the job with TRACE being the most (extremely) verbose. INFO is the recommended log level in most cases. | ||
failedRecordsLogging | string Default: "CLEAN_DIRECTORY" Enum: "NO_STORE" "CLEAN_DIRECTORY" "CREATE_OVERWRITE" "ADD_ALL" Specify whether or not failed records should be saved as XML files in a designated log directory. Also specifies retention policy for the directory, that is, whether to retain files that were saved in previous runs (CLEAN_DIRECTORY = don't retain) and, if so, whether to overwrite any existing files if the same record fails again (CREATE_OVERWRITE) or rather add a sequence number to the new file name in order not to overwrite (ADD_ALL). | ||
maxSavedFailedRecordsPerRun | string^[0-9]*$ Default: "100" Sets a maximum number of files to save in the failed records directory per run. The job log will tell when the limit is reached. | ||
maxSavedFailedRecordsTotal | string^[0-9]*$ Default: "1000" Sets a maximum number of files to be saved in the failed records directory at any given time - as the sum of previously saved records (that were not cleaned up before this run) plus any new records added during the run.The job log will tell when the limit is reached. | ||
mailAddress | string Comma separated list of e-mail addresses that should receive notification on job completion. | ||
mailLevel | string Enum: "OK" "WARN" "ERROR" The minimum severity of a job's completion status that will trigger email notification. | ||
lastHarvestFinished | string Assigned by API. The date and time when the most recent harvest job with this configuration completed. | ||
initiallyHarvested | string Date and time, assigned by Harvester | ||
lastHarvestStarted | string Assigned by Harvester. The date and time when the most recent harvest job with this configuration began. | ||
lastUpdated | string Assigned by Harvester. The date and time when this definition was last modified. | ||
nextHarvestSchedule | string The date and time when a job with this definition should be run (if job is enabled). | ||
amountHarvested | string^[0-9]*$ Assigned by API. Number of records harvested in last run. It seems this should really be an integer, but string is what the WSAPI gives us. | ||
message | string Assigned by API. Message summarising results of last run. | ||
type required | string Value: "oaiPmh" Indicates OAI-PMH job. | ||
metadataPrefix required | string OAI-PMH only. The metadata prefix supported by the OAI-PMH service to harvest from. | ||
oaiSetName required | string OAI-PMH only. The name of a record set offered by the OAI-PMH service to harvest from. | ||
resumptionToken | string OAI-PMH only. PMH identifier for fetching the next batch of records. | ||
clearRtOnError | string Default: "false" Enum: "true" "false" Clear the resumption token for harvests that complete in an error state. This is useful when server errors out and the last resumption token is no longer valid. | ||
fromDate | string yyyy-MM-dd. If empty and no resumption token is set, the Harvester will harvest the full data set from the resource. When this field contains a value, upon completion of the job the Harvester will reset the value of this field to the day prior to the current run date, so subsequent runs will harvest only new records. | ||
untilDate | string yyyy-MM-dd. Upper date limit for selective harvesting. On consecutive runs the Harvester will clear this field making the date interval open-ended. | ||
retryCount | string^[0-9]*$ Default: "2" Indicates how many times Harvester should retry a failed OAI-PMH request. | ||
retryWait | string^[0-9]*$ Default: "60" Indicates how many seconds Harvester should wait before retrying a failed OAI-PMH request. | ||
allowErrors | string Default: "false" Enum: "true" "false" NA for OAI-PMH | ||
json | object Custom configurations in JSON format (has no current applications). | ||
overwrite | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. Will delete all previously harvested data before beginning the next scheduled (or manually triggered) run, if set to true. | ||
keepPartial | string Enum: "true" "false" Applies to Solr but not FOLIO Inventory. When true, partial records harvested during a failed harvest run will be retained in Solr. |
{- "id": "string",
- "name": "string",
- "description": "string",
- "openAccess": "true",
- "storage": {
- "id": "string"
}, - "transformation": {
- "id": "string"
}, - "enabled": "true",
- "harvestImmediately": "true",
- "scheduleString": "string",
- "dateFormat": "string",
- "url": "string",
- "timeout": "300",
- "cacheEnabled": "true",
- "diskRun": "true",
- "storageBatchLimit": "string",
- "recordLimit": "string",
- "laxParsing": "true",
- "constantFields": "string",
- "storeOriginal": "true",
- "managedBy": "string",
- "usedBy": "string",
- "serviceProvider": "string",
- "contactNotes": "string",
- "technicalNotes": "string",
- "logLevel": "ERROR",
- "failedRecordsLogging": "NO_STORE",
- "maxSavedFailedRecordsPerRun": "100",
- "maxSavedFailedRecordsTotal": "1000",
- "mailAddress": "string",
- "mailLevel": "OK",
- "lastHarvestFinished": "string",
- "initiallyHarvested": "string",
- "lastHarvestStarted": "string",
- "lastUpdated": "string",
- "nextHarvestSchedule": "string",
- "amountHarvested": "string",
- "message": "string",
- "type": "oaiPmh",
- "metadataPrefix": "string",
- "oaiSetName": "string",
- "resumptionToken": "string",
- "clearRtOnError": "true",
- "fromDate": "string",
- "untilDate": "string",
- "retryCount": "2",
- "retryWait": "60",
- "allowErrors": "true",
- "json": { },
- "overwrite": "true",
- "keepPartial": "true"
}
Starts a harvest job immediately if possible
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "harvestableId": 0,
- "name": "string",
- "initiated": "string"
}
Get log statements for a harvest job
id required | number Harvest configuration identifier |
offset | string log file start line |
limit | string max log file lines |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
Get failed records for a harvest job
id required | number Harvest configuration identifier |
offset | string result set start row |
limit | string result set max rows |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "failedRecords": [
- {
- "recordErrors": [
- {
- "error": {
- "label": "string",
- "typeOfError": "string",
- "typeOfRecord": "string",
- "transaction": "string",
- "message": { },
- "entity": { }
}
}
], - "original": "string",
- "transformedRecord": { },
- "timeStamp": "string",
- "recordNumber": "string",
- "harvestableId": "string"
}
]
}
Get a failed record for a harvest job
id required | number Harvest configuration identifier |
num required | string number of a failed record |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "recordErrors": [
- {
- "error": {
- "label": "string",
- "typeOfError": "string",
- "typeOfRecord": "string",
- "transaction": "string",
- "message": { },
- "entity": { }
}
}
], - "original": "string",
- "transformedRecord": { },
- "timeStamp": "string",
- "recordNumber": "string",
- "harvestableId": "string"
}
Takes submitted job status, pulls the job config, and stores a copy of its most recent logs
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
status required | string Enum: "NEW" "OK" "WARN" "ERROR" "RUNNING" "FINISHED" "KILLED" The outcome of the harvester job according to Harvester. |
finished | string ISO formatted timestamp for when the job finished. |
started | string ISO formatted timestamp for when the job started. |
amountHarvested | string The number of records harvested in the harvest run. |
message required | string Status message for the outcome of the harvest run. |
{- "status": "NEW",
- "finished": "string",
- "started": "string",
- "amountHarvested": "string",
- "message": "string"
}
Pulls the current job config from Harvester and stores a copy of the most recent log for that job
id required | number Harvest configuration identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
Retrieves list of previous harvest jobs
query | string CQL query, supporting harvestableId, name, type, status, message, and amountHarvested in queries, and the same fields plus started and finished in sorting |
offset | string result set start row |
limit | string result set max rows |
from | string date range start parameter on finished date |
until | string date range end parameter on finished date |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "previousJobs": [
- {
- "id": "string",
- "name": "string",
- "harvestableId": 0,
- "type": "string",
- "url": "string",
- "allowErrors": true,
- "recordLimit": 0,
- "transformation": "string",
- "storage": "string",
- "status": "string",
- "started": "string",
- "finished": "string",
- "amountHarvested": 0,
- "message": "string"
}
]
}
Create job log samples for test purposes etc, for example by import from another FOLIO instance.
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string unique identifier for this report of a single harvest run - a UUID |
name | string The name of the harvest configuration at the time of logging the harvest run. |
harvestableId | integer Unique identifier for the harvest configuration. |
type | string The type of harvest job (bulk XML or OAI-PMH) |
url | string The URL(s) used for retrieving the records that were harvested during this job. |
allowErrors | boolean Indicates whether the job was configured to continue in case of (certain classes of) errors. |
recordLimit | integer Indicates the limit -- if any -- on the maximum number of records to load according to the configuration. |
transformation | string The name of the transformation pipeline that was used for the harvest job. |
storage | string The name of the storage that was used for persisting the records harvested during the job. |
status | string The outcome of the job. This would usually be the status after the job finished but it's possible to retrieve a history entry for a still running job. |
started | string Timestamp indicating when the job began. |
finished | string Timestamp indicating when the job completed. |
amountHarvested | integer The number of (incoming) records that were processed. |
message | string A description of the outcome of the harvest job, for example update statistics or a fatal error. |
{- "id": "string",
- "name": "string",
- "harvestableId": 0,
- "type": "string",
- "url": "string",
- "allowErrors": true,
- "recordLimit": 0,
- "transformation": "string",
- "storage": "string",
- "status": "string",
- "started": "string",
- "finished": "string",
- "amountHarvested": 0,
- "message": "string"
}
{- "id": "string",
- "name": "string",
- "harvestableId": 0,
- "type": "string",
- "url": "string",
- "allowErrors": true,
- "recordLimit": 0,
- "transformation": "string",
- "storage": "string",
- "status": "string",
- "started": "string",
- "finished": "string",
- "amountHarvested": 0,
- "message": "string"
}
Retrieves the failed records of previous harvest jobs
query | string CQL query, supporting recordNumber, harvestableId, harvestableName in queries |
offset | string result set start row |
limit | string result set max rows |
from | string date range parameter on error report timestamp |
until | string date range parameter on error report timestamp |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "failedRecords": [
- {
- "id": "string",
- "harvestJobId": "string",
- "harvestableId": 0,
- "harvestableName": "string",
- "recordErrors": [
- {
- "error": {
- "label": "string",
- "typeOfError": { },
- "typeOfRecord": "string",
- "transaction": "string",
- "message": { },
- "entity": { }
}
}
], - "original": "string",
- "transformedRecord": { },
- "timeStamp": "string",
- "recordNumber": "string"
}
]
}
Retrieves details of a previous harvest job
id required | string <uuid> Harvest job identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "id": "string",
- "name": "string",
- "harvestableId": 0,
- "type": "string",
- "url": "string",
- "allowErrors": true,
- "recordLimit": 0,
- "transformation": "string",
- "storage": "string",
- "status": "string",
- "started": "string",
- "finished": "string",
- "amountHarvested": 0,
- "message": "string"
}
Retrieves the log of a previous harvest job
id required | string <uuid> Harvest job identifier |
query | string CQL, supporting harvestJobId, logLevel, jobLabel, line in query terms |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
Backdoor for creating logs of a previous harvest job without running a job
id required | string <uuid> Harvest job identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
Retrieves the failed records of a previous harvest job
id required | string <uuid> Harvest job identifier |
query | string CQL query, supporting recordNumber, harvestableId, harvestableName in queries |
from | string date range parameter on error report timestamp |
until | string date range parameter on error report timestamp |
offset | string result set start row |
limit | string result set max rows |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "failedRecords": [
- {
- "id": "string",
- "harvestJobId": "string",
- "harvestableId": 0,
- "harvestableName": "string",
- "recordErrors": [
- {
- "error": {
- "label": "string",
- "typeOfError": { },
- "typeOfRecord": "string",
- "transaction": "string",
- "message": { },
- "entity": { }
}
}
], - "original": "string",
- "transformedRecord": { },
- "timeStamp": "string",
- "recordNumber": "string"
}
]
}
Create failed record samples without running a job, for example to import from another FOLIO instance.
id required | string <uuid> Harvest job identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
Array of objects (failedRecordPreviousJob) List of failed records created by a previous harvest job run. | |||||||||||||||||||
Array
|
{- "failedRecords": [
- {
- "id": "string",
- "harvestJobId": "string",
- "harvestableId": 0,
- "harvestableName": "string",
- "recordErrors": [
- {
- "error": {
- "label": "string",
- "typeOfError": { },
- "typeOfRecord": "string",
- "transaction": "string",
- "message": { },
- "entity": { }
}
}
], - "original": "string",
- "transformedRecord": { },
- "timeStamp": "string",
- "recordNumber": "string"
}
]
}
Retrieves a failed record of a previous harvest job
id required | string <uuid> UUID of the failed-record object |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
Create storage configuration
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
type | string Value: "inventoryStorage" Type of storage. |
json required | object Storage configuration parameters in JSON. |
id | string Unique storage identifier. |
name | string Name of the storage definition. |
description | string Free text details about the storage definition. |
enabled | string Default: "false" Enum: "true" "false" Boolean string to indicate if the storage definition can be used. |
url | string Address of the storage service. |
{- "type": "solrStorage",
- "id": "string",
- "name": "string",
- "description": "string",
- "enabled": "true",
- "url": "string"
}
{- "type": "solrStorage",
- "acl": "string",
- "id": "string",
- "name": "string",
- "description": "string",
- "enabled": "true",
- "bulkSize": "string",
- "currentStatus": "string",
- "url": "string"
}
Get brief storage definitions
query | string CQL |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "storages": [
- {
- "id": 0,
- "name": "string",
- "enabled": "true",
- "description": "string"
}
], - "totalRecords": 0
}
Get storage definition
id required | integer Storage definition identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "type": "solrStorage",
- "acl": "string",
- "id": "string",
- "name": "string",
- "description": "string",
- "enabled": "true",
- "bulkSize": "string",
- "currentStatus": "string",
- "url": "string"
}
Update storage definition
id required | integer Storage definition identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
type | string Value: "inventoryStorage" Type of storage. |
json required | object Storage configuration parameters in JSON. |
id | string Unique storage identifier. |
name | string Name of the storage definition. |
description | string Free text details about the storage definition. |
enabled | string Default: "false" Enum: "true" "false" Boolean string to indicate if the storage definition can be used. |
url | string Address of the storage service. |
{- "type": "solrStorage",
- "id": "string",
- "name": "string",
- "description": "string",
- "enabled": "true",
- "url": "string"
}
Create transformation pipeline
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string Unique record identifier. | ||||
name required | string Name of the transformation pipeline. | ||||
description | string Details about the pipeline. | ||||
type required | string Enum: "basicTransformation" "customTransformation" The type of transformation pipeline. | ||||
enabled | string Enum: "true" "false" Indicates if the transformation pipeline can be used by harvest jobs | ||||
parallel | string Default: "false" Enum: "true" "false" Indicates if steps should be run concurrently (each in its own thread). | ||||
Array of objects or objects or objects List of steps that make up the transformation pipeline. In a POST this will be used for attaching the steps to the pipeline. In a PUT this is ignored. | |||||
Array Any of
|
{- "id": "string",
- "name": "string",
- "description": "string",
- "type": "basicTransformation",
- "enabled": "true",
- "parallel": "true",
- "stepAssociations": [
- {
- "position": "string",
- "step": {
- "id": "string"
}
}
]
}
{- "id": "string",
- "name": "string",
- "description": "string",
- "type": "basicTransformation",
- "enabled": "true",
- "parallel": "true",
- "stepAssociations": [
- {
- "id": "string",
- "position": "string",
- "step": {
- "entityType": "xmlTransformationStep",
- "acl": "string",
- "description": "string",
- "inputFormat": "string",
- "name": "string",
- "outputFormat": "string",
- "script": "<'script' omitted from nested displays>",
- "customClass": "string",
- "id": "string",
- "testData": "<'testData' omitted from nested displays>",
- "testOutput": "<'testOutput' omitted from nested displays>"
}, - "transformation": "string"
}
]
}
Get brief transformation definitions
query | string CQL |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "transformations": [
- {
- "id": 0,
- "name": "string",
- "description": "string",
- "inputFormat": "string",
- "outputFormat": "string",
- "type": "string"
}
], - "totalRecords": 0
}
Get transformation pipeline
id required | integer Transformation pipeline identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "id": "string",
- "name": "string",
- "description": "string",
- "type": "basicTransformation",
- "enabled": "true",
- "parallel": "true",
- "stepAssociations": [
- {
- "id": "string",
- "position": "string",
- "step": {
- "entityType": "xmlTransformationStep",
- "acl": "string",
- "description": "string",
- "inputFormat": "string",
- "name": "string",
- "outputFormat": "string",
- "script": "<'script' omitted from nested displays>",
- "customClass": "string",
- "id": "string",
- "testData": "<'testData' omitted from nested displays>",
- "testOutput": "<'testOutput' omitted from nested displays>"
}, - "transformation": "string"
}
]
}
Create new transformation step definition
id | string Unique identifier for the transformation step. |
name required | string A name assigned to the transformation step. |
enabled | string Default: "false" Indicates if this step is available to be used in a transformation pipeline. |
description | string Additional descriptions of the storage definition. |
type required | string Enum: "XmlTransformStep" "CustomTransformStep" Type of transformation step. |
inputFormat | string Free-text indication of the format of input data to the step. |
outputFormat | string Free-text indication of the format of the resulting output from the step. |
testData | string Sample input data for testing. |
testOutput | string Output from testing using the sample test-data. |
customClass | string Only CustomTransformSteps: fully qualified class name of the class performing the transformation. |
script | string Transformation script, typically XSLT. |
{- "id": "string",
- "name": "string",
- "enabled": "false",
- "description": "string",
- "type": "XmlTransformStep",
- "inputFormat": "string",
- "outputFormat": "string",
- "testData": "string",
- "testOutput": "string",
- "customClass": "string",
- "script": "string"
}
{- "acl": "string",
- "id": "string",
- "name": "string",
- "enabled": "false",
- "description": "string",
- "type": "XmlTransformStep",
- "inputFormat": "string",
- "outputFormat": "string",
- "testData": "string",
- "testOutput": "string",
- "customClass": "string",
- "script": "string"
}
Get detailed transformation step definition record
id required | string Step identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "acl": "string",
- "id": "string",
- "name": "string",
- "enabled": "false",
- "description": "string",
- "type": "XmlTransformStep",
- "inputFormat": "string",
- "outputFormat": "string",
- "testData": "string",
- "testOutput": "string",
- "customClass": "string",
- "script": "string"
}
Update a transformation step definition
id required | string Step identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string Unique identifier for the transformation step. |
name required | string A name assigned to the transformation step. |
enabled | string Default: "false" Indicates if this step is available to be used in a transformation pipeline. |
description | string Additional descriptions of the storage definition. |
type required | string Enum: "XmlTransformStep" "CustomTransformStep" Type of transformation step. |
inputFormat | string Free-text indication of the format of input data to the step. |
outputFormat | string Free-text indication of the format of the resulting output from the step. |
testData | string Sample input data for testing. |
testOutput | string Output from testing using the sample test-data. |
customClass | string Only CustomTransformSteps: fully qualified class name of the class performing the transformation. |
script | string Transformation script, typically XSLT. |
{- "id": "string",
- "name": "string",
- "enabled": "false",
- "description": "string",
- "type": "XmlTransformStep",
- "inputFormat": "string",
- "outputFormat": "string",
- "testData": "string",
- "testOutput": "string",
- "customClass": "string",
- "script": "string"
}
{- "acl": "string",
- "id": "string",
- "name": "string",
- "enabled": "false",
- "description": "string",
- "type": "XmlTransformStep",
- "inputFormat": "string",
- "outputFormat": "string",
- "testData": "string",
- "testOutput": "string",
- "customClass": "string",
- "script": "string"
}
Update a transformation step script
id required | string Step identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
property name* additional property | any |
Create new transformation step association
id | string unique identifier for the association | ||||||
required | object or object contains id of step that is associated with a pipeline | ||||||
Any of
| |||||||
transformation required | string Id of the transformation pipeline that the step is associated with. | ||||||
transformationName | string Transient. Optional alternative to the id for looking up the transformation to attach the step to. | ||||||
position required | string The position of the step amongst other transformation steps in the pipeline. |
{- "id": "string",
- "step": {
- "id": "string",
- "name": "string"
}, - "transformation": "string",
- "transformationName": "string",
- "position": "string"
}
{- "id": "string",
- "step": {
- "id": "string",
- "name": "string"
}, - "transformation": "string",
- "transformationName": "string",
- "position": "string"
}
Get a transformation step association
id required | string Association identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
{- "id": "string",
- "step": {
- "id": "string",
- "name": "string"
}, - "transformation": "string",
- "transformationName": "string",
- "position": "string"
}
Update a transformation step association
id required | string Association identifier |
X-Okapi-Tenant | string Okapi Tenant |
X-Okapi-Token | string Okapi Token |
X-Okapi-Url | string Okapi URL |
id | string unique identifier for the association | ||||||
required | object or object contains id of step that is associated with a pipeline | ||||||
Any of
| |||||||
transformation required | string Id of the transformation pipeline that the step is associated with. | ||||||
transformationName | string Transient. Optional alternative to the id for looking up the transformation to attach the step to. | ||||||
position required | string The position of the step amongst other transformation steps in the pipeline. |
{- "id": "string",
- "step": {
- "id": "string",
- "name": "string"
}, - "transformation": "string",
- "transformationName": "string",
- "position": "string"
}
{- "id": "string",
- "step": {
- "id": "string",
- "name": "string"
}, - "transformation": "string",
- "transformationName": "string",
- "position": "string"
}