Skip to main content

batch_prediction_jobs

Overview

Namebatch_prediction_jobs
TypeResource
Idgoogle.aiplatform.batch_prediction_jobs

Fields

NameDatatypeDescription
namestringOutput only. Resource name of the BatchPredictionJob.
completionStatsobjectSuccess and error statistics of processing multiple entities (for example, DataItems or structured data rows) in batch.
createTimestringOutput only. Time when the BatchPredictionJob was created.
dedicatedResourcesobjectA description of resources that are used for performing batch operations, are dedicated to a Model, and need manual configuration.
disableContainerLoggingbooleanFor custom-trained Models and AutoML Tabular Models, the container of the DeployedModel instances will send stderr and stdout streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to Cloud Logging pricing. User can disable container logging by setting this flag to true.
displayNamestringRequired. The user-defined name of this BatchPredictionJob.
encryptionSpecobjectRepresents a customer-managed encryption key spec that can be applied to a top-level resource.
endTimestringOutput only. Time when the BatchPredictionJob entered any of the following states: JOB_STATE_SUCCEEDED, JOB_STATE_FAILED, JOB_STATE_CANCELLED.
errorobjectThe Status type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by gRPC. Each Status message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the API Design Guide.
explanationSpecobjectSpecification of Model explanation.
generateExplanationbooleanGenerate explanation with the batch prediction results. When set to true, the batch prediction output changes based on the predictions_format field of the BatchPredictionJob.output_config object: bigquery: output includes a column named explanation. The value is a struct that conforms to the Explanation object. jsonl: The JSON objects on each line include an additional entry keyed explanation. The value of the entry is a JSON object that conforms to the Explanation object. * csv: Generating explanations for CSV format is not supported. If this field is set to true, either the Model.explanation_spec or explanation_spec must be populated.
inputConfigobjectConfigures the input to BatchPredictionJob. See Model.supported_input_storage_formats for Model's supported input formats, and how instances should be expressed via any of them.
instanceConfigobjectConfiguration defining how to transform batch prediction input instances to the instances that the Model accepts.
labelsobjectThe labels with user-defined metadata to organize BatchPredictionJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
manualBatchTuningParametersobjectManual batch tuning parameters.
modelstringThe name of the Model resource that produces the predictions via this job, must share the same ancestor Location. Starting this job has no impact on any existing deployments of the Model and their resources. Exactly one of model and unmanaged_container_model must be set. The model resource name may contain version id or version alias to specify the version. Example: projects/{project}/locations/{location}/models/{model}@2 or projects/{project}/locations/{location}/models/{model}@golden if no version is specified, the default version will be deployed. The model resource could also be a publisher model. Example: publishers/{publisher}/models/{model} or projects/{project}/locations/{location}/publishers/{publisher}/models/{model}
modelParametersanyThe parameters that govern the predictions. The schema of the parameters may be specified via the Model's PredictSchemata's parameters_schema_uri.
modelVersionIdstringOutput only. The version ID of the Model that produces the predictions via this job.
outputConfigobjectConfigures the output of BatchPredictionJob. See Model.supported_output_storage_formats for supported output formats, and how predictions are expressed via any of them.
outputInfoobjectFurther describes this job's output. Supplements output_config.
partialFailuresarrayOutput only. Partial failures encountered. For example, single files that can't be read. This field never exceeds 20 entries. Status details fields contain standard Google Cloud error details.
resourcesConsumedobjectStatistics information about resource consumption.
serviceAccountstringThe service account that the DeployedModel's container runs as. If not specified, a system generated one will be used, which has minimal permissions and the custom container, if used, may not have enough permission to access other Google Cloud resources. Users deploying the Model must have the iam.serviceAccounts.actAs permission on this service account.
startTimestringOutput only. Time when the BatchPredictionJob for the first time entered the JOB_STATE_RUNNING state.
statestringOutput only. The detailed state of the job.
unmanagedContainerModelobjectContains model information necessary to perform batch prediction without requiring a full model import.
updateTimestringOutput only. Time when the BatchPredictionJob was most recently updated.

Methods

NameAccessible byRequired ParamsDescription
getSELECTbatchPredictionJobsId, locationsId, projectsIdGets a BatchPredictionJob
listSELECTlocationsId, projectsIdLists BatchPredictionJobs in a Location.
createINSERTlocationsId, projectsIdCreates a BatchPredictionJob. A BatchPredictionJob once created will right away be attempted to start.
deleteDELETEbatchPredictionJobsId, locationsId, projectsIdDeletes a BatchPredictionJob. Can only be called on jobs that already finished.
_listEXEClocationsId, projectsIdLists BatchPredictionJobs in a Location.
cancelEXECbatchPredictionJobsId, locationsId, projectsIdCancels a BatchPredictionJob. Starts asynchronous cancellation on the BatchPredictionJob. The server makes the best effort to cancel the job, but success is not guaranteed. Clients can use JobService.GetBatchPredictionJob or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On a successful cancellation, the BatchPredictionJob is not deleted;instead its BatchPredictionJob.state is set to CANCELLED. Any files already outputted by the job are not deleted.