Index
AutoMlForecasting
(message)AutoMlForecastingInputs
(message)AutoMlForecastingInputs.Granularity
(message)AutoMlForecastingInputs.Transformation
(message)AutoMlForecastingInputs.Transformation.AutoTransformation
(message)AutoMlForecastingInputs.Transformation.CategoricalTransformation
(message)AutoMlForecastingInputs.Transformation.NumericTransformation
(message)AutoMlForecastingInputs.Transformation.TextTransformation
(message)AutoMlForecastingInputs.Transformation.TimestampTransformation
(message)AutoMlForecastingMetadata
(message)AutoMlImageClassification
(message)AutoMlImageClassificationInputs
(message)AutoMlImageClassificationInputs.ModelType
(enum)AutoMlImageClassificationMetadata
(message)AutoMlImageClassificationMetadata.SuccessfulStopReason
(enum)AutoMlImageObjectDetection
(message)AutoMlImageObjectDetectionInputs
(message)AutoMlImageObjectDetectionInputs.ModelType
(enum)AutoMlImageObjectDetectionMetadata
(message)AutoMlImageObjectDetectionMetadata.SuccessfulStopReason
(enum)AutoMlImageSegmentation
(message)AutoMlImageSegmentationInputs
(message)AutoMlImageSegmentationInputs.ModelType
(enum)AutoMlImageSegmentationMetadata
(message)AutoMlImageSegmentationMetadata.SuccessfulStopReason
(enum)AutoMlTables
(message)AutoMlTablesInputs
(message)AutoMlTablesInputs.Transformation
(message)AutoMlTablesInputs.Transformation.AutoTransformation
(message)AutoMlTablesInputs.Transformation.CategoricalArrayTransformation
(message)AutoMlTablesInputs.Transformation.CategoricalTransformation
(message)AutoMlTablesInputs.Transformation.NumericArrayTransformation
(message)AutoMlTablesInputs.Transformation.NumericTransformation
(message)AutoMlTablesInputs.Transformation.TextArrayTransformation
(message)AutoMlTablesInputs.Transformation.TextTransformation
(message)AutoMlTablesInputs.Transformation.TimestampTransformation
(message)AutoMlTablesMetadata
(message)AutoMlTextClassification
(message)AutoMlTextClassificationInputs
(message)AutoMlTextExtraction
(message)AutoMlTextExtractionInputs
(message)AutoMlTextSentiment
(message)AutoMlTextSentimentInputs
(message)AutoMlVideoActionRecognition
(message)AutoMlVideoActionRecognitionInputs
(message)AutoMlVideoActionRecognitionInputs.ModelType
(enum)AutoMlVideoClassification
(message)AutoMlVideoClassificationInputs
(message)AutoMlVideoClassificationInputs.ModelType
(enum)AutoMlVideoObjectTracking
(message)AutoMlVideoObjectTrackingInputs
(message)AutoMlVideoObjectTrackingInputs.ModelType
(enum)CustomJobMetadata
(message)CustomTask
(message)ExportEvaluatedDataItemsConfig
(message)HierarchyConfig
(message)HyperparameterTuningJobMetadata
(message)HyperparameterTuningJobSpec
(message)HyperparameterTuningTask
(message)Seq2SeqPlusForecasting
(message)Seq2SeqPlusForecastingInputs
(message)Seq2SeqPlusForecastingInputs.Granularity
(message)Seq2SeqPlusForecastingInputs.Transformation
(message)Seq2SeqPlusForecastingInputs.Transformation.AutoTransformation
(message)Seq2SeqPlusForecastingInputs.Transformation.CategoricalTransformation
(message)Seq2SeqPlusForecastingInputs.Transformation.NumericTransformation
(message)Seq2SeqPlusForecastingInputs.Transformation.TextTransformation
(message)Seq2SeqPlusForecastingInputs.Transformation.TimestampTransformation
(message)Seq2SeqPlusForecastingMetadata
(message)WindowConfig
(message)
AutoMlForecasting
A TrainingJob that trains and uploads an AutoML Forecasting Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
metadata |
The metadata information. |
AutoMlForecastingInputs
Fields | |
---|---|
target_ |
The name of the column that the Model is to predict values for. This column must be unavailable at forecast. |
time_ |
The name of the column that identifies the time series. |
time_ |
The name of the column that identifies time order in the time series. This column must be available at forecast. |
transformations[] |
Each transformation will apply transform function to given input column. And the result will be used for training. When creating transformation for BigQuery Struct column, the column should be flattened using "." as the delimiter. |
optimization_ |
Objective function the model is optimizing towards. The training process creates a model that optimizes the value of the objective function over the validation set. The supported optimization objectives:
|
train_ |
Required. The train budget of creating this model, expressed in milli node hours i.e. 1,000 value in this field means 1 node hour. The training cost of the model will not exceed this budget. The final cost will be attempted to be close to the budget, though may end up being (even) noticeably smaller - at the backend's discretion. This especially may happen when further model training ceases to provide any improvements. If the budget is set to a value known to be insufficient to train a model for the given dataset, the training won't be attempted and will error. The train budget must be between 1,000 and 72,000 milli node hours, inclusive. |
weight_ |
Column name that should be used as the weight column. Higher values in this column give more importance to the row during model training. The column must have numeric values between 0 and 10000 inclusively; 0 means the row is ignored for training. If weight column field is not set, then all rows are assumed to have equal weight of 1. |
time_ |
Column names that should be used as attribute columns. The value of these columns does not vary as a function of time. For example, store ID or item color. |
unavailable_ |
Names of columns that are unavailable when a forecast is requested. This column contains information for the given entity (identified by the time_series_identifier_column) that is unknown before the forecast For example, actual weather on a given day. |
available_ |
Names of columns that are available and provided when a forecast is requested. These columns contain information for the given entity (identified by the time_series_identifier_column column) that is known at forecast. For example, predicted weather for a specific day. |
data_ |
Expected difference in time granularity between rows in the data. |
forecast_ |
The amount of time into the future for which forecasted values for the target are returned. Expressed in number of units defined by the |
context_ |
The amount of time into the past training and prediction data is used for model training and prediction respectively. Expressed in number of units defined by the |
export_ |
Configuration for exporting test set predictions to a BigQuery table. If this configuration is absent, then the export is not performed. |
quantiles[] |
Quantiles to use for minimize-quantile-loss |
hierarchy_ |
Configuration that defines the hierarchical relationship of time series and parameters for hierarchical forecasting strategies. |
window_ |
Config containing strategy for generating sliding windows. |
holiday_ |
The geographical region based on which the holiday effect is applied in modeling by adding holiday categorical array feature that include all holidays matching the date. This option only allowed when data_granularity is day. By default, holiday effect modeling is disabled. To turn it on, specify the holiday region using this option. |
enable_ |
If probabilistic inference is enabled, the model will fit a distribution that captures the uncertainty of a prediction. At inference time, the predictive distribution is used to make a point prediction that minimizes the optimization objective. For example, the mean of a predictive distribution is the point prediction that minimizes RMSE loss. If quantiles are specified, then the quantiles of the distribution are also returned. The optimization objective cannot be minimize-quantile-loss. |
validation_ |
Validation options for the data validation component. The available options are:
|
additional_ |
Additional experiment flags for the time series forcasting training. |
Granularity
A duration of time expressed in time granularity units.
Fields | |
---|---|
unit |
The time granularity unit of this time period. The supported units are:
|
quantity |
The number of granularity_units between data points in the training data. If |
Transformation
Fields | |
---|---|
Union field transformation_detail . The transformation that the training pipeline will apply to the input columns. transformation_detail can be only one of the following: |
|
auto |
|
numeric |
|
categorical |
|
timestamp |
|
text |
AutoTransformation
Training pipeline will infer the proper transformation based on the statistic of dataset.
Fields | |
---|---|
column_ |
|
CategoricalTransformation
Training pipeline will perform following transformation functions.
The categorical string as is--no change to case, punctuation, spelling, tense, and so on.
Convert the category name to a dictionary lookup index and generate an embedding for each index.
Categories that appear less than 5 times in the training dataset are treated as the "unknown" category. The "unknown" category gets its own special lookup index and resulting embedding.
Fields | |
---|---|
column_ |
|
NumericTransformation
Training pipeline will perform following transformation functions.
The value converted to float32.
The z_score of the value.
log(value+1) when the value is greater than or equal to 0. Otherwise, this transformation is not applied and the value is considered a missing value.
z_score of log(value+1) when the value is greater than or equal to 0. Otherwise, this transformation is not applied and the value is considered a missing value.
A boolean value that indicates whether the value is valid.
Fields | |
---|---|
column_ |
|
TextTransformation
Training pipeline will perform following transformation functions.
The text as is--no change to case, punctuation, spelling, tense, and so on.
Convert the category name to a dictionary lookup index and generate an embedding for each index.
Fields | |
---|---|
column_ |
|
TimestampTransformation
Training pipeline will perform following transformation functions.
Apply the transformation functions for Numerical columns.
Determine the year, month, day,and weekday. Treat each value from the timestamp as a Categorical column.
Invalid numerical values (for example, values that fall outside of a typical timestamp range, or are extreme values) receive no special treatment and are not removed.
Fields | |
---|---|
column_ |
|
time_ |
The format in which that time field is expressed. The time_format must either be one of:
(for respectively number of seconds, milliseconds, microseconds and nanoseconds since start of the Unix epoch); or be written in If time_format is not set, then the default format is RFC 3339 |
AutoMlForecastingMetadata
Model metadata specific to AutoML Forecasting.
Fields | |
---|---|
train_ |
Output only. The actual training cost of the model, expressed in milli node hours, i.e. 1,000 value in this field means 1 node hour. Guaranteed to not exceed the train budget. |
evaluated_ |
BigQuery destination uri for exported evaluated examples. |
AutoMlImageClassification
A TrainingJob that trains and uploads an AutoML Image Classification Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
metadata |
The metadata information. |
AutoMlImageClassificationInputs
Fields | |
---|---|
model_ |
|
base_ |
The ID of the |
budget_ |
The training budget of creating this model, expressed in milli node hours i.e. 1,000 value in this field means 1 node hour. The actual metadata.costMilliNodeHours will be equal or less than this value. If further model training ceases to provide any improvements, it will stop without using the full budget and the metadata.successfulStopReason will be |
disable_ |
Use the entire training budget. This disables the early stopping feature. When false the early stopping feature is enabled, which means that AutoML Image Classification might stop training before the entire training budget has been used. |
multi_ |
If false, a single-label (multi-class) Model will be trained (i.e. assuming that for each image just up to one annotation may be applicable). If true, a multi-label Model will be trained (i.e. assuming that for each image multiple annotations may be applicable). |
uptrain_ |
The ID of |
ModelType
Enums | |
---|---|
MODEL_TYPE_UNSPECIFIED |
Should not be set. |
CLOUD |
A Model best tailored to be used within Google Cloud, and which cannot be exported. Default. |
CLOUD_1 |
A model type best tailored to be used within Google Cloud, which cannot be exported externally. Compared to the CLOUD model above, it is expected to have higher prediction accuracy. |
MOBILE_TF_LOW_LATENCY_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow or Core ML model and used on a mobile or edge device afterwards. Expected to have low latency, but may have lower prediction quality than other mobile models. |
MOBILE_TF_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow or Core ML model and used on a mobile or edge device with afterwards. |
MOBILE_TF_HIGH_ACCURACY_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow or Core ML model and used on a mobile or edge device afterwards. Expected to have a higher latency, but should also have a higher prediction quality than other mobile models. |
AutoMlImageClassificationMetadata
Fields | |
---|---|
cost_ |
The actual training cost of creating this model, expressed in milli node hours, i.e. 1,000 value in this field means 1 node hour. Guaranteed to not exceed inputs.budgetMilliNodeHours. |
successful_ |
For successful job completions, this is the reason why the job has finished. |
SuccessfulStopReason
Enums | |
---|---|
SUCCESSFUL_STOP_REASON_UNSPECIFIED |
Should not be set. |
BUDGET_REACHED |
The inputs.budgetMilliNodeHours had been reached. |
MODEL_CONVERGED |
Further training of the Model ceased to increase its quality, since it already has converged. |
AutoMlImageObjectDetection
A TrainingJob that trains and uploads an AutoML Image Object Detection Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
metadata |
The metadata information |
AutoMlImageObjectDetectionInputs
Fields | |
---|---|
model_ |
|
budget_ |
The training budget of creating this model, expressed in milli node hours i.e. 1,000 value in this field means 1 node hour. The actual metadata.costMilliNodeHours will be equal or less than this value. If further model training ceases to provide any improvements, it will stop without using the full budget and the metadata.successfulStopReason will be |
disable_ |
Use the entire training budget. This disables the early stopping feature. When false the early stopping feature is enabled, which means that AutoML Image Object Detection might stop training before the entire training budget has been used. |
uptrain_ |
The ID of |
ModelType
Enums | |
---|---|
MODEL_TYPE_UNSPECIFIED |
Should not be set. |
CLOUD_HIGH_ACCURACY_1 |
A model best tailored to be used within Google Cloud, and which cannot be exported. Expected to have a higher latency, but should also have a higher prediction quality than other cloud models. |
CLOUD_LOW_LATENCY_1 |
A model best tailored to be used within Google Cloud, and which cannot be exported. Expected to have a low latency, but may have lower prediction quality than other cloud models. |
CLOUD_1 |
A model best tailored to be used within Google Cloud, and which cannot be exported. Compared to the CLOUD_HIGH_ACCURACY_1 and CLOUD_LOW_LATENCY_1 models above, it is expected to have higher prediction quality and lower latency. |
MOBILE_TF_LOW_LATENCY_1 |
A model that, in addition to being available within Google Cloud can also be exported (see ModelService.ExportModel) and used on a mobile or edge device with TensorFlow afterwards. Expected to have low latency, but may have lower prediction quality than other mobile models. |
MOBILE_TF_VERSATILE_1 |
A model that, in addition to being available within Google Cloud can also be exported (see ModelService.ExportModel) and used on a mobile or edge device with TensorFlow afterwards. |
MOBILE_TF_HIGH_ACCURACY_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) and used on a mobile or edge device with TensorFlow afterwards. Expected to have a higher latency, but should also have a higher prediction quality than other mobile models. |
CLOUD_STREAMING_1 |
A model best tailored to be used within Google Cloud, and which cannot be exported. Expected to best support predictions in streaming with lower latency and lower prediction quality than other cloud models. |
AutoMlImageObjectDetectionMetadata
Fields | |
---|---|
cost_ |
The actual training cost of creating this model, expressed in milli node hours, i.e. 1,000 value in this field means 1 node hour. Guaranteed to not exceed inputs.budgetMilliNodeHours. |
successful_ |
For successful job completions, this is the reason why the job has finished. |
SuccessfulStopReason
Enums | |
---|---|
SUCCESSFUL_STOP_REASON_UNSPECIFIED |
Should not be set. |
BUDGET_REACHED |
The inputs.budgetMilliNodeHours had been reached. |
MODEL_CONVERGED |
Further training of the Model ceased to increase its quality, since it already has converged. |
AutoMlImageSegmentation
A TrainingJob that trains and uploads an AutoML Image Segmentation Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
metadata |
The metadata information. |
AutoMlImageSegmentationInputs
Fields | |
---|---|
model_ |
|
budget_ |
The training budget of creating this model, expressed in milli node hours i.e. 1,000 value in this field means 1 node hour. The actual metadata.costMilliNodeHours will be equal or less than this value. If further model training ceases to provide any improvements, it will stop without using the full budget and the metadata.successfulStopReason will be |
base_ |
The ID of the |
ModelType
Enums | |
---|---|
MODEL_TYPE_UNSPECIFIED |
Should not be set. |
CLOUD_HIGH_ACCURACY_1 |
A model to be used via prediction calls to uCAIP API. Expected to have a higher latency, but should also have a higher prediction quality than other models. |
CLOUD_LOW_ACCURACY_1 |
A model to be used via prediction calls to uCAIP API. Expected to have a lower latency but relatively lower prediction quality. |
MOBILE_TF_LOW_LATENCY_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as TensorFlow model and used on a mobile or edge device afterwards. Expected to have low latency, but may have lower prediction quality than other mobile models. |
AutoMlImageSegmentationMetadata
Fields | |
---|---|
cost_ |
The actual training cost of creating this model, expressed in milli node hours, i.e. 1,000 value in this field means 1 node hour. Guaranteed to not exceed inputs.budgetMilliNodeHours. |
successful_ |
For successful job completions, this is the reason why the job has finished. |
SuccessfulStopReason
Enums | |
---|---|
SUCCESSFUL_STOP_REASON_UNSPECIFIED |
Should not be set. |
BUDGET_REACHED |
The inputs.budgetMilliNodeHours had been reached. |
MODEL_CONVERGED |
Further training of the Model ceased to increase its quality, since it already has converged. |
AutoMlTables
A TrainingJob that trains and uploads an AutoML Tables Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
metadata |
The metadata information. |
AutoMlTablesInputs
Fields | |
---|---|
prediction_ |
The type of prediction the Model is to produce. "classification" - Predict one out of multiple target values is picked for each row. "regression" - Predict a value based on its relation to other values. This type is available only to columns that contain semantically numeric values, i.e. integers or floating point number, even if stored as e.g. strings. |
target_ |
The column name of the target column that the model is to predict. |
transformations[] |
Each transformation will apply transform function to given input column. And the result will be used for training. When creating transformation for BigQuery Struct column, the column should be flattened using "." as the delimiter. |
optimization_ |
Objective function the model is optimizing towards. The training process creates a model that maximizes/minimizes the value of the objective function over the validation set. The supported optimization objectives depend on the prediction type. If the field is not set, a default objective function is used. classification (binary): "maximize-au-roc" (default) - Maximize the area under the receiver operating characteristic (ROC) curve. "minimize-log-loss" - Minimize log loss. "maximize-au-prc" - Maximize the area under the precision-recall curve. "maximize-precision-at-recall" - Maximize precision for a specified recall value. "maximize-recall-at-precision" - Maximize recall for a specified precision value. classification (multi-class): "minimize-log-loss" (default) - Minimize log loss. regression: "minimize-rmse" (default) - Minimize root-mean-squared error (RMSE). "minimize-mae" - Minimize mean-absolute error (MAE). "minimize-rmsle" - Minimize root-mean-squared log error (RMSLE). |
train_ |
Required. The train budget of creating this model, expressed in milli node hours i.e. 1,000 value in this field means 1 node hour. The training cost of the model will not exceed this budget. The final cost will be attempted to be close to the budget, though may end up being (even) noticeably smaller - at the backend's discretion. This especially may happen when further model training ceases to provide any improvements. If the budget is set to a value known to be insufficient to train a model for the given dataset, the training won't be attempted and will error. The train budget must be between 1,000 and 72,000 milli node hours, inclusive. |
disable_ |
Use the entire training budget. This disables the early stopping feature. By default, the early stopping feature is enabled, which means that AutoML Tables might stop training before the entire training budget has been used. |
weight_ |
Column name that should be used as the weight column. Higher values in this column give more importance to the row during model training. The column must have numeric values between 0 and 10000 inclusively; 0 means the row is ignored for training. If weight column field is not set, then all rows are assumed to have equal weight of 1. |
export_ |
Configuration for exporting test set predictions to a BigQuery table. If this configuration is absent, then the export is not performed. |
additional_ |
Additional experiment flags for the Tables training pipeline. |
Union field additional_optimization_objective_config . Additional optimization objective configuration. Required for maximize-precision-at-recall and maximize-recall-at-precision , otherwise unused. additional_optimization_objective_config can be only one of the following: |
|
optimization_ |
Required when optimization_objective is "maximize-precision-at-recall". Must be between 0 and 1, inclusive. |
optimization_ |
Required when optimization_objective is "maximize-recall-at-precision". Must be between 0 and 1, inclusive. |
Transformation
Fields | |
---|---|
Union field transformation_detail . The transformation that the training pipeline will apply to the input columns. transformation_detail can be only one of the following: |
|
auto |
|
numeric |
|
categorical |
|
timestamp |
|
text |
|
repeated_ |
|
repeated_ |
|
repeated_ |
AutoTransformation
Training pipeline will infer the proper transformation based on the statistic of dataset.
Fields | |
---|---|
column_ |
|
CategoricalArrayTransformation
Treats the column as categorical array and performs following transformation functions. * For each element in the array, convert the category name to a dictionary lookup index and generate an embedding for each index. Combine the embedding of all elements into a single embedding using the mean. * Empty arrays treated as an embedding of zeroes.
Fields | |
---|---|
column_ |
|
CategoricalTransformation
Training pipeline will perform following transformation functions. * The categorical string as is--no change to case, punctuation, spelling, tense, and so on. * Convert the category name to a dictionary lookup index and generate an embedding for each index. * Categories that appear less than 5 times in the training dataset are treated as the "unknown" category. The "unknown" category gets its own special lookup index and resulting embedding.
Fields | |
---|---|
column_ |
|
NumericArrayTransformation
Treats the column as numerical array and performs following transformation functions. * All transformations for Numerical types applied to the average of the all elements. * The average of empty arrays is treated as zero.
Fields | |
---|---|
column_ |
|
invalid_ |
If invalid values is allowed, the training pipeline will create a boolean feature that indicated whether the value is valid. Otherwise, the training pipeline will discard the input row from trainining data. |
NumericTransformation
Training pipeline will perform following transformation functions. * The value converted to float32. * The z_score of the value. * log(value+1) when the value is greater than or equal to 0. Otherwise, this transformation is not applied and the value is considered a missing value. * z_score of log(value+1) when the value is greater than or equal to 0. Otherwise, this transformation is not applied and the value is considered a missing value. * A boolean value that indicates whether the value is valid.
Fields | |
---|---|
column_ |
|
invalid_ |
If invalid values is allowed, the training pipeline will create a boolean feature that indicated whether the value is valid. Otherwise, the training pipeline will discard the input row from trainining data. |
TextArrayTransformation
Treats the column as text array and performs following transformation functions. * Concatenate all text values in the array into a single text value using a space (" ") as a delimiter, and then treat the result as a single text value. Apply the transformations for Text columns. * Empty arrays treated as an empty text.
Fields | |
---|---|
column_ |
|
TextTransformation
Training pipeline will perform following transformation functions. * The text as is--no change to case, punctuation, spelling, tense, and so on. * Tokenize text to words. Convert each words to a dictionary lookup index and generate an embedding for each index. Combine the embedding of all elements into a single embedding using the mean. * Tokenization is based on unicode script boundaries. * Missing values get their own lookup index and resulting embedding. * Stop-words receive no special treatment and are not removed.
Fields | |
---|---|
column_ |
|
TimestampTransformation
Training pipeline will perform following transformation functions. * Apply the transformation functions for Numerical columns. * Determine the year, month, day,and weekday. Treat each value from the * timestamp as a Categorical column. * Invalid numerical values (for example, values that fall outside of a typical timestamp range, or are extreme values) receive no special treatment and are not removed.
Fields | |
---|---|
column_ |
|
time_ |
The format in which that time field is expressed. The time_format must either be one of: * |
invalid_ |
If invalid values is allowed, the training pipeline will create a boolean feature that indicated whether the value is valid. Otherwise, the training pipeline will discard the input row from trainining data. |
AutoMlTablesMetadata
Model metadata specific to AutoML Tables.
Fields | |
---|---|
train_ |
Output only. The actual training cost of the model, expressed in milli node hours, i.e. 1,000 value in this field means 1 node hour. Guaranteed to not exceed the train budget. |
evaluated_ |
BigQuery destination uri for exported evaluated examples. |
AutoMlTextClassification
A TrainingJob that trains and uploads an AutoML Text Classification Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
AutoMlTextClassificationInputs
Fields | |
---|---|
multi_ |
|
AutoMlTextExtraction
A TrainingJob that trains and uploads an AutoML Text Extraction Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
AutoMlTextExtractionInputs
This type has no fields.
AutoMlTextSentiment
A TrainingJob that trains and uploads an AutoML Text Sentiment Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
AutoMlTextSentimentInputs
Fields | |
---|---|
sentiment_ |
A sentiment is expressed as an integer ordinal, where higher value means a more positive sentiment. The range of sentiments that will be used is between 0 and sentimentMax (inclusive on both ends), and all the values in the range must be represented in the dataset before a model can be created. Only the Annotations with this sentimentMax will be used for training. sentimentMax value must be between 1 and 10 (inclusive). |
AutoMlVideoActionRecognition
A TrainingJob that trains and uploads an AutoML Video Action Recognition Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
AutoMlVideoActionRecognitionInputs
Fields | |
---|---|
model_ |
ModelType
Enums | |
---|---|
MODEL_TYPE_UNSPECIFIED |
Should not be set. |
CLOUD |
A model best tailored to be used within Google Cloud, and which c annot be exported. Default. |
MOBILE_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as a TensorFlow or TensorFlow Lite model and used on a mobile or edge device afterwards. |
MOBILE_JETSON_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) to a Jetson device afterwards. |
MOBILE_CORAL_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as a TensorFlow or TensorFlow Lite model and used on a Coral device afterwards. |
AutoMlVideoClassification
A TrainingJob that trains and uploads an AutoML Video Classification Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
AutoMlVideoClassificationInputs
Fields | |
---|---|
model_ |
ModelType
Enums | |
---|---|
MODEL_TYPE_UNSPECIFIED |
Should not be set. |
CLOUD |
A model best tailored to be used within Google Cloud, and which cannot be exported. Default. |
MOBILE_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as a TensorFlow or TensorFlow Lite model and used on a mobile or edge device afterwards. |
MOBILE_JETSON_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) to a Jetson device afterwards. |
AutoMlVideoObjectTracking
A TrainingJob that trains and uploads an AutoML Video ObjectTracking Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
AutoMlVideoObjectTrackingInputs
Fields | |
---|---|
model_ |
ModelType
Enums | |
---|---|
MODEL_TYPE_UNSPECIFIED |
Should not be set. |
CLOUD |
A model best tailored to be used within Google Cloud, and which c annot be exported. Default. |
MOBILE_VERSATILE_1 |
A model that, in addition to being available within Google Cloud, can also be exported (see ModelService.ExportModel) as a TensorFlow or TensorFlow Lite model and used on a mobile or edge device afterwards. |
MOBILE_CORAL_VERSATILE_1 |
A versatile model that is meant to be exported (see ModelService.ExportModel) and used on a Google Coral device. |
MOBILE_CORAL_LOW_LATENCY_1 |
A model that trades off quality for low latency, to be exported (see ModelService.ExportModel) and used on a Google Coral device. |
MOBILE_JETSON_VERSATILE_1 |
A versatile model that is meant to be exported (see ModelService.ExportModel) and used on an NVIDIA Jetson device. |
MOBILE_JETSON_LOW_LATENCY_1 |
A model that trades off quality for low latency, to be exported (see ModelService.ExportModel) and used on an NVIDIA Jetson device. |
CustomJobMetadata
Fields | |
---|---|
backing_ |
The resource name of the CustomJob that has been created to carry out this custom task. |
CustomTask
A TrainingJob that trains a custom code Model.
Fields | |
---|---|
inputs |
The input parameters of this CustomTask. |
metadata |
The metadata information. |
ExportEvaluatedDataItemsConfig
Configuration for exporting test set predictions to a BigQuery table.
Fields | |
---|---|
destination_ |
URI of desired destination BigQuery table. Expected format: If not specified, then results are exported to the following auto-created BigQuery table: |
override_ |
If true and an export destination is specified, then the contents of the destination are overwritten. Otherwise, if the export destination already exists, then the export operation fails. |
HierarchyConfig
Configuration that defines the hierarchical relationship of time series and parameters for hierarchical forecasting strategies.
Fields | |
---|---|
group_ |
A list of time series attribute column names that define the time series hierarchy. Only one level of hierarchy is supported, ex. 'region' for a hierarchy of stores or 'department' for a hierarchy of products. If multiple columns are specified, time series will be grouped by their combined values, ex. ('blue', 'large') for 'color' and 'size', up to 5 columns are accepted. If no group columns are specified, all time series are considered to be part of the same group. |
group_ |
The weight of the loss for predictions aggregated over time series in the same group. |
temporal_ |
The weight of the loss for predictions aggregated over the horizon for a single time series. |
group_ |
The weight of the loss for predictions aggregated over both the horizon and time series in the same hierarchy group. |
HyperparameterTuningJobMetadata
Fields | |
---|---|
backing_ |
The resource name of the HyperparameterTuningJob that has been created to carry out this HyperparameterTuning task. |
best_ |
The resource name of the CustomJob that has been created to run the best Trial of this HyperparameterTuning task. |
HyperparameterTuningJobSpec
Fields | |
---|---|
study_ |
Study configuration of the HyperparameterTuningJob. |
trial_ |
The spec of a trial job. The same spec applies to the CustomJobs created in all the trials. |
max_ |
The desired total number of Trials. |
parallel_ |
The desired number of Trials to run in parallel. |
max_ |
The number of failed Trials that need to be seen before failing the HyperparameterTuningJob. If set to 0, Vertex AI decides how many Trials must fail before the whole job fails. |
HyperparameterTuningTask
A TrainingJob that tunes Hypererparameters of a custom code Model.
Fields | |
---|---|
inputs |
The input parameters of this HyperparameterTuningTask. |
metadata |
The metadata information. |
Seq2SeqPlusForecasting
A TrainingJob that trains and uploads an AutoML Forecasting Model.
Fields | |
---|---|
inputs |
The input parameters of this TrainingJob. |
metadata |
The metadata information. |
Seq2SeqPlusForecastingInputs
Fields | |
---|---|
target_ |
The name of the column that the Model is to predict values for. This column must be unavailable at forecast. |
time_ |
The name of the column that identifies the time series. |
time_ |
The name of the column that identifies time order in the time series. This column must be available at forecast. |
transformations[] |
Each transformation will apply transform function to given input column. And the result will be used for training. When creating transformation for BigQuery Struct column, the column should be flattened using "." as the delimiter. |
optimization_ |
Objective function the model is optimizing towards. The training process creates a model that optimizes the value of the objective function over the validation set. The supported optimization objectives:
|
train_ |
Required. The train budget of creating this model, expressed in milli node hours i.e. 1,000 value in this field means 1 node hour. The training cost of the model will not exceed this budget. The final cost will be attempted to be close to the budget, though may end up being (even) noticeably smaller - at the backend's discretion. This especially may happen when further model training ceases to provide any improvements. If the budget is set to a value known to be insufficient to train a model for the given dataset, the training won't be attempted and will error. The train budget must be between 1,000 and 72,000 milli node hours, inclusive. |
weight_ |
Column name that should be used as the weight column. Higher values in this column give more importance to the row during model training. The column must have numeric values between 0 and 10000 inclusively; 0 means the row is ignored for training. If weight column field is not set, then all rows are assumed to have equal weight of 1. This column must be available at forecast. |
time_ |
Column names that should be used as attribute columns. The value of these columns does not vary as a function of time. For example, store ID or item color. |
unavailable_ |
Names of columns that are unavailable when a forecast is requested. This column contains information for the given entity (identified by the time_series_identifier_column) that is unknown before the forecast For example, actual weather on a given day. |
available_ |
Names of columns that are available and provided when a forecast is requested. These columns contain information for the given entity (identified by the time_series_identifier_column column) that is known at forecast. For example, predicted weather for a specific day. |
data_ |
Expected difference in time granularity between rows in the data. |
forecast_ |
The amount of time into the future for which forecasted values for the target are returned. Expressed in number of units defined by the |
context_ |
The amount of time into the past training and prediction data is used for model training and prediction respectively. Expressed in number of units defined by the |
holiday_ |
The geographical region based on which the holiday effect is applied in modeling by adding holiday categorical array feature that include all holidays matching the date. This option only allowed when data_granularity is day. By default, holiday effect modeling is disabled. To turn it on, specify the holiday region using this option. |
export_ |
Configuration for exporting test set predictions to a BigQuery table. If this configuration is absent, then the export is not performed. |
window_ |
Config containing strategy for generating sliding windows. |
quantiles[] |
Quantiles to use for minimize-quantile-loss |
validation_ |
Validation options for the data validation component. The available options are:
|
additional_ |
Additional experiment flags for the time series forcasting training. |
hierarchy_ |
Configuration that defines the hierarchical relationship of time series and parameters for hierarchical forecasting strategies. |
Granularity
A duration of time expressed in time granularity units.
Fields | |
---|---|
unit |
The time granularity unit of this time period. The supported units are:
|
quantity |
The number of granularity_units between data points in the training data. If |
Transformation
Fields | |
---|---|
Union field transformation_detail . The transformation that the training pipeline will apply to the input columns. transformation_detail can be only one of the following: |
|
auto |
|
numeric |
|
categorical |
|
timestamp |
|
text |
AutoTransformation
Training pipeline will infer the proper transformation based on the statistic of dataset.
Fields | |
---|---|
column_ |
|
CategoricalTransformation
Training pipeline will perform following transformation functions.
The categorical string as is--no change to case, punctuation, spelling, tense, and so on.
Convert the category name to a dictionary lookup index and generate an embedding for each index.
Categories that appear less than 5 times in the training dataset are treated as the "unknown" category. The "unknown" category gets its own special lookup index and resulting embedding.
Fields | |
---|---|
column_ |
|
NumericTransformation
Training pipeline will perform following transformation functions.
The value converted to float32.
The z_score of the value.
log(value+1) when the value is greater than or equal to 0. Otherwise, this transformation is not applied and the value is considered a missing value.
z_score of log(value+1) when the value is greater than or equal to 0. Otherwise, this transformation is not applied and the value is considered a missing value.
Fields | |
---|---|
column_ |
|
TextTransformation
Training pipeline will perform following transformation functions.
The text as is--no change to case, punctuation, spelling, tense, and so on.
Convert the category name to a dictionary lookup index and generate an embedding for each index.
Fields | |
---|---|
column_ |
|
TimestampTransformation
Training pipeline will perform following transformation functions.
Apply the transformation functions for Numerical columns.
Determine the year, month, day,and weekday. Treat each value from the timestamp as a Categorical column.
Invalid numerical values (for example, values that fall outside of a typical timestamp range, or are extreme values) receive no special treatment and are not removed.
Fields | |
---|---|
column_ |
|
time_ |
The format in which that time field is expressed. The time_format must either be one of:
(for respectively number of seconds, milliseconds, microseconds and nanoseconds since start of the Unix epoch); or be written in If time_format is not set, then the default format is RFC 3339 |
Seq2SeqPlusForecastingMetadata
Model metadata specific to Seq2Seq Plus Forecasting.
Fields | |
---|---|
train_ |
Output only. The actual training cost of the model, expressed in milli node hours, i.e. 1,000 value in this field means 1 node hour. Guaranteed to not exceed the train budget. |
evaluated_ |
BigQuery destination uri for exported evaluated examples. |
WindowConfig
Config that contains the strategy used to generate sliding windows in time series training. A window is a series of rows that comprise the context up to the time of prediction, and the horizon following. The corresponding row for each window marks the start of the forecast horizon. Each window is used as an input example for training/evaluation.
Fields | |
---|---|
Union field
|
|
column |
Name of the column that should be used to generate sliding windows. The column should contain either booleans or string booleans; if the value of the row is True, generate a sliding window with the horizon starting at that row. The column will not be used as a feature in training. |
stride_ |
Stride length used to generate input examples. Within one time series, every {$STRIDE_LENGTH} rows will be used to generate a sliding window. |
max_ |
Maximum number of windows that should be generated across all time series. |