google_cloud_pipeline_components.experimental.forecasting package

Google Cloud Pipeline Experimental Forecasting Components.

google_cloud_pipeline_components.experimental.forecasting.ForecastingPrepareDataForTrainOp(input_tables: list, preprocess_metadata: dict, model_feature_columns: list = None)

Prepare data for train Prepares the parameters for the training step.

Converts the input_tables and the output of ForecastingPreprocessingOp to the input parameters of TimeSeriesDatasetCreateOp and AutoMLForecastingTrainingJobRunOp.

Args:
input_tables (JsonArray):

Required. Serialized Json array that specifies input BigQuery tables and specs.

preprocess_metadata (JsonObject):

Required. The output of ForecastingPreprocessingOp that is a serialized dictionary with 2 fields: processed_bigquery_table_uri and column_metadata.

model_feature_columns (JsonArray):

Optional. Serialized list of column names that will be used as input feature in the training step. If None, all columns will be used in training.

Returns:
NamedTuple:
time_series_identifier_column (String):

Name of the column that identifies the time series.

time_series_attribute_columns (JsonArray):

Serialized column names that should be used as attribute columns.

available_at_forecast_columns (JsonArray):

Serialized column names of columns that are available at forecast.

unavailable_at_forecast_columns (JsonArray):

Serialized column names of columns that are unavailable at forecast.

column_transformations (JsonArray):

Serialized transformations to apply to the input columns.

preprocess_bq_uri (String):

The BigQuery table that saves the preprocessing result and will be used as training input.

target_column (String):

The name of the column values of which the Model is to predict.

time_column (String):

Name of the column that identifies time order in the time series.

predefined_split_column (String):

Name of the column that specifies an ML use of the row.

weight_column (String):

Name of the column that should be used as the weight column.

data_granularity_unit (String):

The data granularity unit.

data_granularity_count (Integer):

The number of data granularity units between data points in the training data.

google_cloud_pipeline_components.experimental.forecasting.ForecastingPreprocessingOp()

forecasting_preprocessing Preprocesses BigQuery tables for training or prediction.

Creates a BigQuery table for training or prediction based on the input tables. For training, a primary table is required. Optionally, you can include some attribute tables. For prediction, you need to include all the tables that were used in the training, plus a plan table.

Args:
project (String):

The GCP project id that runs the pipeline.

input_tables (JsonArray):

Serialized Json array that specifies input BigQuery tables and specs.

preprocessing_bigquery_dataset (String):

Optional BigQuery dataset to save the preprocessing result BigQuery table. If not present, a new dataset will be created by the component.

location (String):

Optional location for the BigQuery data, default is US.

Returns:

preprocess_metadata (JsonObject)

google_cloud_pipeline_components.experimental.forecasting.ForecastingValidationOp()

forecasting_validation Validates BigQuery tables for training or prediction.

Validates BigQuery tables for training or prediction based on predefined requirements. For training, a primary table is required. Optionally, you can include some attribute tables. For prediction, you need to include all the tables that were used in the training, plus a plan table.

Args:
input_tables (JsonArray):

Serialized Json array that specifies input BigQuery tables and specs.

validation_theme (String):

Theme to use for validating the BigQuery tables. Acceptable values are FORECASTING_TRAINING and FORECASTING_PREDICTION.

location (String):

Optional location for the BigQuery data, default is US.

Returns:

None