google_cloud_pipeline_components.experimental.forecasting package¶
Google Cloud Pipeline Experimental Forecasting Components.
- google_cloud_pipeline_components.experimental.forecasting.ForecastingPrepareDataForTrainOp(input_tables: list, preprocess_metadata: dict, model_feature_columns: list = None)¶
Prepare data for train Prepares the parameters for the training step.
Converts the input_tables and the output of ForecastingPreprocessingOp to the input parameters of TimeSeriesDatasetCreateOp and AutoMLForecastingTrainingJobRunOp.
- Args:
- input_tables (JsonArray):
Required. Serialized Json array that specifies input BigQuery tables and specs.
- preprocess_metadata (JsonObject):
Required. The output of ForecastingPreprocessingOp that is a serialized dictionary with 2 fields: processed_bigquery_table_uri and column_metadata.
- model_feature_columns (JsonArray):
Optional. Serialized list of column names that will be used as input feature in the training step. If None, all columns will be used in training.
- Returns:
- NamedTuple:
- time_series_identifier_column (String):
Name of the column that identifies the time series.
- time_series_attribute_columns (JsonArray):
Serialized column names that should be used as attribute columns.
- available_at_forecast_columns (JsonArray):
Serialized column names of columns that are available at forecast.
- unavailable_at_forecast_columns (JsonArray):
Serialized column names of columns that are unavailable at forecast.
- column_transformations (JsonArray):
Serialized transformations to apply to the input columns.
- preprocess_bq_uri (String):
The BigQuery table that saves the preprocessing result and will be used as training input.
- target_column (String):
The name of the column values of which the Model is to predict.
- time_column (String):
Name of the column that identifies time order in the time series.
- predefined_split_column (String):
Name of the column that specifies an ML use of the row.
- weight_column (String):
Name of the column that should be used as the weight column.
- data_granularity_unit (String):
The data granularity unit.
- data_granularity_count (Integer):
The number of data granularity units between data points in the training data.
- google_cloud_pipeline_components.experimental.forecasting.ForecastingPreprocessingOp()¶
forecasting_preprocessing Preprocesses BigQuery tables for training or prediction.
Creates a BigQuery table for training or prediction based on the input tables. For training, a primary table is required. Optionally, you can include some attribute tables. For prediction, you need to include all the tables that were used in the training, plus a plan table.
- Args:
- project (String):
The GCP project id that runs the pipeline.
- input_tables (JsonArray):
Serialized Json array that specifies input BigQuery tables and specs.
- preprocessing_bigquery_dataset (String):
Optional BigQuery dataset to save the preprocessing result BigQuery table. If not present, a new dataset will be created by the component.
- location (String):
Optional location for the BigQuery data, default is US.
- Returns:
preprocess_metadata (JsonObject)
- google_cloud_pipeline_components.experimental.forecasting.ForecastingValidationOp()¶
forecasting_validation Validates BigQuery tables for training or prediction.
Validates BigQuery tables for training or prediction based on predefined requirements. For training, a primary table is required. Optionally, you can include some attribute tables. For prediction, you need to include all the tables that were used in the training, plus a plan table.
- Args:
- input_tables (JsonArray):
Serialized Json array that specifies input BigQuery tables and specs.
- validation_theme (String):
Theme to use for validating the BigQuery tables. Acceptable values are FORECASTING_TRAINING and FORECASTING_PREDICTION.
- location (String):
Optional location for the BigQuery data, default is US.
- Returns:
None