Skip to content

Commit 039abd7

Browse files
committed
Added seq2seq sample
1 parent 3d3e0aa commit 039abd7

File tree

2 files changed

+183
-0
lines changed

2 files changed

+183
-0
lines changed
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,98 @@
1+
# Copyright 2022 Google LLC
2+
#
3+
# Licensed under the Apache License, Version 2.0 (the "License");
4+
# you may not use this file except in compliance with the License.
5+
# You may obtain a copy of the License at
6+
#
7+
# https://www.apache.org/licenses/LICENSE-2.0
8+
#
9+
# Unless required by applicable law or agreed to in writing, software
10+
# distributed under the License is distributed on an "AS IS" BASIS,
11+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12+
# See the License for the specific language governing permissions and
13+
# limitations under the License.
14+
15+
from typing import List, Optional
16+
17+
from google.cloud import aiplatform
18+
19+
20+
# [START aiplatform_sdk_create_training_pipeline_forecasting_seq2seq_sample]
21+
def create_training_pipeline_forecasting_seq2seq_sample(
22+
project: str,
23+
display_name: str,
24+
dataset_id: str,
25+
location: str = "us-central1",
26+
model_display_name: str = None,
27+
target_column: str = "target_column",
28+
time_column: str = "date",
29+
time_series_identifier_column: str = "time_series_id",
30+
unavailable_at_forecast_columns: List[str] = [],
31+
available_at_forecast_columns: List[str] = [],
32+
forecast_horizon: int = 1,
33+
data_granularity_unit: str = "week",
34+
data_granularity_count: int = 1,
35+
training_fraction_split: float = 0.8,
36+
validation_fraction_split: float = 0.1,
37+
test_fraction_split: float = 0.1,
38+
budget_milli_node_hours: int = 8000,
39+
timestamp_split_column_name: str = "timestamp_split",
40+
weight_column: str = "weight",
41+
time_series_attribute_columns: List[str] = [],
42+
context_window: int = 0,
43+
export_evaluated_data_items: bool = False,
44+
export_evaluated_data_items_bigquery_destination_uri: Optional[str] = None,
45+
export_evaluated_data_items_override_destination: bool = False,
46+
quantiles: Optional[List[float]] = None,
47+
validation_options: Optional[str] = None,
48+
predefined_split_column_name: Optional[str] = None,
49+
sync: bool = True,
50+
):
51+
aiplatform.init(project=project, location=location)
52+
53+
# Create training job
54+
forecasting_seq2seq_job = aiplatform.SequenceToSequencePlusForecastingTrainingJob(
55+
display_name=display_name, optimization_objective="minimize-rmse"
56+
)
57+
58+
# Retrieve existing dataset
59+
dataset = aiplatform.TimeSeriesDataset(dataset_id)
60+
61+
# Run training job
62+
model = forecasting_seq2seq_job.run(
63+
dataset=dataset,
64+
target_column=target_column,
65+
time_column=time_column,
66+
time_series_identifier_column=time_series_identifier_column,
67+
unavailable_at_forecast_columns=unavailable_at_forecast_columns,
68+
available_at_forecast_columns=available_at_forecast_columns,
69+
forecast_horizon=forecast_horizon,
70+
data_granularity_unit=data_granularity_unit,
71+
data_granularity_count=data_granularity_count,
72+
training_fraction_split=training_fraction_split,
73+
validation_fraction_split=validation_fraction_split,
74+
test_fraction_split=test_fraction_split,
75+
predefined_split_column_name=predefined_split_column_name,
76+
timestamp_split_column_name=timestamp_split_column_name,
77+
weight_column=weight_column,
78+
time_series_attribute_columns=time_series_attribute_columns,
79+
context_window=context_window,
80+
export_evaluated_data_items=export_evaluated_data_items,
81+
export_evaluated_data_items_bigquery_destination_uri=export_evaluated_data_items_bigquery_destination_uri,
82+
export_evaluated_data_items_override_destination=export_evaluated_data_items_override_destination,
83+
quantiles=quantiles,
84+
validation_options=validation_options,
85+
budget_milli_node_hours=budget_milli_node_hours,
86+
model_display_name=model_display_name,
87+
sync=sync,
88+
)
89+
90+
model.wait()
91+
92+
print(model.display_name)
93+
print(model.resource_name)
94+
print(model.uri)
95+
return model
96+
97+
98+
# [END aiplatform_sdk_create_training_pipeline_forecasting_seq2seq_sample]
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,85 @@
1+
# Copyright 2022 Google LLC
2+
#
3+
# Licensed under the Apache License, Version 2.0 (the "License");
4+
# you may not use this file except in compliance with the License.
5+
# You may obtain a copy of the License at
6+
#
7+
# https://www.apache.org/licenses/LICENSE-2.0
8+
#
9+
# Unless required by applicable law or agreed to in writing, software
10+
# distributed under the License is distributed on an "AS IS" BASIS,
11+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12+
# See the License for the specific language governing permissions and
13+
# limitations under the License.
14+
15+
16+
import create_training_pipeline_forecasting_seq2seq_sample
17+
import test_constants as constants
18+
19+
20+
def test_create_training_pipeline_forecasting_seq2seq_sample(
21+
mock_sdk_init,
22+
mock_time_series_dataset,
23+
mock_get_automl_forecasting_seq2seq_training_job,
24+
mock_run_automl_forecasting_seq2seq_training_job,
25+
mock_get_time_series_dataset,
26+
):
27+
28+
create_training_pipeline_forecasting_seq2seq_sample.create_training_pipeline_forecasting_seq2seq_sample(
29+
project=constants.PROJECT,
30+
display_name=constants.DISPLAY_NAME,
31+
dataset_id=constants.RESOURCE_ID,
32+
model_display_name=constants.DISPLAY_NAME_2,
33+
target_column=constants.TABULAR_TARGET_COLUMN,
34+
training_fraction_split=constants.TRAINING_FRACTION_SPLIT,
35+
validation_fraction_split=constants.VALIDATION_FRACTION_SPLIT,
36+
test_fraction_split=constants.TEST_FRACTION_SPLIT,
37+
budget_milli_node_hours=constants.BUDGET_MILLI_NODE_HOURS_8000,
38+
timestamp_split_column_name=constants.TIMESTAMP_SPLIT_COLUMN_NAME,
39+
weight_column=constants.WEIGHT_COLUMN,
40+
time_series_attribute_columns=constants.TIME_SERIES_ATTRIBUTE_COLUMNS,
41+
context_window=constants.CONTEXT_WINDOW,
42+
export_evaluated_data_items=constants.EXPORT_EVALUATED_DATA_ITEMS,
43+
export_evaluated_data_items_bigquery_destination_uri=constants.EXPORT_EVALUATED_DATA_ITEMS_BIGQUERY_DESTINATION_URI,
44+
export_evaluated_data_items_override_destination=constants.EXPORT_EVALUATED_DATA_ITEMS_OVERRIDE_DESTINATION,
45+
quantiles=constants.QUANTILES,
46+
validation_options=constants.VALIDATION_OPTIONS,
47+
predefined_split_column_name=constants.PREDEFINED_SPLIT_COLUMN_NAME,
48+
)
49+
50+
mock_get_time_series_dataset.assert_called_once_with(constants.RESOURCE_ID)
51+
52+
mock_sdk_init.assert_called_once_with(
53+
project=constants.PROJECT, location=constants.LOCATION
54+
)
55+
mock_get_automl_forecasting_seq2seq_training_job.assert_called_once_with(
56+
display_name=constants.DISPLAY_NAME,
57+
optimization_objective="minimize-rmse",
58+
)
59+
mock_run_automl_forecasting_seq2seq_training_job.assert_called_once_with(
60+
dataset=mock_time_series_dataset,
61+
target_column=constants.TABULAR_TARGET_COLUMN,
62+
time_column=constants.FORECASTNG_TIME_COLUMN,
63+
time_series_identifier_column=constants.FORECASTNG_TIME_SERIES_IDENTIFIER_COLUMN,
64+
unavailable_at_forecast_columns=constants.FORECASTNG_UNAVAILABLE_AT_FORECAST_COLUMNS,
65+
available_at_forecast_columns=constants.FORECASTNG_AVAILABLE_AT_FORECAST_COLUMNS,
66+
forecast_horizon=constants.FORECASTNG_FORECAST_HORIZON,
67+
data_granularity_unit=constants.DATA_GRANULARITY_UNIT,
68+
data_granularity_count=constants.DATA_GRANULARITY_COUNT,
69+
training_fraction_split=constants.TRAINING_FRACTION_SPLIT,
70+
validation_fraction_split=constants.VALIDATION_FRACTION_SPLIT,
71+
test_fraction_split=constants.TEST_FRACTION_SPLIT,
72+
budget_milli_node_hours=constants.BUDGET_MILLI_NODE_HOURS_8000,
73+
model_display_name=constants.DISPLAY_NAME_2,
74+
timestamp_split_column_name=constants.TIMESTAMP_SPLIT_COLUMN_NAME,
75+
weight_column=constants.WEIGHT_COLUMN,
76+
time_series_attribute_columns=constants.TIME_SERIES_ATTRIBUTE_COLUMNS,
77+
context_window=constants.CONTEXT_WINDOW,
78+
export_evaluated_data_items=constants.EXPORT_EVALUATED_DATA_ITEMS,
79+
export_evaluated_data_items_bigquery_destination_uri=constants.EXPORT_EVALUATED_DATA_ITEMS_BIGQUERY_DESTINATION_URI,
80+
export_evaluated_data_items_override_destination=constants.EXPORT_EVALUATED_DATA_ITEMS_OVERRIDE_DESTINATION,
81+
quantiles=constants.QUANTILES,
82+
validation_options=constants.VALIDATION_OPTIONS,
83+
predefined_split_column_name=constants.PREDEFINED_SPLIT_COLUMN_NAME,
84+
sync=True,
85+
)

0 commit comments

Comments
 (0)