| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| from __future__ import absolute_import |
|
|
| import base64 |
| import os |
| import requests |
|
|
| import docker |
| import numpy |
| import pytest |
| from botocore.exceptions import ClientError |
|
|
| from sagemaker import utils |
| from sagemaker.amazon.randomcutforest import RandomCutForest |
| from sagemaker.deserializers import StringDeserializer |
| from sagemaker.multidatamodel import MultiDataModel |
| from sagemaker.mxnet import MXNet |
| from sagemaker.predictor import Predictor |
| from sagemaker.serializers import NumpySerializer |
| from sagemaker.utils import sagemaker_timestamp, unique_name_from_base |
| from tests.integ import DATA_DIR, TRAINING_DEFAULT_TIMEOUT_MINUTES |
| from tests.integ.retry import retries |
| from tests.integ.timeout import timeout, timeout_and_delete_endpoint_by_name |
|
|
| ROLE = "SageMakerRole" |
| PRETRAINED_MODEL_PATH_1 = "customer_a/dummy_model.tar.gz" |
| PRETRAINED_MODEL_PATH_2 = "customer_b/dummy_model.tar.gz" |
| string_deserializer = StringDeserializer() |
|
|
|
|
| @pytest.fixture(scope="module") |
| def container_image(sagemaker_session): |
| """Create a Multi-Model image since pre-built ones are not available yet.""" |
| algorithm_name = unique_name_from_base("sagemaker-multimodel-integ-test") |
| ecr_image = _ecr_image_uri(sagemaker_session, algorithm_name) |
|
|
| ecr_client = sagemaker_session.boto_session.client("ecr") |
| username, password = _ecr_login(ecr_client) |
|
|
| docker_client = docker.from_env() |
|
|
| |
| image, build_log = docker_client.images.build( |
| path=os.path.join(DATA_DIR, "multimodel", "container"), |
| tag=algorithm_name, |
| rm=True, |
| ) |
| image.tag(ecr_image, tag="latest") |
|
|
| |
| _create_repository(ecr_client, algorithm_name) |
|
|
| |
| for _ in retries(3, "Upload docker image to ECR repo", seconds_to_sleep=10): |
| try: |
| docker_client.images.push( |
| ecr_image, auth_config={"username": username, "password": password} |
| ) |
| break |
| except requests.exceptions.ConnectionError: |
| |
| pass |
|
|
| yield ecr_image |
|
|
| |
| _delete_repository(ecr_client, algorithm_name) |
|
|
|
|
| def _ecr_image_uri(sagemaker_session, algorithm_name): |
| region = sagemaker_session.boto_region_name |
|
|
| sts_client = sagemaker_session.boto_session.client( |
| "sts", region_name=region, endpoint_url=utils.sts_regional_endpoint(region) |
| ) |
| account_id = sts_client.get_caller_identity()["Account"] |
|
|
| endpoint_data = utils._botocore_resolver().construct_endpoint("ecr", region) |
| return "{}.dkr.{}/{}:latest".format(account_id, endpoint_data["hostname"], algorithm_name) |
|
|
|
|
| def _create_repository(ecr_client, repository_name): |
| """ |
| Creates an ECS Repository (ECR). When a new transform is being registered, |
| we'll need a repository to push the image (and composed model images) to |
| """ |
| try: |
| response = ecr_client.create_repository(repositoryName=repository_name) |
| return response["repository"]["repositoryUri"] |
| except ClientError as e: |
| |
| if "RepositoryAlreadyExistsException" == e.response.get("Error", {}).get("Code"): |
| response = ecr_client.describe_repositories(repositoryNames=[repository_name]) |
| return response["repositories"][0]["repositoryUri"] |
| else: |
| raise |
|
|
|
|
| def _delete_repository(ecr_client, repository_name): |
| """ |
| Deletes an ECS Repository (ECR). After the integration test completes |
| we will remove the repository created during setup |
| """ |
| try: |
| ecr_client.describe_repositories(repositoryNames=[repository_name]) |
| ecr_client.delete_repository(repositoryName=repository_name, force=True) |
| except ecr_client.exceptions.RepositoryNotFoundException: |
| pass |
|
|
|
|
| def _ecr_login(ecr_client): |
| """Get a login credentials for an ecr client.""" |
| login = ecr_client.get_authorization_token() |
| b64token = login["authorizationData"][0]["authorizationToken"].encode("utf-8") |
| username, password = base64.b64decode(b64token).decode("utf-8").split(":") |
| return username, password |
|
|
|
|
| def test_multi_data_model_deploy_pretrained_models( |
| container_image, sagemaker_session, cpu_instance_type |
| ): |
| timestamp = sagemaker_timestamp() |
| endpoint_name = "test-multimodel-endpoint-{}".format(timestamp) |
| model_name = "test-multimodel-{}".format(timestamp) |
|
|
| |
| pretrained_model_data_local_path = os.path.join(DATA_DIR, "sparkml_model", "mleap_model.tar.gz") |
|
|
| with timeout_and_delete_endpoint_by_name(endpoint_name, sagemaker_session): |
| model_data_prefix = os.path.join( |
| "s3://", sagemaker_session.default_bucket(), "multimodel-{}/".format(timestamp) |
| ) |
| multi_data_model = MultiDataModel( |
| name=model_name, |
| model_data_prefix=model_data_prefix, |
| image_uri=container_image, |
| role=ROLE, |
| sagemaker_session=sagemaker_session, |
| ) |
|
|
| |
| multi_data_model.add_model(pretrained_model_data_local_path, PRETRAINED_MODEL_PATH_1) |
| |
| multi_data_model.deploy(1, cpu_instance_type, endpoint_name=endpoint_name) |
| |
| multi_data_model.add_model(pretrained_model_data_local_path, PRETRAINED_MODEL_PATH_2) |
|
|
| endpoint_models = [] |
| for model_path in multi_data_model.list_models(): |
| endpoint_models.append(model_path) |
| assert PRETRAINED_MODEL_PATH_1 in endpoint_models |
| assert PRETRAINED_MODEL_PATH_2 in endpoint_models |
|
|
| predictor = Predictor( |
| endpoint_name=endpoint_name, |
| sagemaker_session=sagemaker_session, |
| serializer=NumpySerializer(), |
| deserializer=string_deserializer, |
| ) |
|
|
| data = numpy.zeros(shape=(1, 1, 28, 28)) |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_1) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_1) |
|
|
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_2) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_2) |
|
|
| |
| sagemaker_session.sagemaker_client.delete_endpoint_config(EndpointConfigName=endpoint_name) |
| multi_data_model.delete_model() |
| with pytest.raises(Exception) as exception: |
| sagemaker_session.sagemaker_client.describe_model(ModelName=multi_data_model.name) |
| assert "Could not find model" in str(exception.value) |
| sagemaker_session.sagemaker_client.describe_endpoint_config(name=endpoint_name) |
| assert "Could not find endpoint" in str(exception.value) |
|
|
|
|
| @pytest.mark.local_mode |
| def test_multi_data_model_deploy_pretrained_models_local_mode(container_image, sagemaker_session): |
| timestamp = sagemaker_timestamp() |
| endpoint_name = "test-multimodel-endpoint-{}".format(timestamp) |
| model_name = "test-multimodel-{}".format(timestamp) |
|
|
| |
| pretrained_model_data_local_path = os.path.join(DATA_DIR, "sparkml_model", "mleap_model.tar.gz") |
|
|
| with timeout(minutes=30): |
| model_data_prefix = os.path.join( |
| "s3://", sagemaker_session.default_bucket(), "multimodel-{}/".format(timestamp) |
| ) |
| multi_data_model = MultiDataModel( |
| name=model_name, |
| model_data_prefix=model_data_prefix, |
| image_uri=container_image, |
| role=ROLE, |
| sagemaker_session=sagemaker_session, |
| ) |
|
|
| |
| multi_data_model.add_model(pretrained_model_data_local_path, PRETRAINED_MODEL_PATH_1) |
| |
| multi_data_model.deploy(1, "local", endpoint_name=endpoint_name) |
| |
| multi_data_model.add_model(pretrained_model_data_local_path, PRETRAINED_MODEL_PATH_2) |
|
|
| endpoint_models = [] |
| for model_path in multi_data_model.list_models(): |
| endpoint_models.append(model_path) |
| assert PRETRAINED_MODEL_PATH_1 in endpoint_models |
| assert PRETRAINED_MODEL_PATH_2 in endpoint_models |
|
|
| predictor = Predictor( |
| endpoint_name=endpoint_name, |
| sagemaker_session=multi_data_model.sagemaker_session, |
| serializer=NumpySerializer(), |
| deserializer=string_deserializer, |
| ) |
|
|
| data = numpy.zeros(shape=(1, 1, 28, 28)) |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_1) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_1) |
|
|
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_2) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_2) |
|
|
| |
| multi_data_model.sagemaker_session.sagemaker_client.delete_endpoint_config( |
| EndpointConfigName=endpoint_name |
| ) |
| multi_data_model.sagemaker_session.delete_endpoint(endpoint_name) |
| multi_data_model.delete_model() |
| with pytest.raises(Exception) as exception: |
| sagemaker_session.sagemaker_client.describe_model(ModelName=multi_data_model.name) |
| assert "Could not find model" in str(exception.value) |
| sagemaker_session.sagemaker_client.describe_endpoint_config(name=endpoint_name) |
| assert "Could not find endpoint" in str(exception.value) |
|
|
|
|
| def test_multi_data_model_deploy_trained_model_from_framework_estimator( |
| container_image, |
| sagemaker_session, |
| cpu_instance_type, |
| mxnet_inference_latest_version, |
| mxnet_inference_latest_py_version, |
| ): |
| timestamp = sagemaker_timestamp() |
| endpoint_name = "test-multimodel-endpoint-{}".format(timestamp) |
| model_name = "test-multimodel-{}".format(timestamp) |
|
|
| with timeout_and_delete_endpoint_by_name(endpoint_name, sagemaker_session): |
| mxnet_model_1 = _mxnet_training_job( |
| sagemaker_session, |
| container_image, |
| mxnet_inference_latest_version, |
| mxnet_inference_latest_py_version, |
| cpu_instance_type, |
| 0.1, |
| ) |
| model_data_prefix = os.path.join( |
| "s3://", sagemaker_session.default_bucket(), "multimodel-{}/".format(timestamp) |
| ) |
| multi_data_model = MultiDataModel( |
| name=model_name, |
| model_data_prefix=model_data_prefix, |
| model=mxnet_model_1, |
| sagemaker_session=sagemaker_session, |
| ) |
|
|
| |
| multi_data_model.add_model(mxnet_model_1.model_data, PRETRAINED_MODEL_PATH_1) |
| |
| multi_data_model.deploy(1, cpu_instance_type, endpoint_name=endpoint_name) |
|
|
| |
| mxnet_model_2 = _mxnet_training_job( |
| sagemaker_session, |
| container_image, |
| mxnet_inference_latest_version, |
| mxnet_inference_latest_py_version, |
| cpu_instance_type, |
| 0.01, |
| ) |
| |
| multi_data_model.add_model(mxnet_model_2.model_data, PRETRAINED_MODEL_PATH_2) |
|
|
| endpoint_models = [] |
| for model_path in multi_data_model.list_models(): |
| endpoint_models.append(model_path) |
| assert PRETRAINED_MODEL_PATH_1 in endpoint_models |
| assert PRETRAINED_MODEL_PATH_2 in endpoint_models |
|
|
| |
| |
| |
| predictor = Predictor( |
| endpoint_name=endpoint_name, |
| sagemaker_session=sagemaker_session, |
| serializer=NumpySerializer(), |
| deserializer=string_deserializer, |
| ) |
|
|
| data = numpy.zeros(shape=(1, 1, 28, 28)) |
| |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_1) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_1) |
|
|
| |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_2) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_2) |
|
|
| |
| sagemaker_session.sagemaker_client.delete_endpoint_config(EndpointConfigName=endpoint_name) |
| multi_data_model.delete_model() |
| with pytest.raises(Exception) as exception: |
| sagemaker_session.sagemaker_client.describe_model(ModelName=model_name) |
| assert "Could not find model" in str(exception.value) |
| sagemaker_session.sagemaker_client.describe_endpoint_config(name=endpoint_name) |
| assert "Could not find endpoint" in str(exception.value) |
|
|
|
|
| def _mxnet_training_job( |
| sagemaker_session, container_image, mxnet_version, py_version, cpu_instance_type, learning_rate |
| ): |
| with timeout(minutes=TRAINING_DEFAULT_TIMEOUT_MINUTES): |
| script_path = os.path.join(DATA_DIR, "mxnet_mnist", "mnist.py") |
| data_path = os.path.join(DATA_DIR, "mxnet_mnist") |
|
|
| mx = MXNet( |
| entry_point=script_path, |
| role=ROLE, |
| framework_version=mxnet_version, |
| py_version=py_version, |
| instance_count=1, |
| instance_type=cpu_instance_type, |
| sagemaker_session=sagemaker_session, |
| hyperparameters={"learning-rate": learning_rate}, |
| ) |
|
|
| train_input = mx.sagemaker_session.upload_data( |
| path=os.path.join(data_path, "train"), key_prefix="integ-test-data/mxnet_mnist/train" |
| ) |
| test_input = mx.sagemaker_session.upload_data( |
| path=os.path.join(data_path, "test"), key_prefix="integ-test-data/mxnet_mnist/test" |
| ) |
|
|
| mx.fit({"train": train_input, "test": test_input}) |
|
|
| |
| |
| return mx.create_model(image_uri=container_image) |
|
|
|
|
| @pytest.mark.slow_test |
| def test_multi_data_model_deploy_train_model_from_amazon_first_party_estimator( |
| container_image, sagemaker_session, cpu_instance_type |
| ): |
| timestamp = sagemaker_timestamp() |
| endpoint_name = "test-multimodel-endpoint-{}".format(timestamp) |
| model_name = "test-multimodel-{}".format(timestamp) |
|
|
| with timeout_and_delete_endpoint_by_name(endpoint_name, sagemaker_session): |
| rcf_model_v1 = __rcf_training_job( |
| sagemaker_session, container_image, cpu_instance_type, 50, 20 |
| ) |
|
|
| model_data_prefix = os.path.join( |
| "s3://", sagemaker_session.default_bucket(), "multimodel-{}/".format(timestamp) |
| ) |
| multi_data_model = MultiDataModel( |
| name=model_name, |
| model_data_prefix=model_data_prefix, |
| model=rcf_model_v1, |
| sagemaker_session=sagemaker_session, |
| ) |
|
|
| |
| multi_data_model.add_model(rcf_model_v1.model_data, PRETRAINED_MODEL_PATH_1) |
| |
| multi_data_model.deploy(1, cpu_instance_type, endpoint_name=endpoint_name) |
| |
| rcf_model_v2 = __rcf_training_job( |
| sagemaker_session, container_image, cpu_instance_type, 70, 20 |
| ) |
| |
| multi_data_model.add_model(rcf_model_v2.model_data, PRETRAINED_MODEL_PATH_2) |
|
|
| |
| endpoint_models = [] |
| for model_path in multi_data_model.list_models(): |
| endpoint_models.append(model_path) |
| assert PRETRAINED_MODEL_PATH_1 in endpoint_models |
| assert PRETRAINED_MODEL_PATH_2 in endpoint_models |
|
|
| |
| |
| |
| predictor = Predictor( |
| endpoint_name=endpoint_name, |
| sagemaker_session=sagemaker_session, |
| serializer=NumpySerializer(), |
| deserializer=string_deserializer, |
| ) |
|
|
| data = numpy.random.rand(1, 14) |
| |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_1) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_1) |
|
|
| |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_2) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_2) |
|
|
| |
| sagemaker_session.sagemaker_client.delete_endpoint_config(EndpointConfigName=endpoint_name) |
| multi_data_model.delete_model() |
| with pytest.raises(Exception) as exception: |
| sagemaker_session.sagemaker_client.describe_model(ModelName=model_name) |
| assert "Could not find model" in str(exception.value) |
| sagemaker_session.sagemaker_client.describe_endpoint_config(name=endpoint_name) |
| assert "Could not find endpoint" in str(exception.value) |
|
|
|
|
| def __rcf_training_job( |
| sagemaker_session, container_image, cpu_instance_type, num_trees, num_samples_per_tree |
| ): |
| job_name = unique_name_from_base("randomcutforest") |
| with timeout(minutes=TRAINING_DEFAULT_TIMEOUT_MINUTES): |
| |
| feature_num = 14 |
| train_input = numpy.random.rand(1000, feature_num) |
|
|
| rcf = RandomCutForest( |
| role=ROLE, |
| instance_count=1, |
| instance_type=cpu_instance_type, |
| num_trees=num_trees, |
| num_samples_per_tree=num_samples_per_tree, |
| eval_metrics=["accuracy", "precision_recall_fscore"], |
| sagemaker_session=sagemaker_session, |
| ) |
|
|
| rcf.fit(records=rcf.record_set(train_input), job_name=job_name) |
|
|
| |
| |
| rcf_model = rcf.create_model() |
| rcf_model.image_uri = container_image |
| return rcf_model |
|
|
|
|
| def test_multi_data_model_deploy_pretrained_models_update_endpoint( |
| container_image, sagemaker_session, cpu_instance_type, alternative_cpu_instance_type |
| ): |
| timestamp = sagemaker_timestamp() |
| endpoint_name = "test-multimodel-endpoint-{}".format(timestamp) |
| model_name = "test-multimodel-{}".format(timestamp) |
|
|
| |
| pretrained_model_data_local_path = os.path.join(DATA_DIR, "sparkml_model", "mleap_model.tar.gz") |
|
|
| with timeout_and_delete_endpoint_by_name(endpoint_name, sagemaker_session): |
| model_data_prefix = os.path.join( |
| "s3://", sagemaker_session.default_bucket(), "multimodel-{}/".format(timestamp) |
| ) |
| multi_data_model = MultiDataModel( |
| name=model_name, |
| model_data_prefix=model_data_prefix, |
| image_uri=container_image, |
| role=ROLE, |
| sagemaker_session=sagemaker_session, |
| ) |
|
|
| |
| multi_data_model.add_model(pretrained_model_data_local_path, PRETRAINED_MODEL_PATH_1) |
| |
| multi_data_model.deploy(1, cpu_instance_type, endpoint_name=endpoint_name) |
| |
| multi_data_model.add_model(pretrained_model_data_local_path, PRETRAINED_MODEL_PATH_2) |
|
|
| |
| endpoint_models = [] |
| for model_path in multi_data_model.list_models(): |
| endpoint_models.append(model_path) |
| assert PRETRAINED_MODEL_PATH_1 in endpoint_models |
| assert PRETRAINED_MODEL_PATH_2 in endpoint_models |
|
|
| predictor = Predictor( |
| endpoint_name=endpoint_name, |
| sagemaker_session=sagemaker_session, |
| serializer=NumpySerializer(), |
| deserializer=string_deserializer, |
| ) |
|
|
| data = numpy.zeros(shape=(1, 1, 28, 28)) |
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_1) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_1) |
|
|
| result = predictor.predict(data, target_model=PRETRAINED_MODEL_PATH_2) |
| assert result == "Invoked model: {}".format(PRETRAINED_MODEL_PATH_2) |
|
|
| endpoint_desc = sagemaker_session.sagemaker_client.describe_endpoint( |
| EndpointName=endpoint_name |
| ) |
| old_config_name = endpoint_desc["EndpointConfigName"] |
|
|
| |
| predictor.update_endpoint( |
| initial_instance_count=1, instance_type=alternative_cpu_instance_type |
| ) |
|
|
| endpoint_desc = sagemaker_session.sagemaker_client.describe_endpoint( |
| EndpointName=endpoint_name |
| ) |
| new_config_name = endpoint_desc["EndpointConfigName"] |
|
|
| new_config = sagemaker_session.sagemaker_client.describe_endpoint_config( |
| EndpointConfigName=new_config_name |
| ) |
| assert old_config_name != new_config_name |
| assert new_config["ProductionVariants"][0]["InstanceType"] == alternative_cpu_instance_type |
| assert new_config["ProductionVariants"][0]["InitialInstanceCount"] == 1 |
|
|
| |
| sagemaker_session.sagemaker_client.delete_endpoint_config( |
| EndpointConfigName=old_config_name |
| ) |
| sagemaker_session.sagemaker_client.delete_endpoint_config( |
| EndpointConfigName=new_config_name |
| ) |
| multi_data_model.delete_model() |
|
|
| with pytest.raises(Exception) as exception: |
| sagemaker_session.sagemaker_client.describe_model(ModelName=model_name) |
| assert "Could not find model" in str(exception.value) |
| sagemaker_session.sagemaker_client.describe_endpoint_config(name=old_config_name) |
| assert "Could not find endpoint" in str(exception.value) |
| sagemaker_session.sagemaker_client.describe_endpoint_config(name=new_config_name) |
| assert "Could not find endpoint" in str(exception.value) |
|
|