@@ -1377,6 +1377,7 @@ def deploy(
1377
1377
managed_instance_scaling : Optional [str ] = None ,
1378
1378
inference_component_name = None ,
1379
1379
routing_config : Optional [Dict [str , Any ]] = None ,
1380
+ model_reference_arn : Optional [str ] = None ,
1380
1381
** kwargs ,
1381
1382
):
1382
1383
"""Deploy this ``Model`` to an ``Endpoint`` and optionally return a ``Predictor``.
@@ -1483,6 +1484,8 @@ def deploy(
1483
1484
{
1484
1485
"RoutingStrategy": sagemaker.enums.RoutingStrategy.RANDOM
1485
1486
}
1487
+ model_reference_arn (Optional [str]): Hub Content Arn of a Model Reference type
1488
+ content (default: None).
1486
1489
Raises:
1487
1490
ValueError: If arguments combination check failed in these circumstances:
1488
1491
- If no role is specified or
@@ -1696,7 +1699,9 @@ def deploy(
1696
1699
instance_type = instance_type ,
1697
1700
accelerator_type = accelerator_type ,
1698
1701
tags = tags ,
1699
- serverless_inference_config = serverless_inference_config
1702
+ serverless_inference_config = serverless_inference_config ,
1703
+ accept_eula = accept_eula ,
1704
+ model_reference_arn = model_reference_arn
1700
1705
)
1701
1706
serverless_inference_config_dict = (
1702
1707
serverless_inference_config ._to_request_dict () if is_serverless else None
0 commit comments