Skip to content

Commit 0042e11

Browse files
authored
Update description for TransformInput()
1 parent e917cb4 commit 0042e11

File tree

1 file changed

+48
-0
lines changed

1 file changed

+48
-0
lines changed

src/sagemaker/inputs.py

Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -165,6 +165,54 @@ class TransformInput(object):
165165
"""Create a class containing all the parameters.
166166
167167
It can be used when calling ``sagemaker.transformer.Transformer.transform()``
168+
169+
Args:
170+
data (str): The S3 location of the input data that the model can consume.
171+
data_type (str): elastic inference accelerator type.
172+
(default: ``'S3Prefix'``)
173+
content_type (str): The multipurpose internet mail extension (MIME) type of the data.
174+
(default: None)
175+
compression_type (str): If your transform data is compressed, specify the compression type.
176+
Valid values: ``'Gzip'``, ``None``
177+
(default: None)
178+
split_type (str): The method to use to split the transform job's data files into smaller
179+
batches.
180+
Valid values: ``'Line'``, ``RecordIO``, ``'TFRecord'``, None
181+
(default: None)
182+
input_filter (str): A JSONPath expression used to select a portion of the input data to pass
183+
to the algorithm. Use the InputFilter parameter to exclude fields, such as an ID column,
184+
from the input. If you want SageMaker to pass the entire input dataset to the algorithm,
185+
accept the default value ``$``. For more information on batch transform data processing,
186+
input, join, and output, see
187+
`Associate Prediction Results with Input Records
188+
<https://docs.aws.amazon.com/sagemaker/latest/dg/batch-transform-data-processing.html?`.
189+
Valid values: ``$``. See `JSONPath Operators
190+
<https://docs.aws.amazon.com/sagemaker/latest/dg/batch-transform-data-processing.html#data-processing-operators>'
191+
for more options.
192+
(default: ``$``)
193+
output_filter (str): A JSONPath expression used to select a portion of the joined dataset to
194+
save in the output file for a batch transform job. If you want SageMaker to store the
195+
entire input dataset in the output file, leave the default value, $. If you specify
196+
indexes that aren't within the dimension size of the joined dataset, you get an error.
197+
Valid values: ``$``. See `JSONPath Operators
198+
<https://docs.aws.amazon.com/sagemaker/latest/dg/batch-transform-data-processing.html#data-processing-operators>'
199+
for more options.
200+
(default: ``$``)
201+
join_source (str): Specifies the source of the data to join with the transformed data. The
202+
default value is ``None``, which specifies not to join the input with the transformed data.
203+
If you want the batch transform job to join the original input data with the transformed data,
204+
set to ``Input``. For more information on
205+
Valid values: ``None``, ``Input``
206+
(default: None)
207+
model_client_config (dict): Configures the timeout and maximum number of retries for
208+
processing a transform job invocation.
209+
210+
* ``'InvocationsTimeoutInSeconds'`` (int) - The timeout value in seconds for an invocation request. The default value is 600.
211+
* ``'InvocationsMaxRetries'`` (int) - The maximum number of retries when invocation requests are failing.
212+
(default: ``{'InvocationsTimeoutInSeconds'=600,'InvocationsMaxRetries'=3}``)
213+
batch_data_capture_config (dict): Specifies configuration related to batch transform job
214+
data capture for use with Amazon SageMaker Model Monitoring.
215+
(default: None)
168216
"""
169217

170218
data: str = attr.ib()

0 commit comments

Comments
 (0)