Multiple Choice
A Machine Learning Specialist is designing a scalable data storage solution for Amazon SageMaker. There is an existing TensorFlow-based model implemented as a train.py script that relies on static training data that is currently stored as TFRecords. Which method of providing training data to Amazon SageMaker would meet the business requirements with the LEAST development overhead?
A) Use Amazon SageMaker script mode and use train.py unchanged. Point the Amazon SageMaker training invocation to the local path of the data without reformatting the training data.
B) Use Amazon SageMaker script mode and use train.py unchanged. Put the TFRecord data into an Amazon S3 bucket. Point the Amazon SageMaker training invocation to the S3 bucket without reformatting the training data.
C) Rewrite the train.py script to add a section that converts TFRecords to protobuf and ingests the protobuf data instead of TFRecords.
D) Prepare the data in the format accepted by Amazon SageMaker. Use AWS Glue or AWS Lambda to reformat and store the data in an Amazon S3 bucket.
Correct Answer:

Verified
Correct Answer:
Verified
Q73: A retail company intends to use machine
Q74: A company uses a long short-term memory
Q75: A Machine Learning Specialist is working with
Q76: A large company has developed a BI
Q77: A Data Scientist needs to migrate an
Q79: A Machine Learning Specialist is using an
Q80: A Machine Learning Specialist has created a
Q81: A monitoring service generates 1 TB of
Q82: A company is observing low accuracy while
Q83: A large consumer goods manufacturer has the