The general process for setting up a training job in Amazon Sagemaker is as follows:
Create or select a S3 bucket to store the input data and output artifacts of the training job.
Prepare the training data and upload it to the S3 bucket.
Create a training script in Python that defines the machine learning model, the hyperparameters, and the input and output channels. This script should account for the SageMaker-specific environment variables that are automatically passed to the training container.
Create an Amazon SageMaker Training Job configuration specifying, for example, the instance type to use or the algorithm to implement.
Launch the training job. This will start an Amazon SageMaker TensorFlow container, for example, based on the specified configuration and using the configured input and output data channels.
Monitor the job progress and the logs and metrics generated by the training container.
Once the job is completed, retrieve the training artifacts from the output S3 bucket or by downloading the output model from the SageMaker console.
If appropriate, deploy the trained model to an Amazon SageMaker Endpoint for inference.
Please start posting anonymously - your entry will be published after you log in or create a new account. This space is reserved only for answers. If you would like to engage in a discussion, please instead post a comment under the question or an answer that you would like to discuss
Asked: 2021-12-31 11:00:00 +0000
Seen: 1 times
Last updated: Mar 07
How can popen() be used to direct streaming data to TAR?
In Python, can a string be utilized to retrieve a dataframe that has the same name as the string?
What is the method for merging field value and text into a singular line for display?
What is the method for programmatic access to a time series?