Remove Accountability Remove Big data Remove Scripts
article thumbnail

Amazon SageMaker Feature Store now supports cross-account sharing, discovery, and access

AWS Machine Learning

SageMaker Feature Store now makes it effortless to share, discover, and access feature groups across AWS accounts. With this launch, account owners can grant access to select feature groups by other accounts using AWS Resource Access Manager (AWS RAM).

article thumbnail

Secure Amazon SageMaker Studio presigned URLs Part 3: Multi-account private API access to Studio

AWS Machine Learning

One important aspect of this foundation is to organize their AWS environment following a multi-account strategy. In this post, we show how you can extend that architecture to multiple accounts to support multiple LOBs. In this post, we show how you can extend that architecture to multiple accounts to support multiple LOBs.

APIs 81
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Create a generative AI–powered custom Google Chat application using Amazon Bedrock

AWS Machine Learning

An AWS account and an AWS Identity and Access Management (IAM) principal with sufficient permissions to create and manage the resources needed for this application. If you don’t have an AWS account, refer to How do I create and activate a new Amazon Web Services account? The script deploys the AWS CDK project in your account.

APIs 123
article thumbnail

Promote pipelines in a multi-environment setup using Amazon SageMaker Model Registry, HashiCorp Terraform, GitHub, and Jenkins CI/CD

AWS Machine Learning

Central model registry – Amazon SageMaker Model Registry is set up in a separate AWS account to track model versions generated across the dev and prod environments. Approve the model in SageMaker Model Registry in the central model registry account. Create a pull request to merge the code into the main branch of the GitHub repository.

Scripts 123
article thumbnail

How Twilio used Amazon SageMaker MLOps pipelines with PrestoDB to enable frequent model retraining and optimized batch transform

AWS Machine Learning

Batch transform The batch transform pipeline consists of the following steps: The pipeline implements a data preparation step that retrieves data from a PrestoDB instance (using a data preprocessing script ) and stores the batch data in Amazon Simple Storage Service (Amazon S3).

Scripts 121
article thumbnail

Use Snowflake as a data source to train ML models with Amazon SageMaker

AWS Machine Learning

We create a custom training container that downloads data directly from the Snowflake table into the training instance rather than first downloading the data into an S3 bucket. Store your Snowflake account credentials in AWS Secrets Manager. Ingest the data in a table in your Snowflake account.

Scripts 128
article thumbnail

Centralize model governance with SageMaker Model Registry Resource Access Manager sharing

AWS Machine Learning

We recently announced the general availability of cross-account sharing of Amazon SageMaker Model Registry using AWS Resource Access Manager (AWS RAM) , making it easier to securely share and discover machine learning (ML) models across your AWS accounts. Mitigation strategies : Implementing measures to minimize or eliminate risks.