Remove APIs Remove Best practices Remove Big data Remove Scripts
article thumbnail

Enable fully homomorphic encryption with Amazon SageMaker endpoints for secure, real-time inferencing

AWS Machine Learning

Applications and services can call the deployed endpoint directly or through a deployed serverless Amazon API Gateway architecture. To learn more about real-time endpoint architectural best practices, refer to Creating a machine learning-powered REST API with Amazon API Gateway mapping templates and Amazon SageMaker.

Scripts 96
article thumbnail

Architect defense-in-depth security for generative AI applications using the OWASP Top 10 for LLMs

AWS Machine Learning

The goal of this post is to empower AI and machine learning (ML) engineers, data scientists, solutions architects, security teams, and other stakeholders to have a common mental model and framework to apply security best practices, allowing AI/ML teams to move fast without trading off security for speed.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

How Amp on Amazon used data to increase customer engagement, Part 1: Building a data analytics platform

AWS Machine Learning

When the message is received by the SQS queue, it triggers the AWS Lambda function to make an API call to the Amp catalog service. The Lambda function retrieves the desired show metadata, filters the metadata, and then sends the output metadata to Amazon Kinesis Data Streams. Data Engineer for Amp on Amazon.

article thumbnail

Securing MLflow in AWS: Fine-grained access control with AWS native services

AWS Machine Learning

In this post, we address these limitations by implementing the access control outside of the MLflow server and offloading authentication and authorization tasks to Amazon API Gateway , where we implement fine-grained access control mechanisms at the resource level using Identity and Access Management (IAM). Adds an IAM authorizer.

APIs 71
article thumbnail

MLOps for batch inference with model monitoring and retraining using Amazon SageMaker, HashiCorp Terraform, and GitLab CI/CD

AWS Machine Learning

Refer to Operating model for best practices regarding a multi-account strategy for ML. When a new version of the model is registered in the model registry, it triggers a notification to the responsible data scientist via Amazon SNS. The central model registry could optionally be placed in a shared services account as well.

Scripts 74
article thumbnail

Build repeatable, secure, and extensible end-to-end machine learning workflows using Kubeflow on AWS

AWS Machine Learning

Prior to our adoption of Kubeflow on AWS, our data scientists used a standardized set of tools and a process that allowed flexibility in the technology and workflow used to train a given model. Each project maintained detailed documentation that outlined how each script was used to build the final model.