Remove APIs Remove Best practices Remove Big data Remove Engineering
article thumbnail

Use AWS PrivateLink to set up private access to Amazon Bedrock

AWS Machine Learning

It allows developers to build and scale generative AI applications using FMs through an API, without managing infrastructure. You can choose from various FMs from Amazon and leading AI startups such as AI21 Labs, Anthropic, Cohere, and Stability AI to find the model that’s best suited for your use case.

APIs 122
article thumbnail

Architect defense-in-depth security for generative AI applications using the OWASP Top 10 for LLMs

AWS Machine Learning

The goal of this post is to empower AI and machine learning (ML) engineers, data scientists, solutions architects, security teams, and other stakeholders to have a common mental model and framework to apply security best practices, allowing AI/ML teams to move fast without trading off security for speed.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

How Amp on Amazon used data to increase customer engagement, Part 2: Building a personalized show recommendation platform using Amazon SageMaker

AWS Machine Learning

Our initial ML model uses 21 batch features computed daily using data captured in the past 2 months. This data includes both playback and app engagement history per user, and grows with the number of users and frequency of app usage. The processing infrastructure is needed only for the duration of the computations. Real-time inference.

article thumbnail

Identify objections in customer conversations using Amazon Comprehend to enhance customer experience without ML expertise

AWS Machine Learning

In this post, we explore how AWS customer Pro360 used the Amazon Comprehend custom classification API , which enables you to easily build custom text classification models using your business-specific labels without requiring you to learn machine learning (ML), to improve customer experience and reduce operational costs.

APIs 61
article thumbnail

How Patsnap used GPT-2 inference on Amazon SageMaker with low latency and cost

AWS Machine Learning

This blog post was co-authored, and includes an introduction, by Zilong Bai, senior natural language processing engineer at Patsnap. They use big data (such as a history of past search queries) to provide many powerful yet easy-to-use patent tools. implement the model and the inference API. gpt2 and predictor.py

APIs 66
article thumbnail

How Amp on Amazon used data to increase customer engagement, Part 1: Building a data analytics platform

AWS Machine Learning

Amp wanted a scalable data and analytics platform to enable easy access to data and perform machine leaning (ML) experiments for live audio transcription, content moderation, feature engineering, and a personal show recommendation service, and to inspect or measure business KPIs and metrics. Data Engineer for Amp on Amazon.

article thumbnail

Enable fully homomorphic encryption with Amazon SageMaker endpoints for secure, real-time inferencing

AWS Machine Learning

Applications and services can call the deployed endpoint directly or through a deployed serverless Amazon API Gateway architecture. To learn more about real-time endpoint architectural best practices, refer to Creating a machine learning-powered REST API with Amazon API Gateway mapping templates and Amazon SageMaker.

Scripts 93