AnnouncementPinecone Connect is now generally available to access Pinecone within your platformGet started

Pinecone and Amazon Web Services (AWS)

Bring scalable Gen AI applications to market with cost-efficiency

KlarnaHubSpotClickUpShortwaveBAIN & COMPANYGoDaddyGongAccentureCVS HealthCourse HeroMidjourney

Building with Pinecone and AWS

Pinecone and Amazon Web Services (AWS) empower you to build highly performant, scalable, and reliable production-ready Gen AI applications with ease.

Amazon Bedrock for Pinecone

Amazon Bedrock provides access to pre-trained foundation models through an API, allowing users to experiment with different foundation models with the ability to easily finetune and augment them.

With this integration, AWS customers can quickly and effortlessly build search and GenAI applications with a workflow called Retrieval Augmented Generation (RAG). RAG powered by Pinecone ensures the most relevant, accurate, and fast responses to end users.

Learn more

Amazon SageMaker for Pinecone

SageMaker is a machine learning service that allows data scientists and developers to easily build, train Large Language Models (LLMs), and then directly deploy them into a production-ready hosted environment.

The integration allows customers to use SageMaker compute and model hosting for LLMs and Pinecone as the knowledge base that keeps our LLMs up to date with the latest information to reduce hallucinations.

Learn more

Pinecone on AWS Marketplace

Develop highly scalable Gen AI apps with Pinecone on the AWS marketplace. Pinecone offers a usage-based pricing model with no minimums or upfront commitments. Easy billing management through your AWS account.

You will be billed per minute from the moment your index is live, with monthly invoices.

Why Build with Pinecone and AWS

High Performance

  • Low latency and high throughput. Speed through data in milliseconds
  • Quick and accurate results with real-time data freshness and metadata filtering
  • Support both vector search and hybrid search
  • Scale beyond billions of vectors cost-effectively without compromising performance

Developer Favorite

  • Get started in seconds with our console that requires no AI expertise
  • Fully managed service without the need to maintain infrastructure, or monitor services
  • Intuitive APIs and SDKs
  • Compatible with any LLMs

Enterprise Ready

  • Enterprise-grade security and compliance: GDPR-ready, SOC2 Type II certified, and HIPPA-compliant
  • Data encryption in transit and at rest
  • Stringent access controls
  • Uptime and response time SLAs
  • title

    Deploying Open Source LLMs for RAG with SageMaker

    In this article, we'll learn how to build LLM + RAG pipelines using open-source models from Hugging Face deployed on AWS SageMaker.

    Read Article
  • title

    Build enterprise-grade Q&A at scale with Open LLMs on AWS

    In this video, we’ll explore how developers can build a reliable and scalable question-answering system on Amazon Web Services (AWS) using open LLMs.

    Watch Video