Workflows
Blocks
Database
Pinecone

Pinecone Block

What it does: Connect to Pinecone's managed vector database for enterprise-grade semantic search and AI knowledge retrieval.

🌲

In simple terms: Pinecone is like a super-powered search engine for your data that understands meaning. When your document collection grows large, Pinecone handles the heavy lifting - managing millions of records while keeping searches lightning fast.

When to Use This

Use Pinecone when you need:

  • ✅ Production-ready vector search at scale
  • ✅ Handle millions of documents efficiently
  • ✅ Enterprise reliability and uptime
  • ✅ Managed infrastructure (no servers to maintain)
  • ✅ Fast, consistent search performance

Example: Power an AI assistant that searches through your company's entire documentation library to answer employee questions instantly.

Key Features

  • Managed Service: No infrastructure to maintain
  • Massive Scale: Handle millions to billions of vectors
  • Fast Queries: Sub-second search responses
  • High Availability: Built-in redundancy and uptime
  • Metadata Filtering: Combine semantic search with traditional filters

Pinecone Concepts Explained

TermWhat It Means
IndexYour searchable database (like a folder for related content)
NamespaceA way to organize content within an index (like subfolders)
VectorThe numerical representation of your text
MetadataAdditional info attached to each record (category, date, source)
EmbeddingThe process of converting text to vectors

Setup Guide

Step 1: Create a Pinecone Account

  1. Go to pinecone.io and sign up
  2. Create a new project
  3. Choose your plan (free tier available)
  4. Get your API key from the dashboard

Step 2: Create an Index

  1. In Pinecone dashboard, click "Create Index"
  2. Name your index (e.g., "knowledge-base")
  3. Choose dimensions (must match your embedding model - typically 1536 for OpenAI)
  4. Select your cloud region
  5. Click Create

Step 3: Configure the Block

Connection Settings:

  • API Key: Your Pinecone API key
  • Index Name: The index you created
  • Namespace: Optional organization (leave empty for default)

Step 4: Choose Your Operation

OperationWhat It Does
UpsertAdd or update vectors in your index
QuerySearch for similar content
DeleteRemove vectors from your index
FetchRetrieve specific vectors by ID

Common Use Cases

Enterprise Knowledge Base

Store your company's documentation, policies, and procedures. Let employees search naturally - "What's our vacation policy?" - and get relevant results instantly.

Customer Support AI

Index all your help articles, FAQs, and product documentation. Your AI assistant can find the perfect answer for any customer question.

Document Similarity

Find related documents, contracts, or content. Upload a document and find others with similar topics or themes.

Product Search

Create semantic product search that understands what customers mean, not just what they type. "Comfortable shoes for walking" finds the right products.

Research and Discovery

Index research papers, reports, or articles. Discover related work based on concepts, not just keywords.

Operations Explained

Upsert (Add/Update)

Add new content or update existing content in your index.

What you provide:

  • Unique ID for each piece of content
  • The text content (will be converted to vector)
  • Optional metadata (category, source, date)

Query (Search)

Find relevant content based on a natural language question or text.

What you provide:

  • Your search query
  • Number of results to return
  • Optional metadata filters

Delete

Remove content from your index.

What you provide:

  • The IDs of vectors to delete
  • Or filter criteria to delete matching records

Fetch

Retrieve specific vectors by their ID.

What you provide:

  • The IDs of vectors you want to retrieve

What You Get Back

After querying Pinecone:

  • Matches: List of relevant content found
  • Scores: Relevance score for each match (0 to 1)
  • Metadata: Any additional info stored with each vector
  • IDs: Unique identifiers for each result

Tips for Success

  1. Choose the right dimensions - Must match your embedding model (OpenAI = 1536, Cohere = 1024)
  2. Use namespaces wisely - Organize content by tenant, category, or type
  3. Add useful metadata - Store source, date, category for later filtering
  4. Plan your index structure - Think about how you'll query before you build
  5. Monitor your usage - Keep track of vector count and query volume

Troubleshooting

ProblemLikely CauseSolution
Dimension mismatchEmbedding model doesn't match indexCreate index with correct dimensions
No resultsEmpty index or very different queryVerify content is indexed, adjust query
Slow queriesVery large result set or indexAdd filters, reduce top_k
Index not foundWrong index name or regionVerify index name and API key
Rate limitsToo many requestsAdd delays or upgrade plan

Best Practices

  • Chunk documents properly - Split large documents into meaningful sections
  • Keep metadata consistent - Use the same fields across all vectors
  • Use namespaces for multi-tenancy - Separate data by customer or category
  • Clean up unused vectors - Delete outdated content regularly
  • Test with real queries - Validate search quality with actual user questions
Indite Documentation v1.4.0
PrivacyTermsSupport