Pinecone Block
What it does: Connect to Pinecone's managed vector database for enterprise-grade semantic search and AI knowledge retrieval.
In simple terms: Pinecone is like a super-powered search engine for your data that understands meaning. When your document collection grows large, Pinecone handles the heavy lifting - managing millions of records while keeping searches lightning fast.
When to Use This
Use Pinecone when you need:
- ✅ Production-ready vector search at scale
- ✅ Handle millions of documents efficiently
- ✅ Enterprise reliability and uptime
- ✅ Managed infrastructure (no servers to maintain)
- ✅ Fast, consistent search performance
Example: Power an AI assistant that searches through your company's entire documentation library to answer employee questions instantly.
Key Features
- Managed Service: No infrastructure to maintain
- Massive Scale: Handle millions to billions of vectors
- Fast Queries: Sub-second search responses
- High Availability: Built-in redundancy and uptime
- Metadata Filtering: Combine semantic search with traditional filters
Pinecone Concepts Explained
| Term | What It Means |
|---|---|
| Index | Your searchable database (like a folder for related content) |
| Namespace | A way to organize content within an index (like subfolders) |
| Vector | The numerical representation of your text |
| Metadata | Additional info attached to each record (category, date, source) |
| Embedding | The process of converting text to vectors |
Setup Guide
Step 1: Create a Pinecone Account
- Go to pinecone.io and sign up
- Create a new project
- Choose your plan (free tier available)
- Get your API key from the dashboard
Step 2: Create an Index
- In Pinecone dashboard, click "Create Index"
- Name your index (e.g., "knowledge-base")
- Choose dimensions (must match your embedding model - typically 1536 for OpenAI)
- Select your cloud region
- Click Create
Step 3: Configure the Block
Connection Settings:
- API Key: Your Pinecone API key
- Index Name: The index you created
- Namespace: Optional organization (leave empty for default)
Step 4: Choose Your Operation
| Operation | What It Does |
|---|---|
| Upsert | Add or update vectors in your index |
| Query | Search for similar content |
| Delete | Remove vectors from your index |
| Fetch | Retrieve specific vectors by ID |
Common Use Cases
Enterprise Knowledge Base
Store your company's documentation, policies, and procedures. Let employees search naturally - "What's our vacation policy?" - and get relevant results instantly.
Customer Support AI
Index all your help articles, FAQs, and product documentation. Your AI assistant can find the perfect answer for any customer question.
Document Similarity
Find related documents, contracts, or content. Upload a document and find others with similar topics or themes.
Product Search
Create semantic product search that understands what customers mean, not just what they type. "Comfortable shoes for walking" finds the right products.
Research and Discovery
Index research papers, reports, or articles. Discover related work based on concepts, not just keywords.
Operations Explained
Upsert (Add/Update)
Add new content or update existing content in your index.
What you provide:
- Unique ID for each piece of content
- The text content (will be converted to vector)
- Optional metadata (category, source, date)
Query (Search)
Find relevant content based on a natural language question or text.
What you provide:
- Your search query
- Number of results to return
- Optional metadata filters
Delete
Remove content from your index.
What you provide:
- The IDs of vectors to delete
- Or filter criteria to delete matching records
Fetch
Retrieve specific vectors by their ID.
What you provide:
- The IDs of vectors you want to retrieve
What You Get Back
After querying Pinecone:
- Matches: List of relevant content found
- Scores: Relevance score for each match (0 to 1)
- Metadata: Any additional info stored with each vector
- IDs: Unique identifiers for each result
Tips for Success
- Choose the right dimensions - Must match your embedding model (OpenAI = 1536, Cohere = 1024)
- Use namespaces wisely - Organize content by tenant, category, or type
- Add useful metadata - Store source, date, category for later filtering
- Plan your index structure - Think about how you'll query before you build
- Monitor your usage - Keep track of vector count and query volume
Troubleshooting
| Problem | Likely Cause | Solution |
|---|---|---|
| Dimension mismatch | Embedding model doesn't match index | Create index with correct dimensions |
| No results | Empty index or very different query | Verify content is indexed, adjust query |
| Slow queries | Very large result set or index | Add filters, reduce top_k |
| Index not found | Wrong index name or region | Verify index name and API key |
| Rate limits | Too many requests | Add delays or upgrade plan |
Best Practices
- Chunk documents properly - Split large documents into meaningful sections
- Keep metadata consistent - Use the same fields across all vectors
- Use namespaces for multi-tenancy - Separate data by customer or category
- Clean up unused vectors - Delete outdated content regularly
- Test with real queries - Validate search quality with actual user questions