Build & Deploy
Get your RAG application up and running quickly with our step-by-step guides. From initial setup to production deployment, we'll walk you through everything you need to know.
Quick Start Guide
1
Create Your Account
Sign up for a free Vectorize account and create your first workspace.
- Visit platform.vectorize.io
- Click "Sign Up" and complete registration
- Verify your email address
- Create your first workspace
2
Connect Your Data Source
Connect to your first data source to start building your knowledge base.
// Using the Vectorize CLI
vectorize sources add \
--type=google-drive \
--name="Company Docs" \
--folder-id="1ABC123DEF456" \
--sync=realtime
Supported data sources include:
- Google Drive, Dropbox, OneDrive
- Notion, Confluence, SharePoint
- GitHub repositories
- Direct file uploads (PDF, DOCX, etc.)
3
Configure Your Pipeline
Set up your embedding model, chunking strategy, and vector database.
// Pipeline configuration
{
"name": "knowledge-base-v1",
"embedding": {
"model": "text-embedding-ada-002",
"dimensions": 1536
},
"chunking": {
"strategy": "recursive",
"chunkSize": 1000,
"overlap": 200,
"separators": ["\n\n", "\n", ". ", " "]
},
"vectorDatabase": {
"provider": "pinecone",
"environment": "us-west1-gcp"
}
}
4
Test Your Pipeline
Run your first queries to ensure everything is working correctly.
// Test query using the API
curl -X POST https://api.vectorize.io/v1/search \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"query": "How do I reset my password?",
"pipeline_id": "your-pipeline-id",
"limit": 5
}'
5
Deploy Your Application
Integrate Vectorize into your application and deploy to production.
// Example Node.js integration
const { VectorizeClient } = require('@vectorize/sdk');
const vectorize = new VectorizeClient({
apiKey: process.env.VECTORIZE_API_KEY
});
app.post('/search', async (req, res) => {
const { query } = req.body;
const results = await vectorize.search({
query,
pipelineId: 'your-pipeline-id',
limit: 10
});
res.json(results);
});
Pipeline Configuration
Embedding Models
Choose the right embedding model for your use case:
- text-embedding-ada-002: General purpose, good for most applications
- voyage-large-2: Optimized for retrieval tasks
- all-MiniLM-L6-v2: Fast and lightweight for real-time applications
Chunking Strategies
Recommended Settings by Content Type
- Technical docs: 800 characters, 20% overlap
- Legal documents: 1200 characters, 15% overlap
- Marketing content: 600 characters, 25% overlap
- Code repositories: Function-based chunking
Vector Database Selection
- Pinecone: Managed, high performance, automatic scaling
- Weaviate: Open source, GraphQL API, hybrid search
- Qdrant: High performance, advanced filtering capabilities
- Milvus: Open source, enterprise features
Deployment Options
Production Checklist
Before deploying to production, ensure you've completed these steps:
- ✅ Test your pipeline with real data
- ✅ Set up monitoring and alerting
- ✅ Configure backup and disaster recovery
- ✅ Implement security best practices
- ✅ Set up rate limiting and caching
- ✅ Configure auto-scaling policies
- ✅ Test failover scenarios
- ✅ Document your deployment process
Next Steps
Once your application is deployed, learn how to:
- Monitor and manage your production systems
- Optimize performance and costs
- Set up automated testing and updates
- Scale your application as usage grows