Data Ingestion & Knowledge Sources
Document formats – PDF, DOCX, PPTX, CSV, TXT, HTML; 5MB free tier limit
Website crawling – Hundreds of thousands of pages indexed under 5 minutes
Google Drive – Native integration with real-time sync for cloud content
SQL databases – MySQL, PostgreSQL, Oracle, SQL Server, AWS/Azure/Google Cloud SQL
⚠️ YouTube, Dropbox, Notion, OneDrive – Zapier middleware required (no native integration)
Document support – PDF, DOCX, HTML automatically indexed (Vectara Platform )
Auto-sync connectors – Cloud storage and enterprise system integrations keep data current
Embedding processing – Background conversion to embeddings enables fast semantic search
1,400+ file formats – PDF, DOCX, Excel, PowerPoint, Markdown, HTML + auto-extraction from ZIP/RAR/7Z archives
Website crawling – Sitemap indexing with configurable depth for help docs, FAQs, and public content
Multimedia transcription – AI Vision, OCR, YouTube/Vimeo/podcast speech-to-text built-in
Cloud integrations – Google Drive, SharePoint, OneDrive, Dropbox, Notion with auto-sync
Knowledge platforms – Zendesk, Freshdesk, HubSpot, Confluence, Shopify connectors
Massive scale – 60M words (Standard) / 300M words (Premium) per bot with no performance degradation
Hybrid Retrieval Architecture ( Core Differentiator)
✅ Three-component system – Elasticsearch + Milvus vectors + XGBoost ML reranking
75.33 NDCG@10 – MTEB vs 73.16 pure vector (3% improvement)
96.50% Recall@20 – Anthropic benchmark vs 90.06% baseline
Models – snowflake-arctic-embed-m, bge-en-icl, voyage-2, OpenAI text-embedding-3-large
✅ Key finding – Open-source models match/exceed paid alternatives in benchmarks
N/A
N/A
98.3% response accuracy – Claimed with 1.2-second average response
Source citation – Visual PDF highlighting with page-level references
⚠️ No published uptime SLA – Service reliability not documented
✅ Enterprise scale – Millisecond responses with heavy traffic (benchmarks )
✅ Hybrid search – Semantic and keyword matching for pinpoint accuracy
✅ Hallucination prevention – Advanced reranking with factual-consistency scoring
Sub-second responses – Optimized RAG with vector search and multi-layer caching
Benchmark-proven – 13% higher accuracy, 34% faster than OpenAI Assistants API
Anti-hallucination tech – Responses grounded only in your provided content
OpenGraph citations – Rich visual cards with titles, descriptions, images
99.9% uptime – Auto-scaling infrastructure handles traffic spikes
Developer Experience ( A P I & S D Ks)
REST API + GraphQL – Bearer token auth with scored passage responses
denser-retriever – MIT-licensed Python package (261 stars, 30 forks)
Docker Compose – Full stack with Elasticsearch and Milvus containers
⚠️ Self-hosted "not production suitable" – Requires additional persistence and secrets config
Rate limits – 200 API calls/month on free tier
Multi-language SDKs – C#, Python, Java, JavaScript with REST API (FAQs )
Clear documentation – Sample code and guides for integration, indexing operations
Secure authentication – Azure AD or custom auth setup for API access
REST API – Full-featured for agents, projects, data ingestion, chat queries
Python SDK – Open-source customgpt-client with full API coverage
Postman collections – Pre-built requests for rapid prototyping
Webhooks – Real-time event notifications for conversations and leads
OpenAI compatible – Use existing OpenAI SDK code with minimal changes
Supported LLMs – GPT-4o, GPT-4o mini, GPT-3.5, Claude (version unspecified)
API keys – Users provide OpenAI or Claude keys via environment
⚠️ No custom fine-tuning – No private model hosting documented
Mockingbird default – In-house model with GPT-4/GPT-3.5 via Azure OpenAI available
Flexible selection – Choose model balancing cost versus quality for use case
Custom prompts – Prompt templates configurable for tone, format, citation rules
GPT-5.1 models – Latest thinking models (Optimal & Smart variants)
GPT-4 series – GPT-4, GPT-4 Turbo, GPT-4o available
Claude 4.5 – Anthropic's Opus available for Enterprise
Auto model routing – Balances cost/performance automatically
Zero API key management – All models managed behind the scenes
Website deployment – JavaScript widget (single line), iFrame, REST API
WordPress – Official plugin with page-specific targeting for no-code install
Zapier – 6,000+ apps with lead form triggers and events
⚠️ No native Slack, Teams, Discord – WhatsApp via Zapier only
⚠️ CRM via Zapier only – HubSpot, Salesforce, Zendesk not native
REST APIs & SDKs – Easy integration into custom applications with comprehensive tooling
Embedded experiences – Search/chat widgets for websites, mobile apps, custom portals
Low-code connectors – Azure Logic Apps and PowerApps simplify workflow integration
Website embedding – Lightweight JS widget or iframe with customizable positioning
CMS plugins – WordPress, WIX, Webflow, Framer, SquareSpace native support
5,000+ app ecosystem – Zapier connects CRMs, marketing, e-commerce tools
MCP Server – Integrate with Claude Desktop, Cursor, ChatGPT, Windsurf
OpenAI SDK compatible – Drop-in replacement for OpenAI API endpoints
LiveChat + Slack – Native chat widgets with human handoff capabilities
Drag-and-drop builder – Theme colors, logos, button sizing, bubbles
Custom domains – Available on paid tiers for white-labeling
Welcome messages – Configure suggested questions and greetings
White-label control – Full theming, logos, CSS customization for brand alignment
Domain restrictions – Bot scope and branding configurable per deployment
Search UI styling – Result cards and search interface match company identity
Full white-labeling included – Colors, logos, CSS, custom domains at no extra cost
2-minute setup – No-code wizard with drag-and-drop interface
Persona customization – Control AI personality, tone, response style via pre-prompts
Visual theme editor – Real-time preview of branding changes
Domain allowlisting – Restrict embedding to approved sites only
No- Code Interface & Usability
Visual builder – Drag-and-drop theme customization without coding
Setup – Single line JavaScript; WordPress plugin for no-code
⚠️ Learning curve – Documentation fragmented across multiple sites
⚠️ ~4-person team – Impacts enterprise support capacity
Azure portal UI – Straightforward index management and settings configuration interface
Low-code options – PowerApps, Logic Apps connectors enable quick non-dev integration
⚠️ Technical complexity – Advanced indexing tweaks require developer expertise vs turnkey tools
2-minute deployment – Fastest time-to-value in the industry
Wizard interface – Step-by-step with visual previews
Drag-and-drop – Upload files, paste URLs, connect cloud storage
In-browser testing – Test before deploying to production
Zero learning curve – Productive on day one
Integrated lead capture – Configurable fields (name, email, company, role, phone)
Conversation-triggered forms – Dynamic deployment based on conversation context
Analytics dashboard – Lead quality, satisfaction scores, conversion trends
✅ 24.8% conversion rate – Claimed on homepage demonstrating effectiveness
N/A
N/A
Multi- Language & Localization
80+ languages – Automatic language detection for global deployments
Multilingual rerankers – jinaai/jina-reranker-v2-base-multilingual support
N/A
N/A
Conversation history – 30-360 days retention by tier
Human handoff – Triggers when complexity exceeds scope
Escalation workflows – Zendesk ticket creation for handoffs
N/A
N/A
Observability & Monitoring
Conversation logs – Retention by tier (30-360 days)
User engagement tracking – Common queries, conversation length, drop-off points
⚠️ No A/B testing – No third-party BI integration (Tableau, PowerBI)
⚠️ No real-time alerting – No documented SLA tracking
Azure portal dashboard – Query latency, index health, usage metrics at-a-glance
Azure Monitor integration – Azure Monitor and App Insights for custom alerts
API log exports – Metrics exportable via API for compliance, analysis reports
Real-time dashboard – Query volumes, token usage, response times
Customer Intelligence – User behavior patterns, popular queries, knowledge gaps
Conversation analytics – Full transcripts, resolution rates, common questions
Export capabilities – API export to BI tools and data warehouses
S Q L Database Chat ( Unique Feature)
✅ Direct SQL connectivity – Conversational BI across major databases
Supported databases – MySQL, PostgreSQL, Oracle, SQL Server, AWS/Azure/Google Cloud SQL
Natural language to SQL – Ask questions, receive database query results
AES-256 encryption – Secure connections with read-only access requirement
N/A
N/A
Free – $0: 1 chatbot, 20 queries/month, 5MB limit
Starter – $19-29/month: 2 chatbots, 1,500 queries/month, 30-day logs
Standard – $89-119/month: 4 chatbots, 7,500 queries/month, custom domain
Business – $399-799/month: 8 chatbots, 15,000 queries/month, priority support
Enterprise – Custom: Private cloud, dedicated support, AWS Marketplace
⚠️ User feedback – "Plans quite restrictive, credit limits reached sooner"
Usage-based pricing – Free tier available, bundles scale with growth (pricing )
Enterprise tiers – Plans scale with query volume, data size for heavy usage
Dedicated deployment – VPC or on-prem options for data isolation requirements
Standard: $99/mo – 60M words, 10 bots
Premium: $449/mo – 300M words, 100 bots
Auto-scaling – Managed cloud scales with demand
Flat rates – No per-query charges
⚠️ NO SOC 2, HIPAA, ISO 27001, GDPR certifications – Not for regulated industries
Private cloud deployments – Enterprise tier for data sovereignty
AES-256 encryption – Database connections with read-only access
AWS infrastructure – Data storage and processing on AWS
✅ Data encryption – Transit and rest encryption, no model training on your content
✅ Compliance certifications – SOC 2, ISO, GDPR, HIPAA (details )
✅ Customer-managed keys – BYOK support with private deployments for full control
SOC 2 Type II + GDPR – Third-party audited compliance
Encryption – 256-bit AES at rest, SSL/TLS in transit
Access controls – RBAC, 2FA, SSO, domain allowlisting
Data isolation – Never trains on your data
✅ denser-retriever – MIT-licensed, 261 GitHub stars, full RAG transparency
Docker Compose deployment – Local experimentation with Elasticsearch and Milvus
Validate benchmarks – Test embeddings, rerankers, chunking on own data
⚠️ Self-hosted "not production suitable" – Denser recommends managed SaaS
N/A
N/A
Founded 2023 – Silicon Valley startup, ~4 employees (bootstrapped)
✅ Founder Zhiheng Huang – Former Amazon Kendra scientist, Amazon Q lead
70+ research papers – 14,000+ citations; BLSTM-CRF 5,400+ citations
N/A
N/A
R A G-as-a- Service Assessment
✅ TRUE RAG PLATFORM – Hybrid retrieval with open-source transparency
Data source flexibility – Good (documents, websites, Google Drive, SQL)
LLM model options – Good (GPT-4o, Claude, multiple embeddings/rerankers)
✅ Open-source transparency – Excellent (MIT-licensed core, published benchmarks)
⚠️ Compliance & certifications – Poor (no SOC 2, HIPAA, ISO 27001)
Best for – Technical teams prioritizing retrieval accuracy and validation
Platform Type – TRUE ENTERPRISE RAG-AS-A-SERVICE: Agent OS for trusted AI
Core Mission – Deploy AI assistants/agents with grounded answers, safe actions, always-on governance
Target Market – Enterprises needing production RAG, white-label APIs, VPC/on-prem deployments
RAG Implementation – Mockingbird LLM (26% better than GPT-4), hybrid search, multi-stage reranking
API-First Architecture – REST APIs, SDKs (C#/Python/Java/JS), Azure integration (Logic Apps/Power BI)
Security & Compliance – SOC 2 Type 2, ISO 27001, GDPR, HIPAA, BYOK, VPC/on-prem
Agent-Ready Platform – Python library, Agent APIs, structured outputs, audit trails, policy enforcement
Advanced RAG Features – Hybrid search, reranking, HHEM scoring, multilingual retrieval (7 languages)
Funding – $53.5M raised ($25M Series A July 2024, FPV/Race Capital)
⚠️ Enterprise complexity – Requires developer expertise for indexing, tuning, agent configuration
⚠️ No no-code builder – Azure portal management but no drag-and-drop chatbot builder
⚠️ Azure ecosystem focus – Best with Azure, less smooth for AWS/GCP cross-cloud flexibility
Use Case Fit – Mission-critical RAG, regulated industries (SOC 2/HIPAA), white-label APIs, VPC/on-prem
Platform type – TRUE RAG-AS-A-SERVICE with managed infrastructure
API-first – REST API, Python SDK, OpenAI compatibility, MCP Server
No-code option – 2-minute wizard deployment for non-developers
Hybrid positioning – Serves both dev teams (APIs) and business users (no-code)
Enterprise ready – SOC 2 Type II, GDPR, WCAG 2.0, flat-rate pricing
vs CustomGPT – Superior retrieval transparency, SQL chat; gaps in compliance
vs Glean – Open-source vs proprietary, lower cost; lacks permissions-aware AI
✅ Unique strengths – Hybrid retrieval benchmarks, founder pedigree, SQL chat
Target audience – Developers building AI chatbots without strict compliance
Market position – Enterprise RAG platform between Azure AI Search and chatbot builders
Target customers – Enterprises needing production RAG, white-label APIs, VPC/on-prem deployments
Key competitors – Azure AI Search, Coveo, OpenAI Enterprise, Pinecone Assistant
Competitive advantages – Mockingbird LLM, hallucination detection, SOC 2/HIPAA compliance, millisecond responses
Pricing advantage – Usage-based with free tier, best value for enterprise RAG infrastructure
Use case fit – Mission-critical RAG, white-label APIs, Azure integration, high-accuracy requirements
Market position – Leading RAG platform balancing enterprise accuracy with no-code usability. Trusted by 6,000+ orgs including Adobe, MIT, Dropbox.
Key differentiators – #1 benchmarked accuracy • 1,400+ formats • Full white-labeling included • Flat-rate pricing
vs OpenAI – 10% lower hallucination, 13% higher accuracy, 34% faster
vs Botsonic/Chatbase – More file formats, source citations, no hidden costs
vs LangChain – Production-ready in 2 min vs weeks of development
✅ Hybrid retrieval – ES + Milvus vectors + XGBoost reranking
75.33 NDCG@10 on MTEB – vs 73.16 pure vector (3% improvement)
96.50% Recall@20 – Anthropic benchmark vs 90.06% baseline
Source citation – Visual PDF highlighting with page references
98.3% accuracy claimed – 1.2-second average response time
✅ Hybrid search – Semantic vector + BM25 keyword matching for pinpoint accuracy
✅ Advanced reranking – Multi-stage pipeline optimizes results before generation with relevance scoring
✅ Factual scoring – HHEM provides reliability score for every response's grounding quality
✅ Citation precision – Mockingbird outperforms GPT-4 on citation metrics, traceable to sources
Multilingual RAG – Cross-lingual: query/retrieve/generate in different languages (7 supported)
Structured outputs – Extract specific information for autonomous agent integration, deterministic data
GPT-4 + RAG – Outperforms OpenAI in independent benchmarks
Anti-hallucination – Responses grounded in your content only
Automatic citations – Clickable source links in every response
Sub-second latency – Optimized vector search and caching
Scale to 300M words – No performance degradation at scale
Customer support chatbots – Website deployment with 24.8% conversion rate
✅ SQL database chat (unique) – Natural language queries against major databases
Technical documentation – Hundreds of thousands of pages indexed under 5 minutes
Multilingual support – 80+ languages with automatic detection
Developer-focused RAG – MIT-licensed denser-retriever for validation
Regulated industries – Health, legal, finance needing accuracy, security, SOC 2 compliance
Enterprise knowledge – Q&A systems with precise answers from large document repositories
Autonomous agents – Structured outputs for deterministic data extraction, decision-making workflows
White-label APIs – Customer-facing search/chat with millisecond responses at enterprise scale
Multilingual support – 7 languages with single knowledge base for multiple locales
High accuracy needs – Citation precision, factual scoring, 0.9% hallucination rate (Mockingbird-2-Echo)
Customer support – 24/7 AI handling common queries with citations
Internal knowledge – HR policies, onboarding, technical docs
Sales enablement – Product info, lead qualification, education
Documentation – Help centers, FAQs with auto-crawling
E-commerce – Product recommendations, order assistance
Documentation – docs.denser.ai, retriever.denser.ai, GitHub READMEs
⚠️ Documentation fragmented – Information scattered across multiple sites
~4-person team – Impacts enterprise support capacity
Open-source community – 261 GitHub stars, 30 forks, MIT license
Enterprise support – Dedicated channels and SLA-backed help for enterprise customers
Microsoft network – Extensive infrastructure, forums, technical guides backed by Microsoft
Comprehensive docs – API references, integration guides, SDKs at docs.vectara.com
Sample code – Pre-built examples, Jupyter notebooks, quick-start guides for rapid integration
Active community – Developer forums for peer support, knowledge sharing, best practices
Documentation hub – Docs, tutorials, API references
Support channels – Email, in-app chat, dedicated managers (Premium+)
Open-source – Python SDK, Postman, GitHub examples
Community – User community + 5,000 Zapier integrations
Limitations & Considerations
⚠️ No compliance certifications – Missing SOC 2, HIPAA, ISO 27001, GDPR
⚠️ Small team (~4 people) – Potential scaling constraints for enterprise
⚠️ Heavy Zapier dependency – No native Slack, Teams, CRM integrations
⚠️ Fragmented documentation – Scattered across docs, retriever docs, GitHub
⚠️ User feedback – "Plans restrictive, credit limits reached sooner"
⚠️ Azure ecosystem focus – Best with Azure services, less smooth for AWS/GCP organizations
⚠️ Developer expertise needed – Advanced indexing requires technical skills vs turnkey no-code tools
⚠️ No drag-and-drop GUI – Azure portal management but no chatbot builder like Tidio/WonderChat
⚠️ Limited model selection – Mockingbird/GPT-4/GPT-3.5 only, no Claude/Gemini/custom models
⚠️ Sales-driven pricing – Contact sales for enterprise pricing, less transparent than self-serve platforms
⚠️ Overkill for simple bots – Enterprise RAG unnecessary for basic FAQ or customer service
Managed service – Less control over RAG pipeline vs build-your-own
Model selection – OpenAI + Anthropic only; no Cohere, AI21, open-source
Real-time data – Requires re-indexing; not ideal for live inventory/prices
Enterprise features – Custom SSO only on Enterprise plan
AI agent capabilities – Process data for intelligent automation with customization
Multi-platform deployment – Launch across websites and messaging with single line
Adaptive learning – Chatbot learns over time using conversation analysis
24/7 availability – Smart AI support with instant answers
Agentic RAG Framework – Python library for autonomous agents: emails, bookings, system integration
Agent APIs (Tech Preview) – Customizable reasoning models, behavioral instructions, tool access controls
LlamaIndex integration – Rapid tool creation connecting Vectara corpora, single-line code generation
Multi-LLM support – OpenAI, Anthropic, Gemini, GROQ, Together.AI, Cohere, AWS Bedrock integration
Step-level audit trails – Source citations, reasoning steps, decision paths for governance compliance
✅ Grounded actions – Document-grounded decisions with citations, 0.9% hallucination rate (Mockingbird-2-Echo)
⚠️ Developer platform – Requires programming expertise, not for non-technical teams
⚠️ No chatbot UI – No polished widgets or turnkey conversational interfaces
⚠️ Tech preview status – Agent APIs subject to change before general availability
Custom AI Agents – Autonomous GPT-4/Claude agents for business tasks
Multi-Agent Systems – Specialized agents for support, sales, knowledge
Memory & Context – Persistent conversation history across sessions
Tool Integration – Webhooks + 5,000 Zapier apps for automation
Continuous Learning – Auto re-indexing without manual retraining
Conversational interface – Chat with customers in friendly manner
Business knowledge integration – Trained on documents, websites, Google Drive
Multi-language support – 80+ languages with automatic detection
Lead capture – Integrated forms (name, email, company, role)
Human handoff – Triggers on complexity with Zendesk tickets
Vector + LLM search – Smart retrieval with generative answers, context-aware responses
Mockingbird LLM – Proprietary model with source citations (details )
Multi-turn conversations – Conversation history tracking for smooth back-and-forth dialogue
✅ #1 accuracy – Median 5/5 in independent benchmarks, 10% lower hallucination than OpenAI
✅ Source citations – Every response includes clickable links to original documents
✅ 93% resolution rate – Handles queries autonomously, reducing human workload
✅ 92 languages – Native multilingual support without per-language config
✅ Lead capture – Built-in email collection, custom forms, real-time notifications
✅ Human handoff – Escalation with full conversation context preserved
Customization & Flexibility ( Behavior & Knowledge)
Behavior customization – Define name, tone, response preferences
File support – PDF, DOCX, XLSX, PPTX, TXT, HTML, CSV, XML
Website crawling – Train bot by crawling URLs for knowledge base
Easy knowledge updates – Add documents, re-crawl, update without rebuild
Flexible deployment – Web widget, dashboard, or API integration
Indexing control – Configure chunk sizes, metadata tags, retrieval parameters
Search weighting – Tune semantic vs lexical search balance per query
Domain tuning – Adjust prompt templates and relevance thresholds for specialty domains
Live content updates – Add/remove content with automatic re-indexing
System prompts – Shape agent behavior and voice through instructions
Multi-agent support – Different bots for different teams
Smart defaults – No ML expertise required for custom behavior
N/A
Microsoft network – Comprehensive docs, forums, technical guides backed by Microsoft
Enterprise support – Dedicated channels and SLA-backed help for enterprise plans
Azure ecosystem – Broad partner network and active developer community access
Comprehensive docs – Tutorials, cookbooks, API references
Email + in-app support – Under 24hr response time
Premium support – Dedicated account managers for Premium/Enterprise
Open-source SDK – Python SDK, Postman, GitHub examples
5,000+ Zapier apps – CRMs, e-commerce, marketing integrations
Additional Considerations N/A
✅ Factual scoring – Hybrid search with reranking provides unique factual-consistency scores
Flexible deployment – Public cloud, VPC, or on-prem for varied compliance needs
Active development – Regular feature releases and integrations keep platform current
Time-to-value – 2-minute deployment vs weeks with DIY
Always current – Auto-updates to latest GPT models
Proven scale – 6,000+ organizations, millions of queries
Multi-LLM – OpenAI + Claude reduces vendor lock-in
N/A
✅ Mockingbird LLM – 26% better than GPT-4 on BERT F1, 0.9% hallucination rate
✅ Mockingbird 2 – 7 languages (EN/ES/FR/AR/ZH/JA/KO), under 10B parameters
GPT-4/GPT-3.5 fallback – Azure OpenAI integration for OpenAI model preference
HHEM + HCM – Hughes Hallucination Evaluation with Correction Model (Mockingbird-2-Echo)
✅ No training on data – Customer data never used for model training/improvement
Custom prompts – Templates configurable for tone, format, citation rules per domain
OpenAI – GPT-5.1 (Optimal/Smart), GPT-4 series
Anthropic – Claude 4.5 Opus/Sonnet (Enterprise)
Auto-routing – Intelligent model selection for cost/performance
Managed – No API keys or fine-tuning required
N/A
✅ SOC 2 Type 2 – Independent audit demonstrating enterprise-grade operational security controls
✅ ISO 27001 + GDPR – Information security management with EU data protection compliance
✅ HIPAA ready – Healthcare compliance with BAAs available for PHI handling
✅ Encryption – TLS 1.3 in transit, AES-256 at rest with BYOK support
✅ Zero data retention – No model training on customer data, content stays private
Private deployments – VPC or on-premise for data sovereignty and network isolation
SOC 2 Type II + GDPR – Regular third-party audits, full EU compliance
256-bit AES encryption – Data at rest; SSL/TLS in transit
SSO + 2FA + RBAC – Enterprise access controls with role-based permissions
Data isolation – Never trains on customer data
Domain allowlisting – Restrict chatbot to approved domains
N/A
30-day free trial – Full enterprise feature access for evaluation before commitment
Usage-based pricing – Pay for query volume and data size with scalable tiers
Free tier – Generous free tier for development, prototyping, small production deployments
Enterprise pricing – Custom pricing for VPC/on-prem installations, heavy usage bundles available
✅ Transparent pricing – No per-seat charges, storage surprises, or model switching fees
Funding – $53.5M raised ($25M Series A July 2024, FPV/Race Capital)
Standard: $99/mo – 10 chatbots, 60M words, 5K items/bot
Premium: $449/mo – 100 chatbots, 300M words, 20K items/bot
Enterprise: Custom – SSO, dedicated support, custom SLAs
7-day free trial – Full Standard access, no charges
Flat-rate pricing – No per-query charges, no hidden costs
Join the Discussion
Loading comments...