Skip to main content

Platform Overview

The Co-mind.ai Private AI Platform provides enterprise-grade access to multiple AI models through a unified, OpenAI-compatible API. Deploy on your own infrastructure with full data sovereignty.

Key Features

Multimodal Support

Process text, images, and audio inputs through a single API.

OpenAI Compatibility

Drop-in replacement for OpenAI API endpoints — use existing SDKs and tools.

Multiple AI Backends

Access vLLM, Ollama, llama.cpp, SambaNova, and more through one interface.

Real-time Streaming

Stream responses via Server-Sent Events for better user experience.

Function/Tool Calling

Extend AI capabilities with custom tools and function calls.

Security Sanitizer

Built-in injection/jailbreak detection, PII redaction, and configurable policies.

Platform Services

Core AI

ServiceDescription
Chat CompletionsOpenAI-compatible chat with streaming, vision, and tool calling
Text CompletionsPrompt-based text generation
EmbeddingsVector embeddings for semantic search and similarity

Knowledge Base (RAG)

ServiceDescription
Knowledge BasesUpload and manage document collections
Stateless ChatChat with knowledge bases — you manage conversation history
Stateful SessionsServer-managed conversation history with KB context
Context RetrievalQuery KBs directly for relevant document chunks

Echo Engine (Audio)

ServiceDescription
Transcription (STT)Synchronous and async audio transcription with speaker diarization
Text-to-Speech (TTS)Convert text to speech with multiple voices and languages
RecordingsManage audio recording metadata and link to transcription jobs

Research & Analysis

ServiceDescription
ResearcherMulti-provider web search, content scraping, deep research with iterative reasoning
Document AnalyzerExtract structured data from documents with human-in-the-loop review
Grant ProposalsAI-assisted grant proposal generation

Administration

ServiceDescription
Tenant ManagementMulti-tenant isolation with sub-organizations
Directory IntegrationLDAP/Active Directory and Microsoft Entra ID (Azure AD)
Security SanitizerConfigurable content safety policies
Audit LogsComprehensive audit trail with filters and aggregation

Supported Models

The platform is model-agnostic — deploy and serve any model through the supported backends. There are no restrictions on which models you can use.

Supported Backends

BackendDescriptionExample Models
vLLMHigh-performance GPU inferenceLlama, Falcon, Qwen, Mistral, Gemma, and any HuggingFace model
OllamaLocal model serverAny model from the Ollama library
llama.cppCPU/GPU inference with GGUF modelsAny GGUF-format model
SambaNovaSambaNova Cloud inferenceLlama, DeepSeek, Qwen, and SambaNova-hosted models
OpenAIOpenAI API passthroughGPT-4o, GPT-4, GPT-3.5, and all OpenAI models
AnthropicAnthropic API passthroughClaude 4, Claude 3.5, and all Anthropic models
Use GET /v1/models to list all available models on your instance and GET /v1/capabilities to check which features each backend supports.

Base URL

Your specific API endpoint URL will be provided with your credentials. The general format is:
https://your-comind-instance.example.com
All API endpoints are prefixed with /v1/.

Next Steps