What is the Mistral AI MCP Server?

The Mistral AI MCP Server gives AI agents like Claude, ChatGPT, and Cursor direct access to Mistral AI. Access Mistral AI models via API — chat with Claude alternatives, generate embeddings, moderate content and manage batch jobs from any AI agent. Powered by the Vinkius AI Gateway — no API keys, no infrastructure, connect in under 2 minutes.

Mistral AI MCP Server: see your AI Agent in action

AI AgentVinkiusMistral AI
You

Vinkius AI Gateway
GDPR·High Security·Kill Switch·Ultra-Low Latency·Plug and Play

Built-in capabilities (10)

cancel_batch

Provide the batch ID. This is useful if you submitted a large batch by mistake and want to stop further processing. Cancel a running batch job

chat

Requires the model ID (e.g. "mistral-large-latest", "mistral-small-latest", "codestral-latest") and messages array in JSON format. Each message must have a "role" ("user", "assistant" or "system") and "content" (text). Optionally set max_tokens, temperature (0-1), top_p (0-1) and tools array for function calling. Returns the assistant's response. Send a chat message to a Mistral model

create_batch

Requires the input file ID (containing JSONL requests) and the endpoint (e.g. "/v1/chat/completions"). Returns the batch with its ID for tracking. Use list_batches and get_batch to monitor progress. Create a batch processing job

delete_file

Provide the file ID from list_files. WARNING: this action is irreversible. Delete an uploaded file from Mistral

embeddings

Requires the model ID and text input (string or array of strings). Returns embedding vectors for each input text. Useful for semantic search, similarity comparison and vector database storage. Generate embeddings using Mistral

get_batch

Provide the batch ID. Get details for a specific batch job

list_batches

Each batch shows its ID, status (queued, running, succeeded, failed, cancelled), input/output file IDs and request counts. List batch processing jobs

list_files

Files are used for fine-tuning, batch processing and document AI. Each file shows its ID, filename, purpose, size and upload date. List files uploaded to Mistral

list_models

Each model returns its ID (e.g. "mistral-large-latest", "mistral-small-latest", "codestral-latest"), display name, capabilities and context window. Use this to discover which models are available and their IDs for use with the chat tool. List all available Mistral AI models

moderate

). Requires the input text (string or array). Returns safety scores for each category. Useful for content filtering and safety checks before processing user input. Moderate text content with Mistral

What this connector unlocks

Connect your Mistral AI account to any AI agent and leverage European-built AI models through natural conversation.

What you can do

  • Model Discovery — List all available Mistral models with their IDs, capabilities and context windows
  • Chat Completions — Send conversations to Mistral models (large, small, codestral, nemo) and receive responses with configurable parameters
  • Embeddings — Generate vector embeddings for semantic search, similarity comparison and vector storage
  • Content Moderation — Check text for harmful categories (violence, hate, sexual, self-harm) with safety scores
  • File Management — List and delete uploaded files used for batch processing and document AI
  • Batch Processing — Create, track and cancel batch jobs for cost-effective asynchronous processing

How it works

1. Subscribe to this server
2. Enter your Mistral AI API Key
3. Start using Mistral models from Claude, Cursor, or any MCP-compatible client

No more switching between API tools to interact with Mistral. Your AI acts as an LLM orchestration layer.

Who is this for?

  • Developers — quickly send messages to Mistral models, generate embeddings and moderate content without writing HTTP code
  • ML Engineers — discover available models, compare capabilities and batch-process many prompts efficiently
  • Content Teams — review model outputs, moderate user-generated content and manage batch processing jobs via conversation

Frequently asked questions

Give your AI agents the power of Mistral AI

Access Mistral AI and 2,500+ MCP servers — ready for your agents to use, right now. No glue code. No custom integrations. Just plug Vinkius AI Gateway and let your agents work.