Perplexity AI MCP Server for LlamaIndex 14 tools — connect in under 2 minutes
LlamaIndex specializes in data-aware AI agents that connect LLMs to structured and unstructured sources. Add Perplexity AI as an MCP tool provider through the Vinkius and your agents can query, analyze, and act on live data alongside your existing indexes.
ASK AI ABOUT THIS MCP SERVER
Vinkius supports streamable HTTP and SSE.
import asyncio
from llama_index.tools.mcp import BasicMCPClient, McpToolSpec
from llama_index.core.agent.workflow import FunctionAgent
from llama_index.llms.openai import OpenAI
async def main():
# Your Vinkius token — get it at cloud.vinkius.com
mcp_client = BasicMCPClient("https://edge.vinkius.com/[YOUR_TOKEN_HERE]/mcp")
mcp_tool_spec = McpToolSpec(client=mcp_client)
tools = await mcp_tool_spec.to_tool_list_async()
agent = FunctionAgent(
tools=tools,
llm=OpenAI(model="gpt-4o"),
system_prompt=(
"You are an assistant with access to Perplexity AI. "
"You have 14 tools available."
),
)
response = await agent.run(
"What tools are available in Perplexity AI?"
)
print(response)
asyncio.run(main())
* Every MCP server runs on Vinkius-managed infrastructure inside AWS - a purpose-built runtime with per-request V8 isolates, Ed25519 signed audit chains, and sub-40ms cold starts optimized for native MCP execution. See our infrastructure
About Perplexity AI MCP Server
Connect your Perplexity AI API key to any AI agent and harness the power of real-time web search with AI-generated answers, citations, and related questions through natural conversation.
LlamaIndex agents combine Perplexity AI tool responses with indexed documents for comprehensive, grounded answers. Connect 14 tools through the Vinkius and query live data alongside vector stores and SQL databases in a single turn — ideal for hybrid search, data enrichment, and analytical workflows.
What you can do
- Answer Questions — Ask any question and get grounded answers with real-time web search and source citations
- Deep Research — Perform exhaustive research on complex topics with comprehensive reports and thorough citations
- Logical Reasoning — Solve complex problems requiring step-by-step analysis and chain-of-thought reasoning
- Domain-Filtered Search — Restrict search results to specific domains for academic, technical, or trusted-source queries
- Recency Filtering — Get answers based on recent information only (hour, day, week, month, or year)
- Multi-Turn Conversations — Maintain context across multiple questions for iterative research sessions
- Structured Output — Get responses in JSON format following a defined schema for programmatic integration
- Visual Results — Include relevant images and related questions in search results
The Perplexity AI MCP Server exposes 14 tools through the Vinkius. Connect it to LlamaIndex in under two minutes — no API keys to rotate, no infrastructure to provision, no vendor lock-in. Your configuration, your data, your control.
How to Connect Perplexity AI to LlamaIndex via MCP
Follow these steps to integrate the Perplexity AI MCP Server with LlamaIndex.
Install dependencies
Run pip install llama-index-tools-mcp llama-index-llms-openai
Replace the token
Replace [YOUR_TOKEN_HERE] with your Vinkius token
Run the agent
Save to agent.py and run: python agent.py
Explore tools
The agent discovers 14 tools from Perplexity AI
Why Use LlamaIndex with the Perplexity AI MCP Server
LlamaIndex provides unique advantages when paired with Perplexity AI through the Model Context Protocol.
Data-first architecture: LlamaIndex agents combine Perplexity AI tool responses with indexed documents for comprehensive, grounded answers
Query pipeline framework lets you chain Perplexity AI tool calls with transformations, filters, and re-rankers in a typed pipeline
Multi-source reasoning: agents can query Perplexity AI, a vector store, and a SQL database in a single turn and synthesize results
Observability integrations show exactly what Perplexity AI tools were called, what data was returned, and how it influenced the final answer
Perplexity AI + LlamaIndex Use Cases
Practical scenarios where LlamaIndex combined with the Perplexity AI MCP Server delivers measurable value.
Hybrid search: combine Perplexity AI real-time data with embedded document indexes for answers that are both current and comprehensive
Data enrichment: query Perplexity AI to augment indexed data with live information before generating user-facing responses
Knowledge base agents: build agents that maintain and update knowledge bases by periodically querying Perplexity AI for fresh data
Analytical workflows: chain Perplexity AI queries with LlamaIndex's data connectors to build multi-source analytical reports
Perplexity AI MCP Tools for LlamaIndex (14)
These 14 tools become available when you connect Perplexity AI to LlamaIndex via MCP:
chat_completion
The Sonar model searches the web, synthesizes information, and provides a concise answer. This is the basic query tool for factual questions, summaries, and general knowledge. Use this for quick lookups where you need accurate, up-to-date information. Ask Perplexity AI a question and get a grounded, cited answer
chat_with_citations
Each claim or fact in the response is linked to its original source. This is essential for research, fact-checking, and academic work where sources matter. The response includes a citations array with URLs of all referenced sources. Ask Perplexity AI and get answers with source citations
chat_with_domain_filter
Provide domains as a comma-separated list (e.g., "arxiv.org,nih.gov,github.com"). Only sources from the specified domains will be used in generating the answer. Use this for domain-specific research, academic papers, or trusted sources only. Citations are automatically included to verify sources. Ask Perplexity AI restricting search to specific domains
chat_with_history
Provide messages as a JSON array of {role: "user"|"assistant"|"system", content: "text"} objects. This enables follow-up questions where the model understands previous context. Use this for complex queries that build on previous answers or require contextual understanding. Example: [{ "role": "user", "content": "What is quantum computing?" }, { "role": "assistant", "content": "Quantum computing uses quantum bits..." }, { "role": "user", "content": "How does it differ from classical computing?" }] Ask Perplexity AI with multi-turn conversation history
chat_with_images
The response includes an images array with URLs to related images found during the search. Use this for visual topics, product searches, or when you need images to accompany the answer. Ask Perplexity AI and get relevant images with the answer
chat_with_recency_filter
Available recency filters: "hour", "day", "week", "month", "year". This ensures the answer is based on recent information only. Use this for news, recent events, or time-sensitive queries where outdated info is not useful. Ask Perplexity AI with results filtered by time recency
chat_with_related_questions
The response includes a related_questions array with suggested questions for further exploration. Use this for research, learning, and discovering related topics you might want to explore. Ask Perplexity AI and get related follow-up questions
deep_research
This model performs extensive web searches and generates detailed reports with thorough citations. It takes longer than regular queries but provides much more depth and breadth. Use this for complex topics, literature reviews, competitive analysis, or thorough investigations. Maximum tokens default to 4096 for comprehensive responses. Perform deep research with exhaustive web search and comprehensive report
follow_up
Provide the conversation history as a JSON array of messages and the follow-up question. This maintains context from previous turns in the conversation. Use this for multi-turn research sessions where each question builds on previous answers. Ask a follow-up question in an ongoing conversation with Perplexity AI
list_models
Use this to discover what models are available before choosing which one to use for your queries. List all available Perplexity AI models
reasoning
This model excels at multi-step reasoning, mathematical problems, code analysis, and chain-of-thought tasks. Use this for problems requiring step-by-step analysis, mathematical proofs, code reviews, or logical deductions. Citations are included where external information is referenced. Ask Perplexity AI for complex logical reasoning and step-by-step analysis
search_query
This combines all search features: cited sources, relevant images, and follow-up questions. Use this when you want the fullest possible search result with all supplementary information. The response includes content, citations array, images array, and related_questions array. Perform a comprehensive web search with citations, images, and related questions
structured_query
The model will return the answer as JSON matching your schema definition. Provide the JSON schema as a string. This is useful for programmatic data extraction, API integrations, and when you need consistent, parseable responses. Example schema: { "type": "object", "properties": { "name": { "type": "string" }, "age": { "type": "number" } } } Ask Perplexity AI and get a structured JSON response following a schema
system_prompt_query
The system prompt defines how the model should respond (e.g., "You are a medical expert...", "Answer in bullet points..."). Use this for specialized queries, role-playing, formatting requirements, or domain-specific expertise. Example system prompt: "You are a senior software architect. Explain concepts with code examples." Ask Perplexity AI with a custom system prompt to set behavior and context
Example Prompts for Perplexity AI in LlamaIndex
Ready-to-use prompts you can give your LlamaIndex agent to start working with Perplexity AI immediately.
"What are the latest developments in quantum computing as of this week?"
"Do deep research on the competitive landscape of electric vehicle manufacturers in Southeast Asia, including market share, pricing strategies, and government incentives."
"Search for news about AI regulation in the European Union from the last month, restricted to europa.eu and reuters.com domains."
Troubleshooting Perplexity AI MCP Server with LlamaIndex
Common issues when connecting Perplexity AI to LlamaIndex through the Vinkius, and how to resolve them.
BasicMCPClient not found
pip install llama-index-tools-mcpPerplexity AI + LlamaIndex FAQ
Common questions about integrating Perplexity AI MCP Server with LlamaIndex.
How does LlamaIndex connect to MCP servers?
Can I combine MCP tools with vector stores?
Does LlamaIndex support async MCP calls?
Connect Perplexity AI with your favorite client
Step-by-step setup guides for every MCP-compatible client and framework:
Anthropic's native desktop app for Claude with built-in MCP support.
AI-first code editor with integrated LLM-powered coding assistance.
GitHub Copilot in VS Code with Agent mode and MCP support.
Purpose-built IDE for agentic AI coding workflows.
Autonomous AI coding agent that runs inside VS Code.
Anthropic's agentic CLI for terminal-first development.
Python SDK for building production-grade OpenAI agent workflows.
Google's framework for building production AI agents.
Type-safe agent development for Python with first-class MCP support.
TypeScript toolkit for building AI-powered web applications.
TypeScript-native agent framework for modern web stacks.
Python framework for orchestrating collaborative AI agent crews.
Leading Python framework for composable LLM applications.
Data-aware AI agent framework for structured and unstructured sources.
Microsoft's framework for multi-agent collaborative conversations.
Connect Perplexity AI to LlamaIndex
Get your token, paste the configuration, and start using 14 tools in under 2 minutes. No API key management needed.
