MCP Integration with Orchestra
Orchestra provides integration with the Model Context Protocol (MCP), allowing your agents to use tools implemented in various programming languages and frameworks. The MCP adapter enables you to extend Orchestra's capabilities beyond Python and integrate with external services and libraries built by the community based on Model Context Protocol (MCP).
What is MCP?Copied!
The Model Context Protocol (MCP), developed by Anthropic, is a standardized protocol for communication between language models and external tools. It allows tools to be implemented in various programming languages while maintaining a consistent interface for language models to interact with them.
The MCPOrchestra AdapterCopied!
Orchestra includes an
MCPOrchestra
- Connects to MCP servers (which can be written in any language)
- Discovers available tools on those servers
- Converts those tools into Python callables with proper docstrings
- Makes them available to your Orchestra agents
Using MCP Tools with OrchestraCopied!
Basic Usage PatternCopied!
import asyncio from mainframe_orchestra import Task, Agent, OpenaiModels, MCPOrchestra, set_verbosity # Optional: Set verbosity level for debugging set_verbosity(2) async def main(): # Create the MCP adapter using async context manager async with MCPOrchestra() as mcp_client: # Connect to an MCP server await mcp_client.connect( server_name="my_server", command="command_to_run_server", args=["arg1", "arg2"], start_server=True, # Optional: start the server if it's not already running server_startup_delay=2.0 # Optional: wait for server to initialize ) # Get tools from the server tools = mcp_client.get_tools() # Create an agent with these tools agent = Agent( agent_id="mcp_agent", # Optional but recommended for identification role="Assistant", goal="Help with tasks using MCP tools", tools=tools, llm=OpenaiModels.gpt_4o ) # Create a task using the agent result = await Task.create( agent=agent, instruction="Your instruction here" ) print(result) if __name__ == "__main__": asyncio.run(main())
Available MCP Tool LibrariesCopied!
Orchestra can work with any MCP-compatible server. Here are some popular MCP tool libraries you can use:
Tool | Description | Installation | Example |
---|---|---|---|
FastMCP | Create Python-based MCP servers with minimal code |
|
Calculator Example |
Fetch MCP | Web fetching and scraping tools |
|
Fetch Example |
Filesystem MCP | File system operations |
|
Filesystem Example |
Playwright MCP | Web automation with Playwright |
|
Playwright Example |
Slack MCP | Slack API integration | See Slack Example | Slack Example |
ExamplesCopied!
Calculator Example with FastMCPCopied!
First, create a FastMCP server file:
# mcp_fast_calc.py from fastmcp import FastMCP # Create a FastMCP server with a name mcp = FastMCP("Calculator") # Define tools using decorators @mcp.tool() def add(a: int, b: int) -> int: """Add two numbers Args: a: The first number to add b: The second number to add Returns: The sum of the two numbers """ return a + b @mcp.tool() def multiply(a: int, b: int) -> int: """Multiply two numbers Args: a: The first number to multiply b: The second number to multiply Returns: The product of the two numbers """ return a * b # You can also define resource endpoints @mcp.resource("greeting://{name}") def get_greeting(name: str) -> str: """Get a personalized greeting Args: name: The name of the person to greet Returns: A personalized greeting """ return f"Hello, {name}!" if __name__ == "__main__": mcp.run()
Then, use it with Orchestra:
# mcp_test_fastmcp.py import asyncio from mainframe_orchestra import Task, Agent, OpenaiModels, MCPOrchestra, set_verbosity set_verbosity(2) # Enable verbose logging async def main(): async with MCPOrchestra() as mcp_client: # Start the FastMCP server and connect to it print("Starting and connecting to FastMCP server...") await mcp_client.connect( server_name="calculator", command="python", args=["mcp_fast_calc.py"], # Path to your FastMCP server script start_server=True, # Start the server process server_startup_delay=2.0 # Wait for server to initialize ) # Get calculator tools calculator_tools = mcp_client.get_tools() # Create a math assistant agent agent = Agent( agent_id="calculator_agent", role="Math Assistant", goal="Help users perform mathematical calculations", tools=calculator_tools, llm=OpenaiModels.gpt_4o ) # Create a task result = await Task.create( agent=agent, instruction="You need to calculate the sum of 42 and 58, and then multiply the result by 5. Respond with the final answer." ) print(result) if __name__ == "__main__": asyncio.run(main())
Web Fetching ExampleCopied!
# mcp_test_fetch.py import asyncio from mainframe_orchestra import Task, Agent, OpenaiModels, MCPOrchestra, set_verbosity set_verbosity(2) async def run_fetch_mcp(): async with MCPOrchestra() as mcp_client: try: # Connect to the Fetch MCP server await mcp_client.connect( server_name="fetch", command="npx", args=["@tokenizin/mcp-npx-fetch"] ) # Get all tools from the MCP server fetch_tools = mcp_client.get_tools() # Define the agent agent = Agent( agent_id="fetch_agent", role="Web Assistant", goal="Help users fetch and analyze web content", tools=fetch_tools, llm=OpenaiModels.gpt_4o ) # Create a task with these tools result = await Task.create( agent=agent, instruction="Fetch the content from docs.orchestra.org and summarize what the website is about.", ) print(result) except Exception as e: print(f"Error occurred: {e}") if __name__ == "__main__": asyncio.run(run_fetch_mcp())
Filesystem ExampleCopied!
# mcp_test_filesystem.py import asyncio import os from mainframe_orchestra import Task, Agent, OpenaiModels, MCPOrchestra, set_verbosity set_verbosity(2) async def run_filesystem_mcp(): async with MCPOrchestra() as mcp_client: try: # Define directories to allow access to (customize these paths) allowed_directories = [ os.path.expanduser("~/Desktop"), # Allow access to Desktop # Add more directories as needed ] # Connect to the Filesystem MCP server await mcp_client.connect( server_name="filesystem", command="npx", args=["@modelcontextprotocol/server-filesystem"] + allowed_directories, # Note: Directories are passed as additional arguments for security ) # Get all tools from the MCP server filesystem_tools = mcp_client.get_tools() # Define the agent agent = Agent( agent_id="filesystem_agent", role="File System Assistant", goal="Help users manage files and directories", attributes="You know to use absolute paths when possible", tools=filesystem_tools, llm=OpenaiModels.gpt_4o ) # Create a task with these tools result = await Task.create( agent=agent, instruction="List the files on my desktop and summarize what you find.", ) print(result) except Exception as e: print(f"Error: {e}") if __name__ == "__main__": asyncio.run(run_filesystem_mcp())
Web Automation with PlaywrightCopied!
# mcp_test_playwright.py import asyncio from mainframe_orchestra import Task, Agent, OpenaiModels, MCPOrchestra, set_verbosity set_verbosity(2) async def run_agent_with_mcp_tools(): async with MCPOrchestra() as mcp_client: try: # Connect to the Playwright MCP server await mcp_client.connect( server_name="playwright", command="npx", args=["playwright-mcp-server"] ) # Get all tools from the MCP server playwright_tools = mcp_client.get_tools() # Define the agent agent = Agent( agent_id="playwright_agent", role="Web Automation Assistant", goal="Help users automate web tasks using Playwright", tools=playwright_tools, llm=OpenaiModels.gpt_4o ) # Create the task with step-by-step instructions result = await Task.create( agent=agent, instruction=""" Please perform the following web automation tasks: 1. Navigate to duckduckgo.com 2. Try to use the selector to find the search input 3. Use the identified selector to search for 'AI Agent news' 4. Click the search button or press Enter to submit the search 5. Find the first result and click it 6. Take a screenshot of the resulting page 7. Summarize what you found """ ) print(result) except Exception as e: print(f"Error occurred: {e}") if __name__ == "__main__": asyncio.run(run_agent_with_mcp_tools())
Slack Integration ExampleCopied!
# mcp_test_slack.py import asyncio import os import shutil from mainframe_orchestra import Task, Agent, OpenaiModels, MCPOrchestra, set_verbosity from dotenv import load_dotenv # Load environment variables from .env file load_dotenv() set_verbosity(2) async def run_slack_mcp(): async with MCPOrchestra() as mcp_client: try: # Get Slack credentials from environment variables slack_bot_token = os.environ.get("SLACK_BOT_TOKEN") slack_team_id = os.environ.get("SLACK_TEAM_ID") if not slack_bot_token or not slack_team_id: print("ERROR: Missing Slack credentials. Set SLACK_BOT_TOKEN and SLACK_TEAM_ID environment variables.") return # Get the path to node executable node_path = shutil.which("node") if not node_path: print("ERROR: Node.js not found. Please install Node.js.") return # Path to the Slack MCP server module # Note: You need to install this first with: # npm install @modelcontextprotocol/server-slack server_path = "node_modules/@modelcontextprotocol/server-slack/dist/index.js" # Connect to the Slack MCP server await mcp_client.connect( server_name="slack", command=node_path, args=[server_path], env={ "SLACK_BOT_TOKEN": slack_bot_token, "SLACK_TEAM_ID": slack_team_id } ) # Get Slack tools slack_tools = mcp_client.get_server_tools("slack") # Create the Slack agent slack_agent = Agent( agent_id="slack_agent", role="Slack Assistant", goal="Help users interact with Slack by using your tools", tools=slack_tools, llm=OpenaiModels.gpt_4o ) # Create a conversation history for context conversation_history = [] # Main conversation loop while True: # Get user input print("\nEnter your instruction (or 'quit' to exit):") user_instruction = input("You: ") if user_instruction.lower() == 'quit': break # Create a task with these tools result = await Task.create( agent=slack_agent, instruction=user_instruction, messages=conversation_history ) # Add the interaction to conversation history conversation_history.append({"role": "user", "content": user_instruction}) conversation_history.append({"role": "assistant", "content": str(result)}) print("\nTask result:") print(result) except Exception as e: print(f"Error occurred: {e}") if __name__ == "__main__": asyncio.run(run_slack_mcp())
Creating Your Own MCP Server with FastMCPCopied!
FastMCP is a Python library that makes it easy to create MCP-compatible servers. Here's how to create a simple server:
from fastmcp import FastMCP # Create a FastMCP server with a name mcp = FastMCP("MyTools") # Define a simple tool @mcp.tool() def my_tool(param1: str, param2: int) -> str: """ Description of what my_tool does Args: param1: Description of param1 param2: Description of param2 Returns: Description of the return value """ # Tool implementation return f"Processed {param1} {param2} times" # Define a resource endpoint @mcp.resource("data://{id}") def get_data(id: str) -> dict: """ Get data for a specific ID Args: id: The ID to get data for Returns: Data associated with the ID """ # Resource implementation return {"id": id, "value": f"Data for {id}"} # Run the server if __name__ == "__main__": mcp.run()
Advanced FeaturesCopied!
Connecting to Multiple ServersCopied!
You can connect to multiple MCP servers and combine their tools:
async with MCPOrchestra() as mcp_client: # Connect to a calculator server await mcp_client.connect( server_name="calculator", command="python", args=["mcp_fast_calc.py"], start_server=True ) # Connect to a web automation server await mcp_client.connect( server_name="playwright", command="npx", args=["playwright-mcp-server"] ) # Get all tools from all servers all_tools = mcp_client.get_tools() # Or get tools from specific servers calculator_tools = mcp_client.get_server_tools("calculator") playwright_tools = mcp_client.get_server_tools("playwright") # Create agents with different tool sets calculator_agent = Agent( agent_id="calculator_agent", role="Math Assistant", tools=calculator_tools, llm=OpenaiModels.gpt_4o ) web_agent = Agent( agent_id="web_agent", role="Web Assistant", tools=playwright_tools, llm=OpenaiModels.gpt_4o ) super_agent = Agent( agent_id="super_agent", role="Super Assistant", tools=all_tools, llm=OpenaiModels.gpt_4o )
Listing Available ToolsCopied!
You can list all available tools from connected MCP servers:
# List tools from all servers tools_info = await mcp_client.list_tools(verbose=True) print(tools_info) # List tools from a specific server calculator_tools_info = await mcp_client.list_tools(server_name="calculator") print(calculator_tools_info)
Installation RequirementsCopied!
Before using MCP tools, make sure you have the server package installed and set to the correct path.
Best Practices when building your own MCP server / tool libraryCopied!
-
Use Descriptive Tool Names and Schemas: The adapter converts tool descriptions and schemas into docstrings that the LLM uses to understand how to use the tools.
-
Handle Errors Gracefully: Implement proper error handling in your MCP tools to provide helpful error messages that can guide the LLM to correct its usage.
-
Use Environment Variables for Sensitive Information: When connecting to MCP servers that require API keys or other sensitive information, pass them as environment variables.
-
Close Connections Properly: Use the
async with
MCPOrchestra
-
Consider Tool Granularity: Create focused, single-purpose tools rather than complex multi-purpose ones for better usability by the LLM.
-
Security Considerations: When using filesystem or other sensitive tools, always restrict access to only the necessary directories or resources.
TroubleshootingCopied!
Common IssuesCopied!
-
Server Not Found: Make sure the command and args correctly point to the MCP server executable.
Error: Cannot find module '@tokenizin/mcp-npx-fetch'
Solution: Install the package with
npm install @tokenizin/mcp-npx-fetch
-
Connection Errors: If you see connection errors, check that:
- The server is running
- You've provided the correct environment variables
- You've waited long enough for the server to start (adjust
server_startup_delay
-
Tool Not Found: If the agent can't find a tool, check that:
- The tool is properly defined in the MCP server
- You're using
get_tools()
get_server_tools(server_name)
Tool Access MethodsCopied!
When working with multiple MCP servers, Orchestra provides two methods to access tools:
get_tools()
vs
get_server_tools(server_name)
Copied!
get_tools()
get_server_tools(server_name)
-
get_tools()
# Get all tools from all connected servers all_tools = mcp_client.get_tools() # Create an agent with access to all tools super_agent = Agent( agent_id="super_agent", role="Super Assistant", tools=all_tools, llm=OpenaiModels.gpt_4o )
-
get_server_tools(server_name)
# Get only tools from the "calculator" server calculator_tools = mcp_client.get_server_tools("calculator") # Create a specialized agent with only calculator tools calculator_agent = Agent( agent_id="calculator_agent", role="Math Assistant", tools=calculator_tools, llm=OpenaiModels.gpt_4o )
Choose
get_tools()
get_server_tools()
OverallCopied!
MCP integration enables Orchestra to work with tools implemented in any programming language, significantly expanding its capabilities. Whether you need to integrate with existing services, leverage specialized libraries, or create custom tools in your preferred language, the MCP adapter provides a standardized way to make these tools available to your Orchestra agents.