Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Connect your AI agent to 500+ apps for discovering tools, managing connections, and executing actions across Gmail, Slack, GitHub, Notion, Google Workspace,...
Connect your AI agent to 500+ apps for discovering tools, managing connections, and executing actions across Gmail, Slack, GitHub, Notion, Google Workspace,...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Composio connects your AI agent to 500+ apps โ Gmail, Slack, GitHub, Notion, Google Workspace, Microsoft (Outlook, Teams), X/Twitter, Figma, Web Search, Browser automation, Meta apps, TikTok, and more โ for seamless cross-app automation. Use this guide to discover tools, manage connections, and execute actions across any supported app. API Base: https://backend.composio.dev/api/v3 Auth: All requests require x-api-key: YOUR_API_KEY
Go to platform.composio.dev and sign in with Google or another account. Navigate to your project settings and copy your API key. export COMPOSIO_API_KEY="your_api_key_here" export COMPOSIO_BASE="https://backend.composio.dev/api/v3" That's it. You can now call any Composio tool.
Every tool is called through one endpoint โ the tool slug goes in the URL: POST /api/v3/tools/execute/{TOOL_SLUG} curl -X POST "$COMPOSIO_BASE/tools/execute/TOOL_SLUG_HERE" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { ... } }' Optional fields on every request: FieldDescriptionargumentsTool-specific input parameters (see each tool below)user_idYour end-user's identifier (for multi-user connected accounts)connected_account_idOverride which connected account to use
Always call this first. Searches 500+ apps to find the right tools for your task. Returns tool schemas, execution plans, connection status, and pitfalls. curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_SEARCH_TOOLS" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "queries": [ { "use_case": "send an email to someone", "known_fields": "recipient_name: John" }, { "use_case": "create a meeting invite", "known_fields": "meeting_date: tomorrow" } ], "session": { "generate_id": true } } }' ParameterRequiredDescriptionqueriesYesArray of search queries (1-7 items)queries[].use_caseYesNormalized English description of the task. Include app name if known. No personal identifiers here.queries[].known_fieldsNoComma-separated key:value pairs of known inputs (e.g., "channel_name: general, timezone: UTC")session.generate_idYes (first call)Set true for new workflows to get a session IDsession.idYes (subsequent)Reuse the session ID returned from first callmodelNoYour LLM model name (e.g., "claude-4.5-sonnet") for optimized planning Response: { "data": { "results": [ { "index": 1, "use_case": "send an email to someone", "primary_tool_slugs": ["GMAIL_SEND_EMAIL"], "related_tool_slugs": ["GMAIL_CREATE_EMAIL_DRAFT"], "toolkits": ["gmail"], "recommended_plan_steps": ["Step 1: ...", "Step 2: ..."], "known_pitfalls": ["Always set user_id to 'me'"], "reference_workbench_snippets": [...] } ], "tool_schemas": { "GMAIL_SEND_EMAIL": { "toolkit": "gmail", "tool_slug": "GMAIL_SEND_EMAIL", "description": "Send an email", "input_schema": { ... } } }, "toolkit_connection_statuses": [ { "toolkit": "gmail", "has_active_connection": false, "status_message": "No active connection. Initiate via COMPOSIO_MANAGE_CONNECTIONS." } ], "time_info": { "current_time_utc": "2025-01-15T10:30:00Z", "current_time_utc_epoch_seconds": 1736935800 }, "session": { "id": "abcd", "instructions": "Pass this session ID in all subsequent calls" } }, "successful": true } Splitting guidelines: 1 query = 1 tool action. Include hidden prerequisites (e.g., "get issue" before "update issue"). Include app names in every sub-query if the user specified one. Translate non-English prompts to English while preserving identifiers. After searching: Review the recommended_plan_steps and known_pitfalls before executing. If a tool has schemaRef instead of input_schema, call COMPOSIO_GET_TOOL_SCHEMAS first. If has_active_connection is false for a toolkit, call COMPOSIO_MANAGE_CONNECTIONS before executing its tools.
Creates or checks OAuth/API connections for toolkits. Returns auth links for user authentication. curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_MANAGE_CONNECTIONS" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "toolkits": ["gmail", "slack"], "session_id": "abcd" } }' ParameterRequiredDescriptiontoolkitsYesToolkit names from SEARCH_TOOLS response (e.g., ["gmail", "github"]). Never invent names.reinitiate_allNoForce reconnection even if already active (default: false)session_idNoSession ID from SEARCH_TOOLS Response: { "data": { "message": "1 active, 1 initiated", "results": { "gmail": { "toolkit": "gmail", "status": "initiated", "redirect_url": "https://accounts.google.com/o/oauth2/...", "instruction": "Click the link to authenticate" }, "slack": { "toolkit": "slack", "status": "active", "has_active_connection": true, "connected_account_id": "ca_xxx", "current_user_info": { "email": "user@example.com" } } }, "summary": { "total_toolkits": 2, "active_connections": 1, "initiated_connections": 1, "failed_connections": 0 } }, "successful": true } Connection workflow: Call SEARCH_TOOLS โ check toolkit_connection_statuses If has_active_connection is false, call MANAGE_CONNECTIONS If status is "initiated", present the redirect_url to the user for authentication Once the user completes auth, the connection becomes active Only execute tools after the connection is confirmed active
Retrieves complete input schemas for tools. Call this when SEARCH_TOOLS returns schemaRef instead of a full input_schema. curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_GET_TOOL_SCHEMAS" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "tool_slugs": ["GMAIL_SEND_EMAIL", "SLACK_SEND_MESSAGE"], "session_id": "abcd" } }' ParameterRequiredDescriptiontool_slugsYesArray of tool slugs from SEARCH_TOOLSsession_idNoSession ID Response: { "data": { "success": true, "tool_schemas": { "GMAIL_SEND_EMAIL": { "toolkit": "gmail", "tool_slug": "GMAIL_SEND_EMAIL", "description": "Send an email via Gmail", "input_schema": { "properties": { "to": { "type": "string", "description": "Recipient email" }, "subject": { "type": "string" }, "body": { "type": "string" } }, "required": ["to", "subject", "body"] } } }, "not_found": [] }, "successful": true }
Executes one or more tools in parallel. This is how you run the tools discovered through SEARCH_TOOLS. curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_MULTI_EXECUTE_TOOL" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "tools": [ { "tool_slug": "GMAIL_SEND_EMAIL", "arguments": { "to": "john@example.com", "subject": "Hello", "body": "Welcome to the team!" } }, { "tool_slug": "SLACK_SEND_MESSAGE", "arguments": { "channel": "#general", "text": "New member joined!" } } ], "sync_response_to_workbench": false, "session_id": "abcd" } }' ParameterRequiredDescriptiontoolsYesArray of tool executions (1-50 items)tools[].tool_slugYesValid slug from SEARCH_TOOLStools[].argumentsYesArguments matching the tool's input schema exactlysync_response_to_workbenchYesSet true if response may be large or needed for later scriptingsession_idNoSession ID Response: { "data": { "results": [ { "tool_slug": "GMAIL_SEND_EMAIL", "index": 0, "response": { "data": { "id": "msg_123", "threadId": "thread_456" }, "successful": true }, "error": null }, { "tool_slug": "SLACK_SEND_MESSAGE", "index": 1, "response": { "data": { "ok": true, "ts": "1234567890.123456" }, "successful": true }, "error": null } ], "total_count": 2, "success_count": 2, "error_count": 0, "remote_file_info": null }, "successful": true } Rules: Only batch tools that are logically independent (no ordering dependencies). Never invent tool slugs or argument fields โ always use what SEARCH_TOOLS returned. Ensure active connections exist before executing. If response is large, it may be saved to a remote file โ use remote_file_info.file_path with the Remote Workbench or Bash tools to process it.
Executes Python code in a persistent remote Jupyter sandbox. Use for processing large tool outputs, bulk operations, and scripting multi-tool chains. curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_REMOTE_WORKBENCH" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "code_to_execute": "result, error = run_composio_tool(\"GMAIL_FETCH_EMAILS\", {\"max_results\": 5, \"user_id\": \"me\"})\nif error:\n print(\"Error:\", error)\nelse:\n emails = result.get(\"data\", {})\n print(\"Fetched:\", len(emails.get(\"messages\", [])))", "thought": "Fetching recent emails for bulk processing", "session_id": "abcd" } }' ParameterRequiredDescriptioncode_to_executeYesPython code to run in the sandboxthoughtNoOne-sentence objective descriptiontimeoutNoExecution timeout in seconds (1-780, default: 600)session_idNoSession ID Response: { "data": { "results": "Fetched: 5", "stdout": "Fetched: 5\n", "stderr": "", "error": "", "sandbox_id_suffix": "a1b2" }, "successful": true } If output exceeds 40,000 characters, it is saved to a file and results_file_path / stdout_file_path is returned instead. Pre-loaded helper functions (do NOT import or redeclare these): FunctionSignatureDescriptionrun_composio_tool(tool_slug: str, arguments: dict) -> tuple[dict, str]Execute a Composio app tool. Returns (response, error).invoke_llm(query: str) -> tuple[str, str]Call an LLM for analysis, summarization, extraction. Max 200k chars input. Returns (response, error).proxy_execute(method, endpoint, toolkit, query_params?, body?, headers?) -> tuple[any, str]Direct API call to a connected toolkit when no Composio tool exists. Returns (response, error).web_search(query: str) -> tuple[str, str]Search the web via Exa AI. Returns (results, error).upload_local_file(*file_paths) -> tuple[dict, str]Upload sandbox files to cloud storage. Returns ({"s3_url": ...}, error).smart_file_extract(sandbox_file_path: str) -> tuple[str, str]Extract text from PDF, images, etc. Returns (text, error). All helpers return (result, error) โ always check error before using result. Coding rules: Split work into small steps; save intermediate results to /tmp/ files. State persists across executions (variables, imports, files). Hard timeout of 4 minutes โ use ThreadPoolExecutor for bulk operations. Always check error from helper functions before using results. Use invoke_llm for summarization and analysis โ it gives better results than ad-hoc filtering. Do NOT call COMPOSIO_* meta tools via run_composio_tool โ only use it for app tools (e.g., GMAIL_SEND_EMAIL). Example โ bulk email processing: import concurrent.futures def process_email(email_id): result, error = run_composio_tool("GMAIL_GET_EMAIL", { "message_id": email_id, "user_id": "me" }) if error: return {"id": email_id, "error": error} subject = result.get("data", {}).get("subject", "") return {"id": email_id, "subject": subject} # Fetch email list emails, err = run_composio_tool("GMAIL_FETCH_EMAILS", { "max_results": 50, "user_id": "me" }) if not err: ids = [m["id"] for m in emails.get("data", {}).get("messages", [])] with concurrent.futures.ThreadPoolExecutor(max_workers=10) as ex: results = list(ex.map(process_email, ids)) print(f"Processed {len(results)} emails")
Executes bash commands in the same persistent sandbox as the workbench. curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_REMOTE_BASH_TOOL" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "command": "cat /home/user/.code_out/response.json | jq \".results[] | .tool_slug\"", "session_id": "abcd" } }' ParameterRequiredDescriptioncommandYesBash command to executesession_idNoSession ID Response: { "data": { "stdout": "\"GMAIL_SEND_EMAIL\"\n\"SLACK_SEND_MESSAGE\"\n", "stdoutLines": 2, "stderr": "", "stderrLines": 0, "sandbox_id_suffix": "a1b2" }, "successful": true } Use cases: Process large tool responses saved to remote files (via jq, awk, sed, grep) File system operations in the sandbox Commands run from /home/user by default 5-minute timeout, max 40,000 chars output per stream
1. COMPOSIO_SEARCH_TOOLS | Find tools for your task | Check connection statuses | Review execution plan and pitfalls 2. COMPOSIO_GET_TOOL_SCHEMAS (if needed) | Get full input schemas for tools with schemaRef 3. COMPOSIO_MANAGE_CONNECTIONS (if needed) | Initiate connections for toolkits without active connections | Present auth URL to user -> user completes OAuth 4. COMPOSIO_MULTI_EXECUTE_TOOL | Execute tools with schema-compliant arguments | Process inline results or note remote file paths 5. COMPOSIO_REMOTE_WORKBENCH / COMPOSIO_REMOTE_BASH_TOOL (if needed) | Process large outputs saved to remote files | Run bulk operations or multi-tool scripts | Upload artifacts via upload_local_file Example โ complete "send email" workflow: # Step 1: Search for email tools curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_SEARCH_TOOLS" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "queries": [{"use_case": "send an email via gmail", "known_fields": "recipient: john@example.com"}], "session": {"generate_id": true} } }' # -> Returns GMAIL_SEND_EMAIL tool with schema # -> Check toolkit_connection_statuses for gmail # -> Save session.id from response (e.g., "abcd") # Step 2: Connect gmail (if not active) curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_MANAGE_CONNECTIONS" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "toolkits": ["gmail"], "session_id": "abcd" } }' # -> If status is "initiated", present redirect_url to user for OAuth # Step 3: Execute the tool curl -X POST "$COMPOSIO_BASE/tools/execute/COMPOSIO_MULTI_EXECUTE_TOOL" \ -H "x-api-key: $COMPOSIO_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "arguments": { "tools": [{ "tool_slug": "GMAIL_SEND_EMAIL", "arguments": { "to": "john@example.com", "subject": "Welcome!", "body": "Welcome to the team, John!", "user_id": "me" } }], "sync_response_to_workbench": false, "session_id": "abcd" } }'
Pass session: {generate_id: true} on your first SEARCH_TOOLS call to get a session ID. Pass that session_id into all subsequent tool calls within the same workflow. Generate a new session when the user pivots to a different task.
Always call SEARCH_TOOLS first โ never guess tool slugs or argument fields. Re-run SEARCH_TOOLS when you need additional tools due to errors or changed requirements. Review recommended_plan_steps and known_pitfalls before executing.
Never execute a toolkit tool without an active connection. Use exact toolkit names from SEARCH_TOOLS โ never invent names. If a connection fails, present the auth URL to the user.
Use schema-compliant arguments only โ check input_schema before every call. Batch independent tools into a single MULTI_EXECUTE call. For large responses, set sync_response_to_workbench: true and process in the workbench.
Only use the workbench for remote file processing or bulk scripting โ not for data already visible inline. Keep code concise; split long operations into multiple workbench calls. Use ThreadPoolExecutor for parallelism within the 4-minute timeout.
ActionTool SlugKey ArgumentsDiscover toolsCOMPOSIO_SEARCH_TOOLSqueries, sessionGet input schemasCOMPOSIO_GET_TOOL_SCHEMAStool_slugsConnect to appsCOMPOSIO_MANAGE_CONNECTIONStoolkits, reinitiate_allExecute toolsCOMPOSIO_MULTI_EXECUTE_TOOLtools, sync_response_to_workbenchRun Python codeCOMPOSIO_REMOTE_WORKBENCHcode_to_execute, thoughtRun bash commandsCOMPOSIO_REMOTE_BASH_TOOLcommand Endpoint for all tools: POST /api/v3/tools/execute/{TOOL_SLUG} Common errors: ErrorMeaning401Invalid or missing API key403Forbidden โ insufficient permissions404Tool or resource not found422Invalid arguments (check input schema)429Rate limited โ back off and retry500Internal server error Rate limits: Respect 429 responses with exponential backoff.
Gotchas and non-obvious behavior discovered during testing. Some tools return schemaRef instead of input_schema. When you see "hasFullSchema": false, you MUST call COMPOSIO_GET_TOOL_SCHEMAS before executing that tool. Response nesting: Tool execution results live at response.data.results[].response.data โ two levels of data. Parse carefully. Workbench state persists across calls. Variables, imports, and files in /tmp/ survive between executions within the same sandbox session. String formatting in workbench code: Avoid f-strings with nested quotes when sending code via JSON. Use string concatenation instead to prevent quoting issues. invoke_llm works well for classification and analysis of tool output. Request explicit JSON format for structured results.
Workflow acceleration for inboxes, docs, calendars, planning, and execution loops.
Largest current source with strong distribution and engagement signals.