Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Moderate text, images, and video using Vettly's content moderation API via MCP server.
Moderate text, images, and video using Vettly's content moderation API via MCP server.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Moderate user-generated content using Vettly's AI-powered content moderation API. This skill uses the @vettly/mcp MCP server to check text, images, and video against configurable moderation policies with auditable decisions.
Add the @vettly/mcp MCP server to your configuration: { "mcpServers": { "vettly": { "command": "npx", "args": ["-y", "@vettly/mcp"], "env": { "VETTLY_API_KEY": "your-api-key" } } } } Get an API key at vettly.dev.
Check text, image, or video content against a Vettly moderation policy. Returns a safety assessment with category scores, the action taken, provider used, latency, and cost. Parameters: content (required) - The content to moderate (text string, or URL for images/video) policyId (required) - The policy ID to use for moderation contentType (optional, default: text) - Type of content: text, image, or video
Validate a Vettly policy YAML without saving it. Returns validation results with any syntax or configuration errors. Use this to test policy changes before deploying them. Parameters: yamlContent (required) - The YAML policy content to validate
List all moderation policies available in your Vettly account. Takes no parameters. Use this to discover available policy IDs before moderating content.
Get usage statistics for your Vettly account including request counts, costs, and moderation outcomes. Parameters: days (optional, default: 30) - Number of days to include in statistics (1-365)
Get recent moderation decisions with optional filtering by outcome, content type, or policy. Parameters: limit (optional, default: 10) - Number of decisions to return (1-50) flagged (optional) - Filter to only flagged content (true) or safe content (false) policyId (optional) - Filter by specific policy ID contentType (optional) - Filter by content type: text, image, or video
Moderate user-generated content (comments, posts, uploads) before publishing Test and validate moderation policy YAML configs during development Audit recent moderation decisions to review flagged content Monitor moderation costs and usage across your account Compare moderation results across different policies
Moderate this user comment for my community forum policy: "I hate this product, it's the worst thing I've ever used and the developers should be ashamed" Call list_policies to find available policies, then moderate_content with the appropriate policy ID and return the safety assessment.
Validate this moderation policy YAML: categories: - name: toxicity threshold: 0.8 action: flag - name: spam threshold: 0.6 action: block Call validate_policy and report any syntax or configuration errors.
Show me all flagged content from the last week Call get_recent_decisions with flagged: true to retrieve recent moderation decisions that were flagged.
Always call list_policies first if you don't know which policy ID to use Use validate_policy to test policy changes before deploying to production Use get_usage_stats to monitor costs and catch unexpected spikes Filter get_recent_decisions by contentType or policyId to narrow results For image and video moderation, pass the content URL rather than raw data
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.