โ† All skills
Tencent SkillHub ยท Developer Tools

Content Moderation

Moderate text, images, and video using Vettly's content moderation API via MCP server.

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

Moderate text, images, and video using Vettly's content moderation API via MCP server.

โฌ‡ 0 downloads โ˜… 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
SKILL.md

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
1.0.0

Documentation

ClawHub primary doc Primary doc: SKILL.md 12 sections Open source page

Content Moderation

Moderate user-generated content using Vettly's AI-powered content moderation API. This skill uses the @vettly/mcp MCP server to check text, images, and video against configurable moderation policies with auditable decisions.

Setup

Add the @vettly/mcp MCP server to your configuration: { "mcpServers": { "vettly": { "command": "npx", "args": ["-y", "@vettly/mcp"], "env": { "VETTLY_API_KEY": "your-api-key" } } } } Get an API key at vettly.dev.

moderate_content

Check text, image, or video content against a Vettly moderation policy. Returns a safety assessment with category scores, the action taken, provider used, latency, and cost. Parameters: content (required) - The content to moderate (text string, or URL for images/video) policyId (required) - The policy ID to use for moderation contentType (optional, default: text) - Type of content: text, image, or video

validate_policy

Validate a Vettly policy YAML without saving it. Returns validation results with any syntax or configuration errors. Use this to test policy changes before deploying them. Parameters: yamlContent (required) - The YAML policy content to validate

list_policies

List all moderation policies available in your Vettly account. Takes no parameters. Use this to discover available policy IDs before moderating content.

get_usage_stats

Get usage statistics for your Vettly account including request counts, costs, and moderation outcomes. Parameters: days (optional, default: 30) - Number of days to include in statistics (1-365)

get_recent_decisions

Get recent moderation decisions with optional filtering by outcome, content type, or policy. Parameters: limit (optional, default: 10) - Number of decisions to return (1-50) flagged (optional) - Filter to only flagged content (true) or safe content (false) policyId (optional) - Filter by specific policy ID contentType (optional) - Filter by content type: text, image, or video

When to Use

Moderate user-generated content (comments, posts, uploads) before publishing Test and validate moderation policy YAML configs during development Audit recent moderation decisions to review flagged content Monitor moderation costs and usage across your account Compare moderation results across different policies

Moderate a user comment

Moderate this user comment for my community forum policy: "I hate this product, it's the worst thing I've ever used and the developers should be ashamed" Call list_policies to find available policies, then moderate_content with the appropriate policy ID and return the safety assessment.

Validate a policy before deploying

Validate this moderation policy YAML: categories: - name: toxicity threshold: 0.8 action: flag - name: spam threshold: 0.6 action: block Call validate_policy and report any syntax or configuration errors.

Review recent flagged content

Show me all flagged content from the last week Call get_recent_decisions with flagged: true to retrieve recent moderation decisions that were flagged.

Tips

Always call list_policies first if you don't know which policy ID to use Use validate_policy to test policy changes before deploying to production Use get_usage_stats to monitor costs and catch unexpected spikes Filter get_recent_decisions by contentType or policyId to narrow results For image and video moderation, pass the content URL rather than raw data

Category context

Code helpers, APIs, CLIs, browser automation, testing, and developer operations.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
1 Docs
  • SKILL.md Primary doc