← All skills
Tencent SkillHub · AI

Visual Gen Ai Language

Maximum control over AI image generation — write structured VGL (Visual Generation Language) JSON that explicitly controls every visual attribute. Define exa...

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

Maximum control over AI image generation — write structured VGL (Visual Generation Language) JSON that explicitly controls every visual attribute. Define exa...

⬇ 0 downloads ★ 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
SKILL.md, references/schema-reference.md

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
1.2.1

Documentation

ClawHub primary doc Primary doc: SKILL.md 23 sections Open source page

Bria VGL — Full Control Over Image Generation

Define every visual attribute as structured JSON instead of hoping natural language gets it right. VGL (Visual Generation Language) gives you explicit, deterministic control over objects, lighting, camera settings, composition, and style for Bria's FIBO models. Related Skill: Use bria-ai to execute these VGL prompts via the Bria API. VGL defines the structured control format; bria-ai handles generation, editing, and background removal.

Core Concept

VGL replaces ambiguous natural language prompts with deterministic JSON that explicitly declares every visual attribute: objects, lighting, camera settings, composition, and style. This ensures reproducible, controllable image generation.

Operation Modes

ModeInputOutputUse CaseGenerateText promptVGL JSONCreate new image from descriptionEditImage + instructionVGL JSONModify reference imageEdit_with_MaskMasked image + instructionVGL JSONFill grey masked regionsCaptionImage onlyVGL JSONDescribe existing imageRefineExisting JSON + editUpdated VGL JSONModify existing prompt

JSON Schema

Output a single valid JSON object with these required keys:

1. short_description (String)

Concise summary of image content, max 200 words. Include key subjects, actions, setting, and mood.

2. objects (Array, max 5 items)

Each object requires: { "description": "Detailed description, max 100 words", "location": "center | top-left | bottom-right foreground | etc.", "relative_size": "small | medium | large within frame", "shape_and_color": "Basic shape and dominant color", "texture": "smooth | rough | metallic | furry | fabric | etc.", "appearance_details": "Notable visual details", "relationship": "Relationship to other objects", "orientation": "upright | tilted 45 degrees | facing left | horizontal | etc." } Human subjects add: { "pose": "Body position description", "expression": "winking | joyful | serious | surprised | calm", "clothing": "Attire description", "action": "What the person is doing", "gender": "Gender description", "skin_tone_and_texture": "Skin appearance" } Object clusters add: { "number_of_objects": 3 } Size guidance: If a person is the main subject, use "medium-to-large" or "large within frame".

3. background_setting (String)

Overall environment, setting, and background elements not in objects.

4. lighting (Object)

{ "conditions": "bright daylight | dim indoor | studio lighting | golden hour | blue hour | overcast", "direction": "front-lit | backlit | side-lit from left | top-down", "shadows": "long, soft shadows | sharp, defined shadows | minimal shadows" }

5. aesthetics (Object)

{ "composition": "rule of thirds | symmetrical | centered | leading lines | medium shot | close-up", "color_scheme": "monochromatic blue | warm complementary | high contrast | pastel", "mood_atmosphere": "serene | energetic | mysterious | joyful | dramatic | peaceful" } For people as main subject, specify shot type in composition: "medium shot", "close-up", "portrait composition".

6. photographic_characteristics (Object)

{ "depth_of_field": "shallow | deep | bokeh background", "focus": "sharp focus on subject | soft focus | motion blur", "camera_angle": "eye-level | low angle | high angle | dutch angle | bird's-eye", "lens_focal_length": "wide-angle | 50mm standard | 85mm portrait | telephoto | macro" } For people: Prefer "standard lens (35mm-50mm)" or "portrait lens (50mm-85mm)". Avoid wide-angle unless specified.

7. style_medium (String)

"photograph" | "oil painting" | "watercolor" | "3D render" | "digital illustration" | "pencil sketch" Default to "photograph" unless explicitly requested otherwise.

8. artistic_style (String)

If not photograph, describe characteristics in max 3 words: "impressionistic, vibrant, textured" For photographs, use "realistic" or similar.

9. context (String)

Describe the image type/purpose: "High-fashion editorial photograph for magazine spread" "Concept art for fantasy video game" "Commercial product photography for e-commerce"

10. text_render (Array)

Default: empty array [] Only populate if user explicitly provides exact text content: { "text": "Exact text from user (never placeholder)", "location": "center | top-left | bottom", "size": "small | medium | large", "color": "white | red | blue", "font": "serif typeface | sans-serif | handwritten | bold impact", "appearance_details": "Metallic finish | 3D effect | etc." } Exception: Universal text integral to objects (e.g., "STOP" on stop sign).

11. edit_instruction (String)

Single imperative command describing the edit/generation.

For Standard Edits (no mask)

Start with action verb, describe changes, never reference "original image": CategoryRewritten InstructionStyle changeTurn the image into the cartoon style.Object attributeChange the dog's color to black and white.Add elementAdd a wide-brimmed felt hat to the subject.Remove objectRemove the book from the subject's hands.Replace objectChange the rose to a bright yellow sunflower.LightingChange the lighting from dark and moody to bright and vibrant.CompositionChange the perspective to a wider shot.Text changeChange the text "Happy Anniversary" to "Hello".QualityRefine the image to obtain increased clarity and sharpness.

For Masked Region Edits

Reference "masked regions" or "masked area" as target: IntentRewritten InstructionObject generationGenerate a white rose with a blue center in the masked region.ExtensionExtend the image into the masked region to create a scene featuring...Background fillCreate the following background in the masked region: A vast ocean extending to horizon.Atmospheric fillFill the background masked area with a clear, bright blue sky with wispy clouds.Subject restorationRestore the area in the mask with a young woman.Environment infillCreate inside the masked area: a greenhouse with rows of plants under glass ceiling.

Standard Edit Mode

Preserve ALL visual properties unless explicitly changed by instruction: Subject identity, pose, appearance Object existence, location, size, orientation Composition, camera angle, lens characteristics Style/medium Only change what the edit strictly requires.

Masked Edit Mode

Preserve all visible (non-masked) portions exactly Fill grey masked regions to blend seamlessly with unmasked areas Match existing style, lighting, and subject matter Never describe grey masks—describe content that fills them

Example Output

{ "short_description": "A professional businesswoman in a navy blazer stands confidently in a modern glass office, holding a tablet. Natural daylight streams through floor-to-ceiling windows, creating a warm, productive atmosphere.", "objects": [ { "description": "A confident businesswoman in her 30s with shoulder-length dark hair, wearing a tailored navy blazer over a white blouse. She holds a tablet in her left hand while gesturing naturally with her right.", "location": "center-right", "relative_size": "large within frame", "shape_and_color": "Human figure, navy and white clothing", "texture": "smooth fabric, professional attire", "appearance_details": "Minimal jewelry, well-groomed professional appearance", "relationship": "Main subject, interacting with tablet", "orientation": "facing slightly left, three-quarter view", "pose": "Standing upright, relaxed professional stance", "expression": "confident, approachable smile", "clothing": "Tailored navy blazer, white silk blouse, dark trousers", "action": "Presenting or reviewing information on tablet", "gender": "female", "skin_tone_and_texture": "Medium warm skin tone, healthy smooth complexion" }, { "description": "A modern tablet device with a bright display showing charts and graphs", "location": "center, held by subject", "relative_size": "small", "shape_and_color": "Rectangular, silver frame with illuminated screen", "texture": "smooth glass and metal", "appearance_details": "Thin profile, business application visible on screen", "relationship": "Held by businesswoman, focus of her attention", "orientation": "vertical, screen facing viewer at slight angle", "pose": null, "expression": null, "clothing": null, "action": null, "gender": null, "skin_tone_and_texture": null, "number_of_objects": null } ], "background_setting": "Modern corporate office interior with floor-to-ceiling windows overlooking a city skyline. Minimalist furniture in neutral tones, potted plants adding touches of green.", "lighting": { "conditions": "bright natural daylight", "direction": "side-lit from left through windows", "shadows": "soft, natural shadows" }, "aesthetics": { "composition": "rule of thirds, medium shot", "color_scheme": "professional blues and neutral whites with warm accents", "mood_atmosphere": "confident, professional, welcoming" }, "photographic_characteristics": { "depth_of_field": "shallow, background slightly soft", "focus": "sharp focus on subject's face and upper body", "camera_angle": "eye-level", "lens_focal_length": "portrait lens (85mm)" }, "style_medium": "photograph", "artistic_style": "realistic", "context": "Corporate portrait photography for company website or LinkedIn professional profile.", "text_render": [], "edit_instruction": "Generate a professional businesswoman in a modern office environment holding a tablet." }

Common Pitfalls

Don't invent text - Keep text_render empty unless user provides exact text Don't over-describe - Max 5 objects, prioritize most important Match the mode - Use correct edit_instruction format for masked vs standard edits Preserve fidelity - Only change what's explicitly requested Be specific - Use concrete values ("85mm portrait lens") not vague terms ("nice camera") Null for irrelevant - Human-specific fields should be null for non-human objects

curl Example

curl -X POST "https://engine.prod.bria-api.com/v2/image/generate" \ -H "api_token: $BRIA_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "structured_prompt": "{\"short_description\": \"...\", ...}", "prompt": "Generate this scene", "aspect_ratio": "16:9" }'

References

Schema Reference - Complete JSON schema with all parameter values bria-ai - API client and endpoint documentation for executing VGL prompts

Category context

Agent frameworks, memory systems, reasoning layers, and model-native orchestration.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
2 Docs
  • SKILL.md Primary doc
  • references/schema-reference.md Docs