Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Generate publication-quality academic diagrams, methodology figures, architecture illustrations, and statistical plots from text descriptions using the Paper...
Generate publication-quality academic diagrams, methodology figures, architecture illustrations, and statistical plots from text descriptions using the Paper...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
Generate publication-quality academic diagrams and statistical plots from text descriptions. Uses a multi-agent pipeline (Retriever β Planner β Stylist β Visualizer β Critic) with iterative refinement.
uv run {baseDir}/scripts/generate.py \ --context "Our framework consists of an encoder module that processes..." \ --caption "Overview of the proposed encoder-decoder architecture" Or from a file: uv run {baseDir}/scripts/generate.py \ --input /path/to/method_section.txt \ --caption "Overview of the proposed method" Options: --iterations N β refinement rounds (default: 3) --auto-refine β loop until critic is satisfied (use for final quality) --aspect RATIO β aspect ratio: 1:1, 2:3, 3:2, 3:4, 4:3, 9:16, 16:9, 21:9 --provider gemini|openai|openrouter β override auto-detected provider --format png|jpeg|webp β output format (default: png) --no-optimize β disable input optimization (on by default)
uv run {baseDir}/scripts/plot.py \ --data '{"model":["GPT-4","Claude","Gemini"],"accuracy":[92.1,94.3,91.8]}' \ --intent "Bar chart comparing model accuracy across benchmarks" Or from a CSV file: uv run {baseDir}/scripts/plot.py \ --data-file /path/to/results.csv \ --intent "Line plot showing training loss over epochs"
uv run {baseDir}/scripts/evaluate.py \ --generated /path/to/generated.png \ --reference /path/to/human_drawn.png \ --context "The methodology section text..." \ --caption "Overview of the framework" Returns scores on: Faithfulness, Readability, Conciseness, Aesthetics.
uv run {baseDir}/scripts/generate.py \ --continue \ --feedback "Make the arrows thicker and use more distinct colors" Or continue a specific run: uv run {baseDir}/scripts/generate.py \ --continue-run run_20260228_143022_a1b2c3 \ --feedback "Add labels to each component box"
The skill auto-installs paperbanana on first use via uv (isolated, no global install). The package is published on PyPI by the llmsresearch team. Required API keys: This skill requires at least one of the following API keys to function. Configure in ~/.openclaw/openclaw.json: Env VariableProviderCostNotesGOOGLE_API_KEYGoogle GeminiFree tier availableRecommended starting pointOPENAI_API_KEYOpenAIPaidBest quality (gpt-5.2 + gpt-image-1.5)OPENROUTER_API_KEYOpenRouterPaidAccess to any model { skills: { entries: { "paperbanana": { env: { // Option A: Google Gemini (free tier β recommended) GOOGLE_API_KEY: "AIza...", // Option B: OpenAI (paid, best quality) // OPENAI_API_KEY: "sk-...", // Option C: OpenRouter (paid, access to any model) // OPENROUTER_API_KEY: "sk-or-...", } } } } } Auto-detection priority: Gemini (free) β OpenAI β OpenRouter. The skill will exit with a clear error if no API key is found.
For provider comparison, model options, and advanced configuration: see {baseDir}/references/providers.md
This skill sends user-provided data to external third-party APIs for diagram generation and evaluation: Text content (context descriptions, captions, feedback) is sent to the configured LLM provider (Gemini, OpenAI, or OpenRouter) for planning and code generation. Generated images may be sent back to the LLM provider for VLM-based evaluation and refinement. CSV/JSON data provided for plot generation is sent to the LLM provider for Matplotlib code generation. Do not use this skill with sensitive, confidential, or proprietary data unless your organization's data policies permit sending that data to the configured provider. All API calls go directly to the provider's endpoints β no intermediate servers are involved. API keys are injected by OpenClaw from your local config (~/.openclaw/openclaw.json) and are never logged or transmitted beyond the provider's API.
PyPI package: paperbanana (β₯0.1.2, installed automatically via uv) Source: llmsresearch/paperbanana on GitHub Skill source: GoatInAHat/openclaw-paperbanana on GitHub Transitive deps: google-genai, openai, matplotlib, Pillow, and others (installed in an isolated uv environment, not globally)
Input optimization is ON by default β enriches context and sharpens captions before generation. Disable with --no-optimize for speed. Generation takes 1-5 minutes depending on iterations and provider. The script prints progress. Output is delivered automatically via the MEDIA: protocol β no manual file handling needed. Run continuation is the natural way to iterate: "make it better" β --continue --feedback "...". Gemini free tier has rate limits (~15 RPM). Keep iterations β€ 3 on free tier.
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.