{
  "schemaVersion": "1.0",
  "item": {
    "slug": "singleshot-prompt-testing",
    "name": "Singleshot Prompt Testing",
    "source": "tencent",
    "type": "skill",
    "category": "开发工具",
    "sourceUrl": "https://clawhub.ai/vincentzhangz/singleshot-prompt-testing",
    "canonicalUrl": "https://clawhub.ai/vincentzhangz/singleshot-prompt-testing",
    "targetPlatform": "OpenClaw"
  },
  "install": {
    "downloadMode": "redirect",
    "downloadUrl": "/downloads/singleshot-prompt-testing",
    "sourceDownloadUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=singleshot-prompt-testing",
    "sourcePlatform": "tencent",
    "targetPlatform": "OpenClaw",
    "installMethod": "Manual import",
    "extraction": "Extract archive",
    "prerequisites": [
      "OpenClaw"
    ],
    "packageFormat": "ZIP package",
    "includedAssets": [
      "QUICKSTART.md",
      "metadata.json",
      "README.md",
      "SKILL.md",
      "RELEASE_NOTES.md"
    ],
    "primaryDoc": "SKILL.md",
    "quickSetup": [
      "Download the package from Yavira.",
      "Extract the archive and review SKILL.md first.",
      "Import or place the package into your OpenClaw setup."
    ],
    "agentAssist": {
      "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
      "steps": [
        "Download the package from Yavira.",
        "Extract it into a folder your agent can access.",
        "Paste one of the prompts below and point your agent at the extracted folder."
      ],
      "prompts": [
        {
          "label": "New install",
          "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete."
        },
        {
          "label": "Upgrade existing",
          "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run."
        }
      ]
    },
    "sourceHealth": {
      "source": "tencent",
      "status": "healthy",
      "reason": "direct_download_ok",
      "recommendedAction": "download",
      "checkedAt": "2026-05-07T17:22:31.273Z",
      "expiresAt": "2026-05-14T17:22:31.273Z",
      "httpStatus": 200,
      "finalUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=afrexai-annual-report",
      "contentType": "application/zip",
      "probeMethod": "head",
      "details": {
        "probeUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=afrexai-annual-report",
        "contentDisposition": "attachment; filename=\"afrexai-annual-report-1.0.0.zip\"",
        "redirectLocation": null,
        "bodySnippet": null
      },
      "scope": "source",
      "summary": "Source download looks usable.",
      "detail": "Yavira can redirect you to the upstream package for this source.",
      "primaryActionLabel": "Download for OpenClaw",
      "primaryActionHref": "/downloads/singleshot-prompt-testing"
    },
    "validation": {
      "installChecklist": [
        "Use the Yavira download entry.",
        "Review SKILL.md after the package is downloaded.",
        "Confirm the extracted package contains the expected setup assets."
      ],
      "postInstallChecks": [
        "Confirm the extracted package includes the expected docs or setup files.",
        "Validate the skill or prompts are available in your target agent workspace.",
        "Capture any manual follow-up steps the agent could not complete."
      ]
    },
    "downloadPageUrl": "https://openagent3.xyz/downloads/singleshot-prompt-testing",
    "agentPageUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing/agent",
    "manifestUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing/agent.md"
  },
  "agentAssist": {
    "summary": "Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.",
    "steps": [
      "Download the package from Yavira.",
      "Extract it into a folder your agent can access.",
      "Paste one of the prompts below and point your agent at the extracted folder."
    ],
    "prompts": [
      {
        "label": "New install",
        "body": "I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete."
      },
      {
        "label": "Upgrade existing",
        "body": "I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run."
      }
    ]
  },
  "documentation": {
    "source": "clawhub",
    "primaryDoc": "SKILL.md",
    "sections": [
      {
        "title": "Description",
        "body": "Prompt cost testing with single shot"
      },
      {
        "title": "Installation",
        "body": "brew tap vincentzhangz/singleshot\nbrew install singleshot\n\nOr: cargo install singleshot"
      },
      {
        "title": "When to Use",
        "body": "Testing new prompts before openclaw implementation\nBenchmarking prompt variations for token efficiency\nComparing model performance and costs\nValidating prompt outputs before production"
      },
      {
        "title": "Core Commands",
        "body": "Always use -d (detail) and -r (report) flags for efficiency analysis:\n\n# Basic test with full metrics\nsingleshot chat -p \"Your prompt\" -P openai -d -r report.md\n\n# Test with config file\nsingleshot chat -l config.md -d -r report.md\n\n# Compare providers\nsingleshot chat -p \"Test\" -P openai -m gpt-4o-mini -d -r openai.md\nsingleshot chat -p \"Test\" -P anthropic -m claude-sonnet-4-20250514 -d -r anthropic.md\n\n# Batch test variations\nfor config in *.md; do\n  singleshot chat -l \"$config\" -d -r \"report-${config%.md}.md\"\ndone"
      },
      {
        "title": "1. Generate Baseline",
        "body": "singleshot chat -p \"Your prompt\" -P openai -d -r baseline.md\ncat baseline.md"
      },
      {
        "title": "2. Optimize & Compare",
        "body": "# Create optimized version, test, and compare\ncat > optimized.md << 'EOF'\n---provider---\nopenai\n---model---\ngpt-4o-mini\n---max_tokens---\n200\n---system---\nExpert. Be concise.\n---prompt---\nYour optimized prompt\nEOF\n\nsingleshot chat -l optimized.md -d -r optimized-report.md\n\n# Compare metrics\necho \"Baseline:\" && grep -E \"(Tokens|Cost)\" baseline.md\necho \"Optimized:\" && grep -E \"(Tokens|Cost)\" optimized-report.md"
      },
      {
        "title": "Report Metrics",
        "body": "Reports contain:\n\n## Token Usage\n- Input Tokens: 245\n- Output Tokens: 180\n- Total Tokens: 425\n\n## Cost (estimated)\n- Input Cost: $0.00003675\n- Output Cost: $0.000108\n- Total Cost: $0.00014475\n\n## Timing\n- Time to First Token: 0.45s\n- Total Time: 1.23s"
      },
      {
        "title": "Optimization Strategies",
        "body": "Test with cheaper models first:\nsingleshot chat -p \"Test\" -P openai -m gpt-4o-mini -d -r report.md\n\n\n\nReduce tokens:\n\nShorten system prompts\nUse --max-tokens to limit output\nAdd \"be concise\" to system prompt\n\n\n\nTest locally (free):\nsingleshot chat -p \"Test\" -P ollama -m llama3.2 -d -r report.md"
      },
      {
        "title": "Example: Full Optimization",
        "body": "# Step 1: Baseline (verbose)\nsingleshot chat \\\n  -p \"How do I write a Rust function to add two numbers?\" \\\n  -s \"You are an expert Rust programmer with 10 years experience\" \\\n  -P openai -d -r v1.md\n\n# Step 2: Read metrics\ncat v1.md\n# Expected: ~130 input tokens, ~400 output tokens\n\n# Step 3: Optimized version\nsingleshot chat \\\n  -p \"Rust function: add(a: i32, b: i32) -> i32\" \\\n  -s \"Rust expert. Code only.\" \\\n  -P openai --max-tokens 100 -d -r v2.md\n\n# Step 4: Compare\necho \"=== COMPARISON ===\"\ngrep \"Total Cost\" v1.md v2.md\ngrep \"Total Tokens\" v1.md v2.md"
      },
      {
        "title": "Quick Reference",
        "body": "# Test with full details\nsingleshot chat -p \"prompt\" -P openai -d -r report.md\n\n# Extract metrics\ngrep -E \"(Input|Output|Total)\" report.md\n\n# Compare reports\ndiff report1.md report2.md\n\n# Vision test\nsingleshot chat -p \"Describe\" -i image.jpg -P openai -d -r report.md\n\n# List models\nsingleshot models -P openai\n\n# Test connection\nsingleshot ping -P openai"
      },
      {
        "title": "Environment Variables",
        "body": "export OPENAI_API_KEY=\"sk-...\"\nexport ANTHROPIC_API_KEY=\"sk-ant-...\"\nexport OPENROUTER_API_KEY=\"sk-or-...\""
      },
      {
        "title": "Best Practices",
        "body": "Always use -d for detailed token metrics\nAlways use -r to save reports\nAlways cat reports to analyze metrics\nTest variations and compare costs\nSet --max-tokens to control costs\nUse gpt-4o-mini for testing (cheaper)"
      },
      {
        "title": "Troubleshooting",
        "body": "No metrics: Ensure -d flag is used\nNo report file: Ensure -r flag is used\nHigh costs: Switch to gpt-4o-mini or Ollama\nConnection issues: Run singleshot ping -P <provider>"
      }
    ],
    "body": "Singleshot Prompt Testing & Optimization Skill\nDescription\n\nPrompt cost testing with single shot\n\nInstallation\nbrew tap vincentzhangz/singleshot\nbrew install singleshot\n\n\nOr: cargo install singleshot\n\nWhen to Use\nTesting new prompts before openclaw implementation\nBenchmarking prompt variations for token efficiency\nComparing model performance and costs\nValidating prompt outputs before production\nCore Commands\n\nAlways use -d (detail) and -r (report) flags for efficiency analysis:\n\n# Basic test with full metrics\nsingleshot chat -p \"Your prompt\" -P openai -d -r report.md\n\n# Test with config file\nsingleshot chat -l config.md -d -r report.md\n\n# Compare providers\nsingleshot chat -p \"Test\" -P openai -m gpt-4o-mini -d -r openai.md\nsingleshot chat -p \"Test\" -P anthropic -m claude-sonnet-4-20250514 -d -r anthropic.md\n\n# Batch test variations\nfor config in *.md; do\n  singleshot chat -l \"$config\" -d -r \"report-${config%.md}.md\"\ndone\n\nReport Analysis Workflow\n1. Generate Baseline\nsingleshot chat -p \"Your prompt\" -P openai -d -r baseline.md\ncat baseline.md\n\n2. Optimize & Compare\n# Create optimized version, test, and compare\ncat > optimized.md << 'EOF'\n---provider---\nopenai\n---model---\ngpt-4o-mini\n---max_tokens---\n200\n---system---\nExpert. Be concise.\n---prompt---\nYour optimized prompt\nEOF\n\nsingleshot chat -l optimized.md -d -r optimized-report.md\n\n# Compare metrics\necho \"Baseline:\" && grep -E \"(Tokens|Cost)\" baseline.md\necho \"Optimized:\" && grep -E \"(Tokens|Cost)\" optimized-report.md\n\nReport Metrics\n\nReports contain:\n\n## Token Usage\n- Input Tokens: 245\n- Output Tokens: 180\n- Total Tokens: 425\n\n## Cost (estimated)\n- Input Cost: $0.00003675\n- Output Cost: $0.000108\n- Total Cost: $0.00014475\n\n## Timing\n- Time to First Token: 0.45s\n- Total Time: 1.23s\n\nOptimization Strategies\n\nTest with cheaper models first:\n\nsingleshot chat -p \"Test\" -P openai -m gpt-4o-mini -d -r report.md\n\n\nReduce tokens:\n\nShorten system prompts\nUse --max-tokens to limit output\nAdd \"be concise\" to system prompt\n\nTest locally (free):\n\nsingleshot chat -p \"Test\" -P ollama -m llama3.2 -d -r report.md\n\nExample: Full Optimization\n# Step 1: Baseline (verbose)\nsingleshot chat \\\n  -p \"How do I write a Rust function to add two numbers?\" \\\n  -s \"You are an expert Rust programmer with 10 years experience\" \\\n  -P openai -d -r v1.md\n\n# Step 2: Read metrics\ncat v1.md\n# Expected: ~130 input tokens, ~400 output tokens\n\n# Step 3: Optimized version\nsingleshot chat \\\n  -p \"Rust function: add(a: i32, b: i32) -> i32\" \\\n  -s \"Rust expert. Code only.\" \\\n  -P openai --max-tokens 100 -d -r v2.md\n\n# Step 4: Compare\necho \"=== COMPARISON ===\"\ngrep \"Total Cost\" v1.md v2.md\ngrep \"Total Tokens\" v1.md v2.md\n\nQuick Reference\n# Test with full details\nsingleshot chat -p \"prompt\" -P openai -d -r report.md\n\n# Extract metrics\ngrep -E \"(Input|Output|Total)\" report.md\n\n# Compare reports\ndiff report1.md report2.md\n\n# Vision test\nsingleshot chat -p \"Describe\" -i image.jpg -P openai -d -r report.md\n\n# List models\nsingleshot models -P openai\n\n# Test connection\nsingleshot ping -P openai\n\nEnvironment Variables\nexport OPENAI_API_KEY=\"sk-...\"\nexport ANTHROPIC_API_KEY=\"sk-ant-...\"\nexport OPENROUTER_API_KEY=\"sk-or-...\"\n\nBest Practices\nAlways use -d for detailed token metrics\nAlways use -r to save reports\nAlways cat reports to analyze metrics\nTest variations and compare costs\nSet --max-tokens to control costs\nUse gpt-4o-mini for testing (cheaper)\nTroubleshooting\nNo metrics: Ensure -d flag is used\nNo report file: Ensure -r flag is used\nHigh costs: Switch to gpt-4o-mini or Ollama\nConnection issues: Run singleshot ping -P <provider>"
  },
  "trust": {
    "sourceLabel": "tencent",
    "provenanceUrl": "https://clawhub.ai/vincentzhangz/singleshot-prompt-testing",
    "publisherUrl": "https://clawhub.ai/vincentzhangz/singleshot-prompt-testing",
    "owner": "vincentzhangz",
    "version": "0.1.0",
    "license": null,
    "verificationStatus": "Indexed source record"
  },
  "links": {
    "detailUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing",
    "downloadUrl": "https://openagent3.xyz/downloads/singleshot-prompt-testing",
    "agentUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing/agent",
    "manifestUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/singleshot-prompt-testing/agent.md"
  }
}