โ† All skills
Tencent SkillHub ยท Developer Tools

QA & Test Engineering Command Center

Comprehensive QA system for planning strategy, writing tests, analyzing coverage, automating pipelines, performance and security testing, defect triage, and...

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

Comprehensive QA system for planning strategy, writing tests, analyzing coverage, automating pipelines, performance and security testing, defect triage, and...

โฌ‡ 0 downloads โ˜… 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
README.md, SKILL.md

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
1.0.0

Documentation

ClawHub primary doc Primary doc: SKILL.md 34 sections Open source page

QA & Test Engineering Command Center

Complete quality assurance system โ€” from test strategy to automation frameworks, coverage analysis, and release readiness. Works for any stack, any team size.

When to Use

Planning test strategy for a new feature or project Writing unit, integration, or E2E tests Reviewing test quality and coverage gaps Setting up test automation and CI/CD quality gates Performance testing and load analysis Security testing checklist Bug triage and defect management Release readiness assessment

Strategy Brief

Before writing any tests, define the strategy: # test-strategy.yaml project: "[name]" scope: "[feature/module/full product]" risk_level: high | medium | low stack: language: "[TypeScript/Python/Java/Go]" framework: "[React/Express/Django/Spring]" test_runner: "[Jest/Vitest/pytest/JUnit/Go test]" e2e_tool: "[Playwright/Cypress/Selenium]" # What are we testing? test_scope: - area: "[e.g., Auth module]" risk: high test_types: [unit, integration, e2e] priority: 1 - area: "[e.g., Settings page]" risk: low test_types: [unit] priority: 3 # What's NOT in scope (and why) exclusions: - "[e.g., Third-party widget โ€” covered by vendor]" # Quality targets targets: line_coverage: 80 branch_coverage: 70 critical_path_coverage: 100 max_flaky_rate: 2% max_test_duration_unit: 10ms max_test_duration_integration: 500ms max_test_duration_e2e: 30s

Risk-Based Test Allocation

Not everything needs the same testing depth. Use the risk matrix: Risk LevelUnit TestsIntegrationE2EManual/ExploratoryCritical (payments, auth, data loss)95%+ coverageFull API coverageHappy + error pathsExploratory sessionHigh (core features, user-facing)85%+ coverageKey integrationsHappy pathSpot checkMedium (secondary features)70%+ coverageCritical paths onlySmoke onlyOn releaseLow (admin, internal tools)50%+ coverageNoneNoneNone

Test Pyramid

Follow the pyramid โ€” not the ice cream cone: / E2E \ โ† Few (5-10%) โ€” slow, expensive, brittle / Integr. \ โ† Some (15-25%) โ€” API contracts, DB queries / Unit \ โ† Many (65-80%) โ€” fast, isolated, cheap Anti-pattern: Ice cream cone (mostly E2E, few unit tests) = slow CI, flaky builds, expensive maintenance. Decision rule: Can this be tested at a lower level? โ†’ Test it there.

Anatomy of a Good Unit Test

Every unit test follows AAA (Arrange-Act-Assert): 1. ARRANGE โ€” Set up test data, mocks, state 2. ACT โ€” Call the function/method under test 3. ASSERT โ€” Verify the output matches expectations

Unit Test Checklist (per function)

For each function/method, verify: Happy path โ€” expected input โ†’ expected output Edge cases โ€” empty input, null/undefined, zero, max values Boundary values โ€” off-by-one, min-1, max+1 Error handling โ€” invalid input โ†’ correct error thrown Return types โ€” correct type, shape, structure Side effects โ€” does it modify state it shouldn't? Idempotency โ€” calling twice gives same result?

What to Mock (and What NOT to Mock)

Mock these: External APIs (HTTP calls, third-party services) Database queries (in unit tests only) File system operations Date/time (use fake timers) Random number generators Environment variables DO NOT mock these: The function under test itself Pure utility functions (test them directly) Data transformations Simple value objects Mock rule of thumb: If removing the mock would make the test hit the network, file system, or database โ†’ mock it. Otherwise โ†’ don't.

Test Naming Convention

Use the pattern: [unit] [scenario] [expected result] Examples: calculateTotal returns 0 for empty cart validateEmail throws for missing @ symbol parseDate handles ISO 8601 with timezone offset

Coverage Analysis

Metrics that matter: MetricTargetWhyLine coverage80%+Basic completenessBranch coverage70%+Catches missed if/else pathsFunction coverage90%+Ensures all functions are testedCritical path coverage100%Business-critical code fully verified Coverage traps to avoid: 100% line coverage โ‰  good tests (assertions matter more than lines hit) Coverage on generated code inflates numbers Trivial getters/setters pad coverage without value Coverage should INCREASE over time, never decrease

What Integration Tests Cover

Integration tests verify that components work TOGETHER: API endpoint โ†’ middleware โ†’ handler โ†’ database โ†’ response Service A calls Service B and handles the response Message queue producer โ†’ consumer โ†’ side effect Auth flow: login โ†’ token โ†’ authenticated request

Integration Test Patterns

Pattern 1: API Contract Testing 1. Start test server (or use supertest/httptest) 2. Send HTTP request with specific payload 3. Assert: status code, response body shape, headers 4. Assert: database state changed correctly 5. Assert: side effects triggered (emails, events) Pattern 2: Database Integration 1. Start test database (SQLite in-memory or test container) 2. Run migrations 3. Seed test data 4. Execute query/operation 5. Assert: data matches expectations 6. Teardown (truncate or rollback transaction) Pattern 3: External Service 1. Record real API response (VCR/nock/wiremock) 2. Replay recorded response in tests 3. Assert: your code handles the response correctly 4. Also test: timeout, 500 error, malformed response

Integration Test Checklist

Happy path โ€” full flow works end-to-end Auth โ€” unauthenticated returns 401, wrong role returns 403 Validation โ€” bad payload returns 400 with error details Not found โ€” missing resource returns 404 Conflict โ€” duplicate create returns 409 Rate limiting โ€” excessive requests return 429 Database constraints โ€” unique violations, foreign keys Concurrency โ€” two simultaneous writes don't corrupt data Timeout handling โ€” external service timeout โ†’ graceful fallback

E2E Strategy

E2E tests verify complete user journeys. They're expensive โ€” be strategic: Test these E2E: User registration โ†’ email verification โ†’ first login Purchase flow โ†’ payment โ†’ confirmation Critical business workflows (the ones that make money) Cross-browser/device smoke tests DON'T test these E2E: Individual form validations (unit test) API error handling (integration test) Edge cases (lower-level tests) Visual styling (visual regression tools)

E2E Test Template

test_name: "[User journey name]" preconditions: - "[User is logged in]" - "[Product exists in catalog]" steps: - action: "Navigate to /products" verify: "Product list is visible" - action: "Click 'Add to Cart' on Product A" verify: "Cart badge shows 1" - action: "Click 'Checkout'" verify: "Checkout form displayed" - action: "Fill payment details and submit" verify: "Order confirmation page with order ID" postconditions: - "Order exists in database with status 'paid'" - "Confirmation email sent" max_duration: 30s

Flaky Test Management

Flaky tests are the #1 CI killer. Handle them: Flaky Test Triage: Identify โ€” Track test pass rates over 10+ runs Classify โ€” Why is it flaky? Timing/race condition โ†’ Add explicit waits, not sleep() Test data dependency โ†’ Isolate test data per run External service โ†’ Mock it or use test container Browser rendering โ†’ Use visibility checks, not delays Quarantine โ€” Move to @flaky suite, run separately Fix or delete โ€” Flaky test unfixed for 2 weeks โ†’ delete it Flaky rate target: < 2% of total test runs

Performance Test Types

TypePurposeWhenLoad testNormal traffic handlingBefore every releaseStress testFind breaking pointQuarterly or before scalingSpike testSudden traffic burstBefore marketing campaignsSoak testMemory leaks over timeMonthly or after major changesCapacity testMax users/throughputPlanning infrastructure

Performance Test Plan

test_name: "[API/Page] Load Test" target: "[URL or endpoint]" baseline: p50_response: "[current p50 ms]" p95_response: "[current p95 ms]" p99_response: "[current p99 ms]" error_rate: "[current %]" scenarios: - name: "Normal load" vus: 50 # virtual users duration: 5m ramp_up: 30s thresholds: p95_response: "< 500ms" error_rate: "< 1%" - name: "Peak load" vus: 200 duration: 10m ramp_up: 1m thresholds: p95_response: "< 2000ms" error_rate: "< 5%" - name: "Stress test" vus: 500 duration: 5m ramp_up: 2m # Find the breaking point โ€” no thresholds, observe

Performance Metrics Dashboard

Track these per endpoint: MetricGreenYellowRedp50 response< 200ms200-500ms> 500msp95 response< 500ms500ms-2s> 2sp99 response< 1s1-5s> 5sError rate< 0.1%0.1-1%> 1%Throughput> baseline80-100% baseline< 80%CPU usage< 60%60-80%> 80%Memory usage< 70%70-85%> 85%DB query time< 50ms avg50-200ms> 200ms

Common Performance Fixes

SymptomLikely CauseFixSlow API responseN+1 queriesBatch/join queriesMemory climbingObject retentionProfile heap, fix leaksTimeout spikesConnection pool exhaustionIncrease pool, add queuingSlow page loadLarge bundleCode split, lazy loadDB bottleneckMissing indexAdd index on WHERE/JOIN columnsHigh CPUSynchronous computeMove to worker/queue

Security Test Checklist

Run through these for every feature/release: Authentication & Authorization: Passwords hashed with bcrypt/argon2 (not MD5/SHA1) Session tokens are random, sufficient length (128+ bits) JWT tokens have short expiry (15 min access, 7 day refresh) Failed login rate limiting (5 attempts โ†’ lockout) Password reset tokens expire (1 hour max) Role-based access enforced server-side (not just UI) Can't access other users' data by changing IDs in URL Input Validation: SQL injection โ€” parameterized queries everywhere XSS โ€” output encoding, CSP headers CSRF โ€” tokens on state-changing requests Path traversal โ€” validate file paths, no ../ Command injection โ€” never pass user input to shell File upload โ€” validate type, size, scan for malware JSON/XML parsing โ€” depth limits, entity expansion disabled Data Protection: HTTPS everywhere (HSTS header) Sensitive data encrypted at rest PII not logged (mask in log output) API keys not in client-side code CORS configured correctly (not *) Security headers set (X-Frame-Options, X-Content-Type-Options) Infrastructure: Dependencies scanned for CVEs (npm audit / pip audit) Docker images scanned (Trivy/Snyk) Secrets not in code/env files (use vault) Error messages don't leak internals Admin endpoints behind VPN/IP allowlist

OWASP Top 10 Quick Reference

#VulnerabilityTest ForA01Broken Access ControlAccess other users' resources, bypass role checksA02Cryptographic FailuresWeak hashing, plaintext secrets, expired certsA03InjectionSQL, XSS, command, LDAP injectionA04Insecure DesignBusiness logic flaws, missing rate limitsA05Security MisconfigurationDefault creds, verbose errors, open portsA06Vulnerable ComponentsOutdated deps with known CVEsA07Authentication FailuresBrute force, weak passwords, session fixationA08Data Integrity FailuresUnsigned updates, CI/CD pipeline injectionA09Logging FailuresMissing audit logs, no alerting on breachesA10SSRFInternal network access via user-controlled URLs

Bug Report Template

bug_id: "[auto or manual]" title: "[Short description of the bug]" severity: P0-critical | P1-high | P2-medium | P3-low reporter: "[name]" date: "[YYYY-MM-DD]" environment: os: "[OS + version]" browser: "[Browser + version]" app_version: "[version/commit]" steps_to_reproduce: 1. "[Step 1]" 2. "[Step 2]" 3. "[Step 3]" expected_result: "[What should happen]" actual_result: "[What actually happens]" frequency: "always | intermittent | once" screenshots: "[links]" logs: "[relevant log output]"

Severity Classification

LevelDefinitionSLAExampleP0 CriticalSystem down, data loss, security breachFix in 4 hoursPayment processing brokenP1 HighMajor feature broken, no workaroundFix in 24 hoursUsers can't loginP2 MediumFeature broken with workaroundFix this sprintSearch returns wrong results sometimesP3 LowMinor issue, cosmeticFix when convenientButton alignment off by 2px

Bug Triage Process (Weekly)

1. Review all new bugs (unassigned) 2. For each bug: a. Reproduce โ€” can you trigger it? b. Classify severity (P0-P3) c. Estimate fix effort (S/M/L) d. Assign to owner + sprint e. Link to related bugs/stories 3. Review P0/P1 bugs from last week โ€” are they fixed? 4. Close bugs that can't be reproduced (after 2 attempts) 5. Update metrics dashboard

Bug Metrics Dashboard

Track weekly: MetricFormulaTargetBug escape rateBugs found in prod / total bugs< 10%Mean time to fix (P0)Avg hours from report to deploy< 8 hoursMean time to fix (P1)Avg hours from report to deploy< 48 hoursBug reopen rateReopened bugs / closed bugs< 5%Test escape analysisBugs that SHOULD have been caughtTrack & reduceOpen bug countTotal open by severityTrending down

Release Checklist

Before shipping to production: Code Quality: All unit tests passing All integration tests passing E2E smoke suite passing No new lint warnings/errors Code reviewed and approved No known P0/P1 bugs open for this release Coverage & Quality Gates: Line coverage โ‰ฅ target (80%) Branch coverage โ‰ฅ target (70%) No coverage decrease from last release Mutation testing score โ‰ฅ 60% (if applicable) Performance: Load test passed (within thresholds) No performance regressions vs baseline Bundle size within budget Security: Dependency audit clean (no critical/high CVEs) Security checklist completed Secrets rotated if needed Operational Readiness: Monitoring/alerts configured for new features Rollback plan documented Feature flags in place for risky changes Database migration tested and reversible Runbook updated

Release Readiness Score

Score 0-100 across 5 dimensions: DimensionWeightScoringTest coverage25%100 if targets met, -10 per gap areaBug status25%100 if 0 P0/P1, -20 per open P0, -10 per P1Performance20%100 if all green, -15 per yellow, -30 per redSecurity20%100 if clean, -25 per critical, -15 per highOperational10%100 if checklist complete, -20 per missing item Ship threshold: โ‰ฅ 80 overall, no dimension below 60

Pipeline Quality Gates

Configure these gates in your CI pipeline: # Quality gate configuration gates: - name: "Lint" stage: pre-commit command: "npm run lint" blocking: true - name: "Unit Tests" stage: commit command: "npm test -- --coverage" blocking: true thresholds: pass_rate: 100% coverage_line: 80% coverage_branch: 70% - name: "Integration Tests" stage: merge command: "npm run test:integration" blocking: true thresholds: pass_rate: 100% - name: "Security Scan" stage: merge command: "npm audit --audit-level=high" blocking: true - name: "E2E Smoke" stage: staging command: "npm run test:e2e:smoke" blocking: true thresholds: pass_rate: 100% - name: "Performance" stage: staging command: "npm run test:perf" blocking: false # Alert only thresholds: p95_regression: 20%

Test Automation Maturity Model

Rate your team 1-5: LevelDescriptionCharacteristics1 โ€” ManualAll testing is manualNo automation, long release cycles2 โ€” ReactiveSome unit tests, no CITests written after bugs, not before3 โ€” StructuredTest pyramid, CI pipelineUnit + integration, automated on push4 โ€” ProactiveFull automation, quality gatesE2E + perf + security in pipeline, TDD5 โ€” OptimizedSelf-healing, predictiveFlaky auto-quarantine, AI-assisted testing, continuous deployment

Weekly Test Health Review

review_date: "[YYYY-MM-DD]" metrics: total_tests: 0 pass_rate_7d: "0%" flaky_tests: 0 flaky_rate: "0%" avg_suite_duration: "0s" coverage_line: "0%" coverage_branch: "0%" actions: quarantined: [] # Tests moved to flaky suite deleted: [] # Tests removed (obsolete/unfixable) fixed: [] # Flaky tests fixed this week added: [] # New tests added trends: coverage_delta: "+0%" # vs last week flaky_delta: "+0" # vs last week duration_delta: "+0s" # vs last week notes: ""

Test Maintenance Rules

No commented-out tests โ€” delete or fix, never comment No skipped tests > 2 weeks โ€” fix or remove No test duplication โ€” each behavior tested once at the right level Test names must be readable โ€” someone new should understand what broke Shared test utilities โ€” common setup in fixtures/factories, not copy-pasted Test data isolation โ€” each test creates its own data, cleans up after No magic numbers โ€” use named constants in assertions Assertion messages โ€” custom messages on complex assertions

Common Test Anti-Patterns

Anti-PatternProblemFixSleeping testssleep(2000) instead of waitingUse explicit waits/pollingTest interdependenceTest B relies on Test A's stateIsolate โ€” each test sets up its own stateAssertionless testsTest runs code but doesn't assertAdd meaningful assertionsBrittle selectorsCSS selectors that break on redesignUse data-testid or aria rolesGod testOne test verifying 20 thingsSplit into focused testsMock overloadEverything mocked, nothing real testedOnly mock external boundariesHardcoded dataTests break when seed data changesUse factories/buildersIgnoring test output"It passed, ship it"Review WHY it passed โ€” is the assertion meaningful?

Quick Reference: Natural Language Commands

Tell the agent: "Create test strategy for [feature]" โ†’ Generates strategy brief "Write unit tests for [function/file]" โ†’ AAA-structured tests with edge cases "Review test coverage for [module]" โ†’ Gap analysis + recommendations "Write integration tests for [API endpoint]" โ†’ Full HTTP test suite "Plan E2E tests for [user journey]" โ†’ E2E test template "Run security checklist for [feature]" โ†’ OWASP-based security review "Triage these bugs: [list]" โ†’ Severity classification + assignment "Release readiness check" โ†’ Full readiness score + blockers "Performance test plan for [endpoint]" โ†’ Load/stress test configuration "Fix flaky test [name]" โ†’ Root cause analysis + fix strategy

Category context

Code helpers, APIs, CLIs, browser automation, testing, and developer operations.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
2 Docs
  • SKILL.md Primary doc
  • README.md Docs