Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Run AI-powered mock interviews using Fuku.ai's free public service.
Run AI-powered mock interviews using Fuku.ai's free public service.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Turn a folder full of resumes into structured AI interview sessions. This skill uses Fuku.ai's free public API to generate AI interview reports.
This is a Fuku.ai-specific skill that leverages their free, anonymous interview service. No user account, API key, or login required.
This skill uses shared anonymous credentials provided by Fuku.ai for public access: ItemValueUpload Endpointhttps://hapi.fuku.ai/hr/rc/anon/file/uploadJob APIhttps://hapi.fuku.ai/hr/rc/anon/job/invite/ai_interviewX-NUMBER Headerjob-Z4nV8cQ1LmT7XpR2bH9sJdK6WyEaF0uid Query Param1873977344885133312 These are fixed, shared credentials for Fuku.ai's free tier. All users of this skill use the same endpoints and identifiers. This is intentional—the service is designed for anonymous, no-login usage.
No user credentials required: The service is free and public No environment variables: Endpoints and credentials are hardcoded by design Not self-hostable: This skill only works with Fuku.ai's hosted service Privacy consideration: Resume files are sent to Fuku.ai's servers. Review their privacy policy before uploading sensitive documents.
Collects three mandatory inputs: job title, company name, report email. Scans a resume folder for PDF/DOC/DOCX files (up to 100). Uploads each resume to Fuku.ai's public endpoint and captures the returned file URLs. Creates an AI interview job via Fuku.ai's API using the shared anonymous credentials. Logs minimal job metadata locally and confirms the report email destination.
Install & Prepare cd skills/ai-interview && npm install (installs axios + form-data). Subsequent runs only need node run.js .... Gather Inputs Ask the user for: job title, company, email for reports (validate email format). Ask for the resume folder path inside the workspace. Confirm contents before proceeding. Scan Folder Accept only .pdf, .doc, .docx files. Abort if folder is empty or missing. Upload Resumes From skills/ai-interview/, run node run.js --folder <dir> --jobTitle <title> --company <company> --email <email>. The script auto-detects .pdf/.doc/.docx files (up to 100), uploads each to Fuku.ai's upload endpoint, and captures the returned file URLs. On any failed upload, the script aborts and reports the .desc field from the API. Trigger Interview Job The same script immediately calls the interview creation endpoint with payload { jobTitle, company, email, fileUrls }. Authentication uses hardcoded credentials (see "Hardcoded Configuration" table above). Expect response { "code": 0, "data": { "id", "company", "title", ... } }. If code is not 0, the script surfaces the error and stops. Report Back Confirm job creation, list resumes included, and restate the email destination. Persist only the minimal identifiers (id, company, title) into ai-interview/jobs/<timestamp>.json—no need to keep full payloads. Remind the user that AI interview reports are delivered directly to the email they provided. A typical success response looks like: { "code": 0, "data": { "id": "5b16b2d2f5e947f78244246a9f24e2cb", "company": "FUKU", "title": "cfoe", ... (truncated) }, "desc": "successful" }
Email: Must match /^[^@\s]+@[^@\s]+\.[^@\s]+$/. File count: Maximum 100 resumes per batch. Upload errors: The Fuku API must return code: 0; otherwise surface the .desc field and ask whether to retry or skip that file. PII handling: Do not log resume contents—only file names are logged (not full remote URLs). HTTPS: Both endpoints use HTTPS. Data destination: Resume files are sent to Fuku.ai's third-party service. Review their privacy policy before uploading sensitive documents.
Minimal audit trail stored under ai-interview/jobs/: ai-interview/ jobs/ 2026-02-27T08-30-00Z.json # job identifiers only Each file contains only the essential identifiers (no resume data or full API responses): { "timestamp": "2026-02-27T08:30:00Z", "jobId": "5b16b2d2f5e947f78244246a9f24e2cb", "company": "FUKU", "title": "cfoe" }
Use dummy resumes and a test email address for initial runs. Validate that the email receives the AI interview report before marking the job done. Note: This skill only works with Fuku.ai's production endpoints—there is no staging/mock mode.
IssueCauseFixUpload returns 413File too largeCompress resume or raise server limitfileUrls emptyUpload failed silentlyCheck upload response for success flagAPI 400Missing fieldsEnsure jobTitle/company/email/fileUrls filled
"Great! Need the job title, company name, a mailbox for the interview report, and the folder path containing the resumes (PDF/DOC/DOCX)."
Automate email notifications to confirm when the interview report is delivered (future enhancement). Add optional metadata per candidate (experience, notes) by extending the payload. Happy interviewing! 🎙️
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.