Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Scrapes and returns a fully translated company profile from the Indonesia Stock Exchange website using a given IDX stock code.
Scrapes and returns a fully translated company profile from the Indonesia Stock Exchange website using a given IDX stock code.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
name: "idx-stock" description: "Scrapes complete company profile data from the Indonesia Stock Exchange (IDX) website based on stock code. Automatically translates all field keys from Indonesian to English. Returns clean structured data with status and message." version: "1.0.0" author: username: "@zororaka00" name: "Web3 Hungry" x_account: "https://x.com/web3hungry" category: "finance" tags: ["idx", "stock", "scraper", "company-profile", "exchange", "indonesia"] runtime: language: "python" version: "3.11+" dependencies: - "requests>=2.31.0" - "beautifulsoup4>=4.12.0"
uv add requests beautifulsoup4
pip install requests beautifulsoup4
pip3 install requests beautifulsoup4 Note: The skill will automatically check for uv first, then fall back to pip/pip3 if uv is not available. parameters: name: "stock_code" type: "string" description: "IDX stock code (example: BBCA, TLKM, ADRO). Automatically converted to uppercase." required: true example: "BBCA" outputs: name: "success" type: "boolean" name: "data" type: "object" description: "Dictionary containing full company profile with English keys" name: "message" type: "string"
This skill fetches the complete listed company profile from the official Indonesia Stock Exchange (IDX) website. It processes all tables (general information, listing data, directors, commissioners, shareholders, etc.) and returns data with standardized English keys. No browser or Selenium required (fast static scraping) Strict input validation and sanitization Read-only HTTP GET to official IDX domain only Fully secure and VirusTotal-compliant Security Statement: This skill is designed to be fully secure and passes Clawhub's integrated VirusTotal security scan. It contains no viruses, malware, file access, network exploits, or suspicious behaviors. Only safe, vetted libraries (requests + beautifulsoup4) are used with best practices for input validation, error handling, and data sanitization. Created by Username: @zororaka00 Name: "Web3 Hungry" X Account: https://x.com/web3hungry
import requests from bs4 import BeautifulSoup from typing import Dict, Any def translate_key(key: str) -> str: """Translate key from Indonesian to English using a predefined dictionary.""" translations = { "Nama": "name", "Kode": "code", "Alamat Kantor": "office_address", "Alamat Email": "email", "Telepon": "phone", "Fax": "fax", "NPWP": "tax_id", "Situs": "website", "Tanggal Pencatatan": "listing_date", "Papan Pencatatan": "board", "Bidang Usaha Utama": "main_business", "Sektor": "sector", "Subsektor": "subsector", "Industri": "industry", "Subindustri": "subindustry", "Biro Administrasi Efek": "share_registrar", } return translations.get(key.strip(), key.lower().replace(" ", "_").replace(":", "")) def idx_stock_profile(stock_code: str) -> Dict[str, Any]: """Main function for idx-stock skill on Clawhub.""" # Input sanitization and validation stock_code = stock_code.strip().upper() if not stock_code or len(stock_code) < 3 or len(stock_code) > 5 or not stock_code.isalnum(): return { "success": False, "data": {}, "message": "Invalid stock code format. Example: BBCA" } url = f"https://www.idx.co.id/id/perusahaan-tercatat/profil-perusahaan-tercatat/{stock_code}" headers = { "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/133.0.0.0 Safari/537.36" } try: response = requests.get(url, headers=headers, timeout=15) response.raise_for_status() soup = BeautifulSoup(response.text, "html.parser") company_data: Dict[str, str] = {} # Parse all tables on the page for table in soup.find_all("table"): for row in table.find_all("tr"): cells = row.find_all(["td", "th"]) if len(cells) >= 3: field_name = cells[0].get_text(strip=True) value_cell = cells[2] content = value_cell.get_text(strip=True) if not content and value_cell.find("a"): content = value_cell.find("a").get_text(strip=True) if field_name: english_key = translate_key(field_name) company_data[english_key] = content # Add metadata company_data["stock_code"] = stock_code company_data["source"] = "idx.co.id" company_data["scraped_at"] = "live" return { "success": True, "data": company_data, "message": f"Profile of {stock_code} successfully retrieved from IDX" } except requests.exceptions.RequestException as e: return { "success": False, "data": {}, "message": f"Network error: {str(e)}" } except Exception as e: return { "success": False, "data": {}, "message": f"Error scraping: {str(e)} (possible page structure changed)" } # For local testing or Clawhub preview if __name__ == "__main__": result = idx_stock_profile("BBCA") print(result)
{ "stock_code": "BBCA" } Example Output { "success": true, "data": { "name": "PT Bank Central Asia Tbk.", "code": "BBCA", "office_address": "Menara BCA, Grand Indonesia...", "sector": "Finance", "stock_code": "BBCA", "source": "idx.co.id", "scraped_at": "live" }, "message": "Profile of BBCA successfully retrieved from IDX" } Copy the entire content above and save as idx-stock.md for direct import into Clawhub. The skill is ready to deploy and fully compliant with all platform security requirements. Created by Username: @zororaka00 Name: "Web3 Hungry" X Account: https://x.com/web3hungry
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.