← All skills
Tencent SkillHub Β· AI

Senior Ml Engineer

ML engineering skill for productionizing models, building MLOps pipelines, and integrating LLMs. Covers model deployment, feature stores, drift monitoring, RAG systems, and cost optimization.

skill openclawclawhub Free
0 Downloads
0 Stars
0 Installs
0 Score
High Signal

ML engineering skill for productionizing models, building MLOps pipelines, and integrating LLMs. Covers model deployment, feature stores, drift monitoring, RAG systems, and cost optimization.

⬇ 0 downloads β˜… 0 stars Unverified but indexed

Install for OpenClaw

Quick setup
  1. Download the package from Yavira.
  2. Extract the archive and review SKILL.md first.
  3. Import or place the package into your OpenClaw setup.

Requirements

Target platform
OpenClaw
Install method
Manual import
Extraction
Extract archive
Prerequisites
OpenClaw
Primary doc
SKILL.md

Package facts

Download mode
Yavira redirect
Package format
ZIP package
Source platform
Tencent SkillHub
What's included
SKILL.md, references/llm_integration_guide.md, references/mlops_production_patterns.md, references/rag_system_architecture.md, scripts/ml_monitoring_suite.py, scripts/model_deployment_pipeline.py

Validation

  • Use the Yavira download entry.
  • Review SKILL.md after the package is downloaded.
  • Confirm the extracted package contains the expected setup assets.

Install with your agent

Agent handoff

Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.

  1. Download the package from Yavira.
  2. Extract it into a folder your agent can access.
  3. Paste one of the prompts below and point your agent at the extracted folder.
New install

I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.

Upgrade existing

I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.

Trust & source

Release facts

Source
Tencent SkillHub
Verification
Indexed source record
Version
2.1.1

Documentation

ClawHub primary doc Primary doc: SKILL.md 24 sections Open source page

Senior ML Engineer

Production ML engineering patterns for model deployment, MLOps infrastructure, and LLM integration.

Table of Contents

Model Deployment Workflow MLOps Pipeline Setup LLM Integration Workflow RAG System Implementation Model Monitoring Reference Documentation Tools

Model Deployment Workflow

Deploy a trained model to production with monitoring: Export model to standardized format (ONNX, TorchScript, SavedModel) Package model with dependencies in Docker container Deploy to staging environment Run integration tests against staging Deploy canary (5% traffic) to production Monitor latency and error rates for 1 hour Promote to full production if metrics pass Validation: p95 latency < 100ms, error rate < 0.1%

Container Template

FROM python:3.11-slim COPY requirements.txt . RUN pip install --no-cache-dir -r requirements.txt COPY model/ /app/model/ COPY src/ /app/src/ HEALTHCHECK CMD curl -f http://localhost:8080/health || exit 1 EXPOSE 8080 CMD ["uvicorn", "src.server:app", "--host", "0.0.0.0", "--port", "8080"]

Serving Options

OptionLatencyThroughputUse CaseFastAPI + UvicornLowMediumREST APIs, small modelsTriton Inference ServerVery LowVery HighGPU inference, batchingTensorFlow ServingLowHighTensorFlow modelsTorchServeLowHighPyTorch modelsRay ServeMediumHighComplex pipelines, multi-model

MLOps Pipeline Setup

Establish automated training and deployment: Configure feature store (Feast, Tecton) for training data Set up experiment tracking (MLflow, Weights & Biases) Create training pipeline with hyperparameter logging Register model in model registry with version metadata Configure staging deployment triggered by registry events Set up A/B testing infrastructure for model comparison Enable drift monitoring with alerting Validation: New models automatically evaluated against baseline

Feature Store Pattern

from feast import Entity, Feature, FeatureView, FileSource user = Entity(name="user_id", value_type=ValueType.INT64) user_features = FeatureView( name="user_features", entities=["user_id"], ttl=timedelta(days=1), features=[ Feature(name="purchase_count_30d", dtype=ValueType.INT64), Feature(name="avg_order_value", dtype=ValueType.FLOAT), ], online=True, source=FileSource(path="data/user_features.parquet"), )

Retraining Triggers

TriggerDetectionActionScheduledCron (weekly/monthly)Full retrainPerformance dropAccuracy < thresholdImmediate retrainData driftPSI > 0.2Evaluate, then retrainNew data volumeX new samplesIncremental update

LLM Integration Workflow

Integrate LLM APIs into production applications: Create provider abstraction layer for vendor flexibility Implement retry logic with exponential backoff Configure fallback to secondary provider Set up token counting and context truncation Add response caching for repeated queries Implement cost tracking per request Add structured output validation with Pydantic Validation: Response parses correctly, cost within budget

Provider Abstraction

from abc import ABC, abstractmethod from tenacity import retry, stop_after_attempt, wait_exponential class LLMProvider(ABC): @abstractmethod def complete(self, prompt: str, **kwargs) -> str: pass @retry(stop=stop_after_attempt(3), wait=wait_exponential(min=1, max=10)) def call_llm_with_retry(provider: LLMProvider, prompt: str) -> str: return provider.complete(prompt)

Cost Management

ProviderInput CostOutput CostGPT-4$0.03/1K$0.06/1KGPT-3.5$0.0005/1K$0.0015/1KClaude 3 Opus$0.015/1K$0.075/1KClaude 3 Haiku$0.00025/1K$0.00125/1K

RAG System Implementation

Build retrieval-augmented generation pipeline: Choose vector database (Pinecone, Qdrant, Weaviate) Select embedding model based on quality/cost tradeoff Implement document chunking strategy Create ingestion pipeline with metadata extraction Build retrieval with query embedding Add reranking for relevance improvement Format context and send to LLM Validation: Response references retrieved context, no hallucinations

Vector Database Selection

DatabaseHostingScaleLatencyBest ForPineconeManagedHighLowProduction, managedQdrantBothHighVery LowPerformance-criticalWeaviateBothHighLowHybrid searchChromaSelf-hostedMediumLowPrototypingpgvectorSelf-hostedMediumMediumExisting Postgres

Chunking Strategies

StrategyChunk SizeOverlapBest ForFixed500-1000 tokens50-100General textSentence3-5 sentences1 sentenceStructured textSemanticVariableBased on meaningResearch papersRecursiveHierarchicalParent-childLong documents

Model Monitoring

Monitor production models for drift and degradation: Set up latency tracking (p50, p95, p99) Configure error rate alerting Implement input data drift detection Track prediction distribution shifts Log ground truth when available Compare model versions with A/B metrics Set up automated retraining triggers Validation: Alerts fire before user-visible degradation

Drift Detection

from scipy.stats import ks_2samp def detect_drift(reference, current, threshold=0.05): statistic, p_value = ks_2samp(reference, current) return { "drift_detected": p_value < threshold, "ks_statistic": statistic, "p_value": p_value }

Alert Thresholds

MetricWarningCriticalp95 latency> 100ms> 200msError rate> 0.1%> 1%PSI (drift)> 0.1> 0.2Accuracy drop> 2%> 5%

MLOps Production Patterns

references/mlops_production_patterns.md contains: Model deployment pipeline with Kubernetes manifests Feature store architecture with Feast examples Model monitoring with drift detection code A/B testing infrastructure with traffic splitting Automated retraining pipeline with MLflow

LLM Integration Guide

references/llm_integration_guide.md contains: Provider abstraction layer pattern Retry and fallback strategies with tenacity Prompt engineering templates (few-shot, CoT) Token optimization with tiktoken Cost calculation and tracking

RAG System Architecture

references/rag_system_architecture.md contains: RAG pipeline implementation with code Vector database comparison and integration Chunking strategies (fixed, semantic, recursive) Embedding model selection guide Hybrid search and reranking patterns

Model Deployment Pipeline

python scripts/model_deployment_pipeline.py --model model.pkl --target staging Generates deployment artifacts: Dockerfile, Kubernetes manifests, health checks.

RAG System Builder

python scripts/rag_system_builder.py --config rag_config.yaml --analyze Scaffolds RAG pipeline with vector store integration and retrieval logic.

ML Monitoring Suite

python scripts/ml_monitoring_suite.py --config monitoring.yaml --deploy Sets up drift detection, alerting, and performance dashboards.

Tech Stack

CategoryToolsML FrameworksPyTorch, TensorFlow, Scikit-learn, XGBoostLLM FrameworksLangChain, LlamaIndex, DSPyMLOpsMLflow, Weights & Biases, KubeflowDataSpark, Airflow, dbt, KafkaDeploymentDocker, Kubernetes, TritonDatabasesPostgreSQL, BigQuery, Pinecone, Redis

Category context

Agent frameworks, memory systems, reasoning layers, and model-native orchestration.

Source: Tencent SkillHub

Largest current source with strong distribution and engagement signals.

Package contents

Included in package
4 Docs2 Scripts
  • SKILL.md Primary doc
  • references/llm_integration_guide.md Docs
  • references/mlops_production_patterns.md Docs
  • references/rag_system_architecture.md Docs
  • scripts/ml_monitoring_suite.py Scripts
  • scripts/model_deployment_pipeline.py Scripts