# Send Senior Ml Engineer to your agent
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
## Fast path
- Download the package from Yavira.
- Extract it into a folder your agent can access.
- Paste one of the prompts below and point your agent at the extracted folder.
## Suggested prompts
### New install

```text
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
```
### Upgrade existing

```text
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
```
## Machine-readable fields
```json
{
  "schemaVersion": "1.0",
  "item": {
    "slug": "senior-ml-engineer",
    "name": "Senior Ml Engineer",
    "source": "tencent",
    "type": "skill",
    "category": "AI 智能",
    "sourceUrl": "https://clawhub.ai/alirezarezvani/senior-ml-engineer",
    "canonicalUrl": "https://clawhub.ai/alirezarezvani/senior-ml-engineer",
    "targetPlatform": "OpenClaw"
  },
  "install": {
    "downloadUrl": "/downloads/senior-ml-engineer",
    "sourceDownloadUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=senior-ml-engineer",
    "sourcePlatform": "tencent",
    "targetPlatform": "OpenClaw",
    "packageFormat": "ZIP package",
    "primaryDoc": "SKILL.md",
    "includedAssets": [
      "SKILL.md",
      "references/llm_integration_guide.md",
      "references/mlops_production_patterns.md",
      "references/rag_system_architecture.md",
      "scripts/ml_monitoring_suite.py",
      "scripts/model_deployment_pipeline.py"
    ],
    "downloadMode": "redirect",
    "sourceHealth": {
      "source": "tencent",
      "slug": "senior-ml-engineer",
      "status": "healthy",
      "reason": "direct_download_ok",
      "recommendedAction": "download",
      "checkedAt": "2026-05-03T15:25:46.781Z",
      "expiresAt": "2026-05-10T15:25:46.781Z",
      "httpStatus": 200,
      "finalUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=senior-ml-engineer",
      "contentType": "application/zip",
      "probeMethod": "head",
      "details": {
        "probeUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=senior-ml-engineer",
        "contentDisposition": "attachment; filename=\"senior-ml-engineer-2.1.1.zip\"",
        "redirectLocation": null,
        "bodySnippet": null,
        "slug": "senior-ml-engineer"
      },
      "scope": "item",
      "summary": "Item download looks usable.",
      "detail": "Yavira can redirect you to the upstream package for this item.",
      "primaryActionLabel": "Download for OpenClaw",
      "primaryActionHref": "/downloads/senior-ml-engineer"
    },
    "validation": {
      "installChecklist": [
        "Use the Yavira download entry.",
        "Review SKILL.md after the package is downloaded.",
        "Confirm the extracted package contains the expected setup assets."
      ],
      "postInstallChecks": [
        "Confirm the extracted package includes the expected docs or setup files.",
        "Validate the skill or prompts are available in your target agent workspace.",
        "Capture any manual follow-up steps the agent could not complete."
      ]
    }
  },
  "links": {
    "detailUrl": "https://openagent3.xyz/skills/senior-ml-engineer",
    "downloadUrl": "https://openagent3.xyz/downloads/senior-ml-engineer",
    "agentUrl": "https://openagent3.xyz/skills/senior-ml-engineer/agent",
    "manifestUrl": "https://openagent3.xyz/skills/senior-ml-engineer/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/senior-ml-engineer/agent.md"
  }
}
```
## Documentation

### Senior ML Engineer

Production ML engineering patterns for model deployment, MLOps infrastructure, and LLM integration.

### Table of Contents

Model Deployment Workflow
MLOps Pipeline Setup
LLM Integration Workflow
RAG System Implementation
Model Monitoring
Reference Documentation
Tools

### Model Deployment Workflow

Deploy a trained model to production with monitoring:

Export model to standardized format (ONNX, TorchScript, SavedModel)
Package model with dependencies in Docker container
Deploy to staging environment
Run integration tests against staging
Deploy canary (5% traffic) to production
Monitor latency and error rates for 1 hour
Promote to full production if metrics pass
Validation: p95 latency < 100ms, error rate < 0.1%

### Container Template

FROM python:3.11-slim

COPY requirements.txt .
RUN pip install --no-cache-dir -r requirements.txt

COPY model/ /app/model/
COPY src/ /app/src/

HEALTHCHECK CMD curl -f http://localhost:8080/health || exit 1

EXPOSE 8080
CMD ["uvicorn", "src.server:app", "--host", "0.0.0.0", "--port", "8080"]

### Serving Options

OptionLatencyThroughputUse CaseFastAPI + UvicornLowMediumREST APIs, small modelsTriton Inference ServerVery LowVery HighGPU inference, batchingTensorFlow ServingLowHighTensorFlow modelsTorchServeLowHighPyTorch modelsRay ServeMediumHighComplex pipelines, multi-model

### MLOps Pipeline Setup

Establish automated training and deployment:

Configure feature store (Feast, Tecton) for training data
Set up experiment tracking (MLflow, Weights & Biases)
Create training pipeline with hyperparameter logging
Register model in model registry with version metadata
Configure staging deployment triggered by registry events
Set up A/B testing infrastructure for model comparison
Enable drift monitoring with alerting
Validation: New models automatically evaluated against baseline

### Feature Store Pattern

from feast import Entity, Feature, FeatureView, FileSource

user = Entity(name="user_id", value_type=ValueType.INT64)

user_features = FeatureView(
    name="user_features",
    entities=["user_id"],
    ttl=timedelta(days=1),
    features=[
        Feature(name="purchase_count_30d", dtype=ValueType.INT64),
        Feature(name="avg_order_value", dtype=ValueType.FLOAT),
    ],
    online=True,
    source=FileSource(path="data/user_features.parquet"),
)

### Retraining Triggers

TriggerDetectionActionScheduledCron (weekly/monthly)Full retrainPerformance dropAccuracy < thresholdImmediate retrainData driftPSI > 0.2Evaluate, then retrainNew data volumeX new samplesIncremental update

### LLM Integration Workflow

Integrate LLM APIs into production applications:

Create provider abstraction layer for vendor flexibility
Implement retry logic with exponential backoff
Configure fallback to secondary provider
Set up token counting and context truncation
Add response caching for repeated queries
Implement cost tracking per request
Add structured output validation with Pydantic
Validation: Response parses correctly, cost within budget

### Provider Abstraction

from abc import ABC, abstractmethod
from tenacity import retry, stop_after_attempt, wait_exponential

class LLMProvider(ABC):
    @abstractmethod
    def complete(self, prompt: str, **kwargs) -> str:
        pass

@retry(stop=stop_after_attempt(3), wait=wait_exponential(min=1, max=10))
def call_llm_with_retry(provider: LLMProvider, prompt: str) -> str:
    return provider.complete(prompt)

### Cost Management

ProviderInput CostOutput CostGPT-4$0.03/1K$0.06/1KGPT-3.5$0.0005/1K$0.0015/1KClaude 3 Opus$0.015/1K$0.075/1KClaude 3 Haiku$0.00025/1K$0.00125/1K

### RAG System Implementation

Build retrieval-augmented generation pipeline:

Choose vector database (Pinecone, Qdrant, Weaviate)
Select embedding model based on quality/cost tradeoff
Implement document chunking strategy
Create ingestion pipeline with metadata extraction
Build retrieval with query embedding
Add reranking for relevance improvement
Format context and send to LLM
Validation: Response references retrieved context, no hallucinations

### Vector Database Selection

DatabaseHostingScaleLatencyBest ForPineconeManagedHighLowProduction, managedQdrantBothHighVery LowPerformance-criticalWeaviateBothHighLowHybrid searchChromaSelf-hostedMediumLowPrototypingpgvectorSelf-hostedMediumMediumExisting Postgres

### Chunking Strategies

StrategyChunk SizeOverlapBest ForFixed500-1000 tokens50-100General textSentence3-5 sentences1 sentenceStructured textSemanticVariableBased on meaningResearch papersRecursiveHierarchicalParent-childLong documents

### Model Monitoring

Monitor production models for drift and degradation:

Set up latency tracking (p50, p95, p99)
Configure error rate alerting
Implement input data drift detection
Track prediction distribution shifts
Log ground truth when available
Compare model versions with A/B metrics
Set up automated retraining triggers
Validation: Alerts fire before user-visible degradation

### Drift Detection

from scipy.stats import ks_2samp

def detect_drift(reference, current, threshold=0.05):
    statistic, p_value = ks_2samp(reference, current)
    return {
        "drift_detected": p_value < threshold,
        "ks_statistic": statistic,
        "p_value": p_value
    }

### Alert Thresholds

MetricWarningCriticalp95 latency> 100ms> 200msError rate> 0.1%> 1%PSI (drift)> 0.1> 0.2Accuracy drop> 2%> 5%

### MLOps Production Patterns

references/mlops_production_patterns.md contains:

Model deployment pipeline with Kubernetes manifests
Feature store architecture with Feast examples
Model monitoring with drift detection code
A/B testing infrastructure with traffic splitting
Automated retraining pipeline with MLflow

### LLM Integration Guide

references/llm_integration_guide.md contains:

Provider abstraction layer pattern
Retry and fallback strategies with tenacity
Prompt engineering templates (few-shot, CoT)
Token optimization with tiktoken
Cost calculation and tracking

### RAG System Architecture

references/rag_system_architecture.md contains:

RAG pipeline implementation with code
Vector database comparison and integration
Chunking strategies (fixed, semantic, recursive)
Embedding model selection guide
Hybrid search and reranking patterns

### Model Deployment Pipeline

python scripts/model_deployment_pipeline.py --model model.pkl --target staging

Generates deployment artifacts: Dockerfile, Kubernetes manifests, health checks.

### RAG System Builder

python scripts/rag_system_builder.py --config rag_config.yaml --analyze

Scaffolds RAG pipeline with vector store integration and retrieval logic.

### ML Monitoring Suite

python scripts/ml_monitoring_suite.py --config monitoring.yaml --deploy

Sets up drift detection, alerting, and performance dashboards.

### Tech Stack

CategoryToolsML FrameworksPyTorch, TensorFlow, Scikit-learn, XGBoostLLM FrameworksLangChain, LlamaIndex, DSPyMLOpsMLflow, Weights & Biases, KubeflowDataSpark, Airflow, dbt, KafkaDeploymentDocker, Kubernetes, TritonDatabasesPostgreSQL, BigQuery, Pinecone, Redis
## Trust
- Source: tencent
- Verification: Indexed source record
- Publisher: alirezarezvani
- Version: 2.1.1
## Source health
- Status: healthy
- Item download looks usable.
- Yavira can redirect you to the upstream package for this item.
- Health scope: item
- Reason: direct_download_ok
- Checked at: 2026-05-03T15:25:46.781Z
- Expires at: 2026-05-10T15:25:46.781Z
- Recommended action: Download for OpenClaw
## Links
- [Detail page](https://openagent3.xyz/skills/senior-ml-engineer)
- [Send to Agent page](https://openagent3.xyz/skills/senior-ml-engineer/agent)
- [JSON manifest](https://openagent3.xyz/skills/senior-ml-engineer/agent.json)
- [Markdown brief](https://openagent3.xyz/skills/senior-ml-engineer/agent.md)
- [Download page](https://openagent3.xyz/downloads/senior-ml-engineer)