faion-ml-ops

ML operations: fine-tuning (LoRA, QLoRA), model evaluation, cost optimization, observability.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "faion-ml-ops" with this command: npx skills add faionfaion/faion-network/faionfaion-faion-network-faion-ml-ops

Entry point: /faion-net — invoke this skill for automatic routing to the appropriate domain.

ML Ops Skill

Communication: User's language. Code: English.

Purpose

Handles ML model operations. Covers fine-tuning, evaluation, cost management, and observability.

Context Discovery

Auto-Investigation

Check these project signals before asking questions:

SignalWhere to CheckWhat to Look For
Dependenciesrequirements.txttransformers, peft, openai, tiktoken, langsmith
Training data/data, /datasetsJSONL files for fine-tuning
Logs/metricsGrep for "langsmith", "wandb", "mlflow"Existing observability tools
Cost trackingGrep for "tiktoken", "count_tokens"Token counting implementation

Discovery Questions

question: "What ML operation are you working on?"
header: "Operation Type"
multiSelect: false
options:
  - label: "Fine-tuning LLM"
    description: "Custom model training (OpenAI API, LoRA, QLoRA)"
  - label: "Model evaluation"
    description: "Benchmark performance, LLM-as-judge"
  - label: "Cost optimization"
    description: "Reduce API costs, prompt caching, batching"
  - label: "Observability/monitoring"
    description: "Track LLM usage, traces, performance"
question: "For fine-tuning: dataset size and approach?"
header: "Fine-tuning Strategy"
multiSelect: false
options:
  - label: "<100 examples - use few-shot prompting instead"
    description: "Too small for fine-tuning, improve prompts"
  - label: "100-1000 examples - OpenAI fine-tuning"
    description: "Use OpenAI API fine-tuning endpoint"
  - label: ">1000 examples - LoRA/QLoRA"
    description: "Efficient parameter fine-tuning"
  - label: "Not fine-tuning"
    description: "Skip this question"
question: "Which observability tools?"
header: "Monitoring Stack"
multiSelect: true
options:
  - label: "LangSmith (recommended)"
    description: "LangChain native tracing"
  - label: "Langfuse (open-source)"
    description: "Self-hosted observability"
  - label: "Custom logging"
    description: "Build custom tracking"
  - label: "None yet"
    description: "Starting from scratch"

Scope

AreaCoverage
Fine-tuningLoRA, QLoRA, OpenAI fine-tuning, datasets
EvaluationMetrics, benchmarks, frameworks
Cost OptimizationToken management, caching, batch APIs
ObservabilityLLM monitoring, tracing, logging

Quick Start

TaskFiles
Fine-tune OpenAIfine-tuning-openai-basics.md → fine-tuning-openai-production.md
Fine-tune LoRAlora-qlora.md → finetuning-basics.md
Cost optimizationllm-cost-basics.md → cost-reduction-strategies.md
Evaluationevaluation-metrics.md → evaluation-framework.md
Observabilityllm-observability.md → llm-observability-stack-2026.md

Methodologies (15)

Fine-tuning (5):

  • finetuning-basics: Fundamentals, when to fine-tune
  • finetuning-datasets: Data preparation, quality
  • fine-tuning-openai-basics: OpenAI API fine-tuning
  • fine-tuning-openai-production: Production deployment
  • lora-qlora: Efficient fine-tuning, parameter selection

Evaluation (3):

  • evaluation-metrics: Accuracy, F1, perplexity, task metrics
  • evaluation-framework: LLM-as-judge, human eval
  • evaluation-benchmarks: MMLU, HumanEval, industry benchmarks

Cost Optimization (2):

  • llm-cost-basics: Token counting, pricing models
  • cost-reduction-strategies: Caching, compression, batching

Observability (5):

  • llm-observability: Fundamentals, why monitor
  • llm-observability-stack: Tools selection
  • llm-observability-stack-2026: Latest tools (LangSmith, Langfuse)
  • llm-management-observability: End-to-end management

Code Examples

OpenAI Fine-tuning

from openai import OpenAI

client = OpenAI()

# Upload training data
file = client.files.create(
    file=open("training_data.jsonl", "rb"),
    purpose="fine-tune"
)

# Create fine-tuning job
job = client.fine_tuning.jobs.create(
    training_file=file.id,
    model="gpt-4o-mini-2024-07-18",
    hyperparameters={"n_epochs": 3}
)

# Monitor
while True:
    job = client.fine_tuning.jobs.retrieve(job.id)
    if job.status == "succeeded":
        break

LoRA Fine-tuning

from peft import LoraConfig, get_peft_model
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3-8b")

lora_config = LoraConfig(
    r=16,
    lora_alpha=32,
    target_modules=["q_proj", "v_proj"],
    lora_dropout=0.1,
    bias="none"
)

model = get_peft_model(model, lora_config)

Cost Tracking

import tiktoken

def count_tokens(text, model="gpt-4o"):
    encoding = tiktoken.encoding_for_model(model)
    return len(encoding.encode(text))

def estimate_cost(prompt, completion, model="gpt-4o"):
    prompt_tokens = count_tokens(prompt, model)
    completion_tokens = count_tokens(completion, model)

    # GPT-4o pricing
    prompt_cost = prompt_tokens * 0.000005
    completion_cost = completion_tokens * 0.000015

    return prompt_cost + completion_cost

LLM Observability with LangSmith

from langsmith import traceable

@traceable
def rag_pipeline(query: str) -> str:
    # Retrieval
    docs = retrieve(query)

    # Generation
    response = generate(query, docs)

    return response

Fine-tuning Decision Matrix

ScenarioApproach
Small dataset (<100 examples)Few-shot prompting
Medium dataset (100-1000)OpenAI fine-tuning
Large dataset (>1000)LoRA/QLoRA
Custom behaviorFine-tuning
New knowledgeRAG (not fine-tuning)

Cost Reduction Strategies

StrategySavingsTrade-off
Prompt caching90% on cachedCold start cost
Batch API50%24h latency
Smaller models80%+Lower quality
Context pruningVariableMay lose context
Output limitsVariableTruncated responses

Evaluation Frameworks

FrameworkUse Case
LangSmithProduction monitoring, traces
LangfuseOpen-source observability
PromptLayerPrompt versioning
Weights & BiasesExperiment tracking

Related Skills

SkillRelationship
faion-llm-integrationProvides APIs to optimize
faion-rag-engineerRAG evaluation
faion-devops-engineerModel deployment

ML Ops v1.0 | 15 methodologies

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

faion-automation-tooling

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

faion-ai-agents

No summary provided by upstream source.

Repository SourceNeeds Review
General

faion-ppc-manager

No summary provided by upstream source.

Repository SourceNeeds Review