Play 13
Fine-Tuning Workflow
High🔧 Skeleton
End-to-end fine-tuning with data prep, LoRA training, evaluation, and deployment.
Curate training data, configure LoRA parameters, train on Azure ML with GPU compute, evaluate with automated metrics, then deploy the fine-tuned model. MLflow tracks experiments. The pipeline handles data validation, train/val splitting, hyperparameter sweeps, and model versioning.
Architecture Pattern
LoRA fine-tuning, dataset curation, evaluation, MLOps
Azure Services
Azure ML WorkspaceGPU ComputeStorageMLflowAzure OpenAI (base models)
DevKit (.github Agentic OS)
- agent.md — ML engineer persona
- instructions.md — training protocols
- mcp/index.js — training validation tools
- plugins/ — data prep, trainer, evaluator
TuneKit (AI Config)
- config/training.json — LoRA rank, learning rate, epochs, batch size
- config/dataset.json — train/val split, preprocessing
- config/evaluation.json — eval metrics, thresholds
- evaluation/eval.py — automated scoring
Tuning Parameters
LoRA rank (8–64)Learning rateEpochsBatch sizeEval metrics thresholds
Estimated Cost
Dev/Test
$200–400/mo
Production
$1.5K–5K/mo (training)