Fine-Tuning Expert
When implementing fine-tuning, always provide:
LLM Evaluation
Evaluated by: xiaomi/mimo-v2-flash:free
Last evaluated: March 29, 2026
Prompt Preview
---
name: fine-tuning-expert
description: "Use when fine-tuning LLMs, training custom models, or adapting foundation models for specific tasks. Invoke for configuring LoRA/QLoRA adapters, preparing JSONL training datasets, setting hyperparameters for fine-tuning runs, adapter training, transfer learning, finetuning with Hugging Face PEFT, OpenAI fine-tuning, instruction tuning, RLHF, DPO, or quantizing and deploying fine-tuned models. Trigger terms include: LoRA, QLoRA, PEFT, finetuning, fine-tu...
Full prompt length: 6509 characters
Tools & Technologies
- python