Agent skill
fine-tuning
LLM fine-tuning and prompt-tuning techniques
Stars
163
Forks
31
Install this agent skill to your Project
npx add-skill https://github.com/majiayu000/claude-skill-registry/tree/main/skills/development/fine-tuning
SKILL.md
Fine-Tuning Skill
Bonded to: prompt-optimization-agent
Quick Start
bash
Skill("custom-plugin-prompt-engineering:fine-tuning")
Parameter Schema
yaml
parameters:
tuning_method:
type: enum
values: [full, lora, qlora, prompt_tuning, prefix_tuning]
default: lora
dataset_size:
type: enum
values: [small, medium, large]
description: "<1k, 1k-10k, >10k examples"
compute_budget:
type: enum
values: [low, medium, high]
default: medium
Tuning Methods Comparison
| Method | Parameters | Compute | Quality | Best For |
|---|---|---|---|---|
| Full Fine-tune | All | Very High | Highest | Maximum customization |
| LoRA | ~0.1% | Low | High | Resource-constrained |
| QLoRA | ~0.1% | Very Low | Good | Consumer GPUs |
| Prompt Tuning | <0.01% | Minimal | Good | Simple tasks |
| Prefix Tuning | ~0.1% | Low | Good | Generation tasks |
Dataset Preparation
Format Templates
yaml
formats:
instruction:
template: |
### Instruction
{instruction}
### Response
{response}
chat:
template: |
<|user|>
{user_message}
<|assistant|>
{assistant_response}
completion:
template: "{input}{output}"
Quality Criteria
yaml
quality_checklist:
- [ ] No duplicate examples
- [ ] Consistent formatting
- [ ] Diverse examples
- [ ] Balanced categories
- [ ] High-quality outputs
- [ ] No harmful content
Training Configuration
yaml
training_config:
hyperparameters:
learning_rate: 2e-5
batch_size: 8
epochs: 3
warmup_ratio: 0.1
lora_config:
r: 16
alpha: 32
dropout: 0.05
target_modules: ["q_proj", "v_proj"]
evaluation:
eval_steps: 100
save_steps: 500
metric: loss
Evaluation Framework
| Metric | Purpose | Target |
|---|---|---|
| Loss | Training progress | Decreasing |
| Accuracy | Task performance | >90% |
| Perplexity | Model confidence | <10 |
| Human eval | Quality assessment | Preferred >80% |
Troubleshooting
| Issue | Cause | Solution |
|---|---|---|
| Overfitting | Small dataset | Add regularization |
| Underfitting | Low epochs | Increase training |
| Catastrophic forgetting | Aggressive tuning | Lower learning rate |
| Poor generalization | Data bias | Diversify dataset |
References
See: Hugging Face PEFT, OpenAI Fine-tuning Guide
Didn't find tool you were looking for?