Why use Fine-tuning
- Create domain experts: Train models on your specific data to excel at specialized tasks like legal analysis or medical coding
- Ensure consistent outputs: Get reliable formatting and structure that matches your exact specifications every time
- Reduce costs and latency: Smaller fine-tuned models often outperform larger generic models at specific tasks
Helicone's dataset curation interface for preparing fine-tuning data
Quick Start
1
Curate your dataset
Filter and select high-quality request-response pairs from your Helicone logs:
Advanced filtering options
Advanced filtering options
Use Helicone’s powerful filtering to find the best training data:
- Score thresholds for quality control
- User feedback filters
- Token count ranges
- Custom property combinations
2
Export for fine-tuning
Export your curated dataset in the format your fine-tuning platform requires:
3
Create fine-tuning job
Use your exported dataset with your chosen platform:
Configuration Options
Basic Settings
Core options for dataset curation and export:| Setting | Type | Description | Default | Example |
|---|---|---|---|---|
minScore | number | Minimum score for including requests | 0 | 0.8 |
maxTokens | number | Maximum tokens per example | ∞ | 4096 |
format | string | Export format for platform | "openai" | "anthropic" |
includeMetadata | boolean | Include Helicone metadata | false | true |
Advanced Settings
| Setting | Type | Description | Default | Example |
|---|---|---|---|---|
validationSplit | number | Percentage for validation set | 0.2 | 0.15 |
deduplication | boolean | Remove duplicate examples | true | false |
sampling | object | Sampling configuration | {} | {"method": "random", "size": 1000} |
Detailed Explanations
Validation Split
Validation Split
Control how your dataset is divided for training and validation:Larger validation sets help detect overfitting but reduce training data.
Deduplication Strategies
Deduplication Strategies
Remove duplicate or near-duplicate examples to improve training:
Use Cases
- Customer Support Agent
- Code Generation Specialist
- Domain-Specific Expert
Fine-tune a model on your best support conversations for consistent, brand-aligned responses:
Understanding Fine-tuning
When to Use Fine-tuning vs RAG
Fine-tuning and RAG (Retrieval Augmented Generation) solve different problems: What Fine-tuning is best for:- Teaching consistent behavior and output formats
- Improving performance on specific tasks
- Reducing latency and costs with smaller models
- Encoding domain knowledge into the model
- Working with frequently changing information
- Handling large knowledge bases
- Maintaining source attribution
- Quick updates without retraining
Dataset Quality Guidelines
The quality of your fine-tuning dataset determines model performance: Key principles:- Diversity: Include varied examples covering edge cases
- Consistency: Ensure similar inputs have similar outputs
- Quality: Only include high-quality, verified examples
- Quantity: Start with 50-100 examples minimum
Fine-tuning Workflow Best Practices
Data preparation checklist:- Review examples for quality and consistency
- Remove PII and sensitive information
- Balance dataset across different use cases
- Validate format matches platform requirements