Environment-Driven Improvement
+46% Performance Gain
See how fine-tuning Qwen2.5-Coder-7B on CodeBlue's analytics tasks dramatically improves performance across all difficulty levels.
Correctness
Baseline
42.0%
69.0%
Fine-tuned
71.0%
Efficiency
Baseline
65.0%
26.2%
Fine-tuned
82.0%
Avg Turns
Baseline
5.2
40.4%
Fine-tuned
3.1
Success Rate
Baseline
38.0%
76.3%
Fine-tuned
67.0%
Performance Comparison
Model Details
Base ModelQwen2.5-Coder-7B
Parameters7B
ArchitectureTransformer
Training Config
Dataset2,847 examples
Epochs3
Learning Rate2e-5
Batch Size8
Training Time4.2 hours
HardwareA100 80GB
Biggest Gains
L6 (Hard)+150%
L5 (Medium+)+107%
L4 (Medium)+90%
Want these gains for your model?
Our environment and dataset are open-source. Fine-tune your own model on CodeBlue tasks.