CodeBlue

|Fine-tuning Gains
Back to Home
Environment-Driven Improvement

+46% Performance Gain

See how fine-tuning Qwen2.5-Coder-7B on CodeBlue's analytics tasks dramatically improves performance across all difficulty levels.

Correctness

Baseline

42.0%

69.0%

Fine-tuned

71.0%

Efficiency

Baseline

65.0%

26.2%

Fine-tuned

82.0%

Avg Turns

Baseline

5.2

40.4%

Fine-tuned

3.1

Success Rate

Baseline

38.0%

76.3%

Fine-tuned

67.0%

Performance Comparison

Model Details

Base ModelQwen2.5-Coder-7B
Parameters7B
ArchitectureTransformer

Training Config

Dataset2,847 examples
Epochs3
Learning Rate2e-5
Batch Size8
Training Time4.2 hours
HardwareA100 80GB

Biggest Gains

L6 (Hard)+150%
L5 (Medium+)+107%
L4 (Medium)+90%

Want these gains for your model?

Our environment and dataset are open-source. Fine-tune your own model on CodeBlue tasks.