Papers Explained 306: Critique Fine-Tuning

Ritvik Rastogi
4 min readFeb 10, 2025

--

Critique Fine-Tuning (CFT) is a strategy where models learn to critique noisy responses rather than simply imitate correct ones. Inspired by human learning processes that emphasize critical thinking, CFT encourages deeper analysis and nuanced understanding. It challenges the commonly used Supervised Fine-Tuning (SFT) paradigm,which imitates annotated responses for given instructions.

The project is available on GitHub.

Method & Dataset

Comparison between CFT and SFT.

Most experiments are based on WebInstruct, an instruction dataset collected from online educational resources and quiz websites. WebInstruct spans a wide range of topics, including Mathematics (65%), Physics (8%), Chemistry (4%), Business (10%), Humanities (4%), and more. The responses in WebInstruct are extracted and refined by large language models such as Qwen-72B and Mixtral, making them highly prone to noise due to the lack of verification or quality control.

The following subsets are curated from WebInstruct:

  • WebInstruct-SFT: A 50K subset directly sampled from the original WebInstruct dataset. This subset has a very high error ratio (over 50%).
  • WebInstruct-verified: Samples from WebInstruct are adopted and GPT-4o-1120 is prompted to judge whether the original answers are correct or not. The top 50K samples are retained as ”verified” SFT data.
  • WebInstruct-GPT-4o: A 50K subset that reuses questions from WebInstruct-SFT but replaces the answers with those generated by GPT-4o-1120.
  • WebInstruct-CFT: A 50K subset derived from WebInstruct-SFT, where GPT-4o-1120 provides detailed critiques of the original responses. Approximately 56% of the responses in this subset are judged as ”correct,” while the rest are considered ”wrong.” Despite containing some critique errors introduced by GPT-4o, this dataset is comparable in quality to WebInstruct-GPT-4o.
  • WebInstruct-CFT-Tiny: A smaller version of WebInstruct-CFT, containing only 4K examples, designed for training our 32B model.
Comparison of CFT datasets with existing SFT datasets.

Training Objective

The training objective of the approach is straightforward. Question x and noisy response y are concatenated as input, and the model parameters are optimized to generate the critique c. Formally, the training loss is:

Experiment Setup

Three different SFT settings and one CFT setting are evaluated in the experiments.

  1. SFT: directly training on original noisy responses
  2. SFT-verified: training on responses validated by GPT-4o
  3. SFT-gpt4o: training on responses generated by GPT-4o. For CFT, the model is trained using curated CFT datasets.

MATH-500 is used as the validation set and the best-performing checkpoint after training on the entire dataset for 1 epoch is selected.

Evaluation

Performance comparison of SFT and CFT on different base models.
  • CFT consistently outperforms all SFT baselines across different base models on mathematical reasoning benchmarks.
  • For example, CFT achieved a 3.5% absolute improvement over SFT-GPT4o on DeepSeek-Math-7B.
  • CFT showed a substantial 10.4% improvement over SFT-verified on Qwen2.5–7B.
  • CFT surpassed SFT-GPT4o by 6.7% on Qwen2.5-Math-7B.
  • Qwen2.5-Math-7B is found to be a stronger base model, and when enhanced with CFT, it achieved the best performance among the tested 7B models.
Training dynamics comparison of different methods on Qwen2.5-Math-7B across MATH and Minerva-Math.
  • CFT demonstrates faster convergence and maintains higher performance throughout training compared to SFT variants.
  • CFT achieves approximately 80% accuracy on MATH and 40% on Minerva-Math, compared to lower accuracies for SFT-V and SFT-G.
Performance comparison of CFT models vs. other reasoning-specialized models.
  • The 7B CFT model (Qwen2.5-Math-7B-CFT) achieves the highest average performance (48.1%) among 7B-scale models with significantly less training data (50K samples).
  • It outperforms other specialized math models like Deepseek-Math-7B-Instruct, Mathstral-7B, and NuminaMath-7B-CoT.
  • The 7B CFT model shows competitive performance compared to much larger models like Qwen2.5-Math-72B-Instruct, despite having significantly fewer parameters and less training data.
Performance Comparison of 32B Models across Mathematical Reasoning Benchmarks.
  • The 32B CFT model (Qwen2.5–32B-Instruct-CFT) is highly data-efficient, achieving optimal performance with only 4K training samples, compared to 17K for Sky-T1–32B-Preview.
Performance comparison of CFT using different teacher critique models on Qwen2.5-Math-7B.
  • CFT is effective even with a modest critique model (GPT-4o-mini), and performance improves further with a stronger critique model (GPT-4o-1120).

Paper

Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate 2501.17703

Hungry for more insights?

Don’t miss out on exploring other fascinating threads in this series. Simply click here and uncover the state-of-the-art research!

Do Subscribe for weekly updates!!

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

--

--

Ritvik Rastogi
Ritvik Rastogi

Written by Ritvik Rastogi

Data Scientist, 2x Kaggle Expert

No responses yet

Write a response