PinnedI truly appreciate your kind words, It’s an honor to be acknowledged by the author, and I’m really…Jan 31Jan 31
PinnedThanks for the appreciation, Its surreal for me to get acknowledged from the author itself.Feb 1, 2024Feb 1, 2024
Papers Explained 365: DeepMathDeepMath-103K is a new dataset designed for advancing mathematical reasoning research. It comprises 103,000 mathematical problems with a…14h ago14h ago
Papers Explained 364: OmniMathOmniMath is a comprehensive and challenging benchmark specifically designed to assess LLMs’ mathematical reasoning at the Olympiad level…1d ago1d ago
Papers Explained 363: UltraLongThis work introduces an efficient training recipe for building ultra-long context LLMs from aligned instruct models, pushing the boundaries…2d ago2d ago
Papers Explained 362: Llama-NemotronLlama-Nemotron is an open family of heterogeneous reasoning models available in Nano (8B), Super (49B), and Ultra (253B) sizes, designed…5d ago5d ago
Papers Explained 361: OpenCodeReasoningOpenCodeReasoning is a publicly available synthetic dataset for code reasoning, comprising 736,712 Python code solutions with accompanying…6d ago6d ago
Papers Explained 360: Nemotron CrossThinkNemotron-Crossthink is a framework that systematically incorporates multi-domain corpora, including both synthetic and real-world…May 7May 7
Papers Explained 359: Phi-4-Mini-ReasoningThis paper presents a systematic training recipe for SLMs that consists of four steps:May 6May 6
Papers Explained 358: Phi-4-ReasoningPhi-4-reasoning is a 14-billion parameter reasoning model that achieves strong performance on complex reasoning tasks. It is trained via…May 5May 5