Skip to content

Top Stories

Top Stories

Primary Menu
  • Breaking News
  • UNIT CONVERTER
  • QR Code Generator
  • SEO META TAG GENERATOR
  • Background Remover Tool
  • Image Enhancer Tool
  • Image Converter Tool
  • Image Compressor Tool
  • Keyword Research Tool
  • Paint Tool
  • About Us
  • Contact Us
  • Privacy Policy
HOME PAGE
  • Home
  • Uncategorized
  • AbstRaL: Teaching LLMs Abstract Reasoning via Reinforcement to Boost Robustness on GSM Benchmarks
  • Uncategorized

AbstRaL: Teaching LLMs Abstract Reasoning via Reinforcement to Boost Robustness on GSM Benchmarks

VedVision HeadLines July 6, 2025
AbstRaL: Teaching LLMs Abstract Reasoning via Reinforcement to Boost Robustness on GSM Benchmarks


Recent research indicates that LLMs, particularly smaller ones, frequently struggle with robust reasoning. They tend to perform well on familiar questions but falter when those same problems are slightly altered, such as changing names or numbers, or adding irrelevant but related information. This weakness, known as poor out-of-distribution (OOD) generalization, results in notable accuracy drops, even in simple math tasks. One promising solution is to create synthetic variations of reasoning problems, helping models learn to focus on the underlying logic rather than surface details. Strengthening reasoning in this manner is crucial for developing more general and reliable AI systems.

Abstracting the Core Logic of LLM Reasoning Failures

LLMs have demonstrated impressive reasoning capabilities, yet they often falter when exposed to distribution shifts, such as changes in phrasing, numerical values, or the introduction of distractions. This vulnerability is evident across benchmarks in logic, mathematics, and commonsense reasoning. Prior solutions have relied on data augmentation to expose models to a broader variety of inputs, improving robustness but increasing computational demands. Researchers have also explored formats such as abstraction-of-thought and chain-of-abstraction to teach abstract reasoning, while planning techniques like chain-of-thought and tree-of-thought aid step-by-step problem-solving. Reinforcement learning and preference-based methods provide additional support for reasoning skill development beyond pattern memorization.

AbstRaL’s Symbolic Learning Method to Improve Reasoning Consistency

Researchers from Apple and EPFL propose AbstRaL, a method that teaches LLMs to understand abstract reasoning patterns rather than memorizing surface details. Instead of generating many varied training examples, which is computationally costly, AbstRaL helps LLMs learn the underlying structure of reasoning problems using reinforcement learning. This method connects these abstract patterns to symbolic tools, enabling more reliable problem-solving. Tested on GSM benchmarks, AbstRaL significantly improves LLM performance, especially when faced with input changes or distracting information. It outperforms models trained only with supervised learning by promoting more consistent and context-independent reasoning.

Four Steps to Abstract Symbolic Reasoning via AbstRaL

AbstRaL is a four-step framework designed to teach LLMs to reason abstractly rather than rely on surface patterns. First, it identifies key variables in a question and replaces them with symbolic placeholders. Then, using specially crafted data (GranulAR), the model learns to reason step-by-step with these abstract symbols. Next, it retrieves the general reasoning structure (abstraction) from the symbolic answer. Finally, it uses this abstraction with the original values to compute the correct answer. Reinforcement learning with two rewards, one for correctness and another for symbolic similarity, further improves the model’s ability to generate accurate, context-independent reasoning patterns.

GSM8K Variations Reveal AbstRaL’s Robustness Across LLM Sizes

The researchers evaluate AbstRaL on math reasoning tasks using models such as Llama-3 and Qwen2, training them with a dataset called GranulAR that rewrites math problems in an abstract symbolic form. This helps models focus on structure rather than surface details. They test robustness using altered versions of GSM8K problems, changing numbers, names, and phrasing. Compared to baselines like standard Chain-of-Thought prompting, AbstRaL shows stronger consistency and less accuracy drop on these variations. Especially for smaller models, it improves reliability across reworded inputs. The results suggest that teaching models to reason abstractly makes them more adaptable and less reliant on memorized patterns.

Teaching LLMs Abstract Thinking through Reinforcement Yields Robust Reasoning

In conclusion, AbstRaL is a method designed to enhance abstract reasoning in LLMs, making them more resilient to superficial changes in problems. Unlike traditional fine-tuning or data augmentation, AbstRaL uses reinforcement learning to train models on GranulAR rationales that mix Socratic chain-of-thought with detailed abstraction. This approach helps models strip away surface-level distractions and better connect with symbolic tools. Tested on challenging GSM8K perturbation benchmarks, AbstRaL notably reduces performance drops under distribution shifts, particularly in smaller models. The study shows that learning to abstract improves reasoning robustness more effectively than relying solely on direct supervision.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter, Youtube and Spotify and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter.


Sana Hassan, a consulting intern at Marktechpost and dual-degree student at IIT Madras, is passionate about applying technology and AI to address real-world challenges. With a keen interest in solving practical problems, he brings a fresh perspective to the intersection of AI and real-life solutions.



Source link

Continue Reading

Previous: Designer of Kate Middleton’s necklace speaks out after Princess of Wales’s latest move
Next: BBC Antiques Roadshow expert admits ‘I’ve never seen one’ as he declares item ‘one of the rarest ever’

Related News

World economy could get carved up into these 3 trading blocs
  • Uncategorized

World economy could get carved up into these 3 trading blocs

VedVision HeadLines July 6, 2025
Whales Power Bitcoin Cash ($BCH) to 8-Month High as Golden Cross Signals Breakout
  • Uncategorized

Whales Power Bitcoin Cash ($BCH) to 8-Month High as Golden Cross Signals Breakout

VedVision HeadLines July 6, 2025
A Brief History Of Wallet Clustering
  • Uncategorized

A Brief History Of Wallet Clustering

VedVision HeadLines July 6, 2025

Recent Posts

  • Bobby Brazier shares true feelings about BBC EastEnders exit as actor admits ‘it’s my first real job’
  • King Charles downs glass of whisky as he steps out on final day of Scottish tour
  • World economy could get carved up into these 3 trading blocs
  • Kapil Sharma calls Archana Puran Singh ‘Anaconda’, accuses her of locking Navjot Singh Sidhu; her reply wins hearts | Television News
  • Whales Power Bitcoin Cash ($BCH) to 8-Month High as Golden Cross Signals Breakout

Recent Comments

No comments to show.

Archives

  • July 2025
  • June 2025
  • May 2025
  • April 2025

Categories

  • Current Affairs
  • Shopping
  • Uncategorized

You may have missed

Bobby Brazier shares true feelings about BBC EastEnders exit as actor admits ‘it’s my first real job’
  • Current Affairs

Bobby Brazier shares true feelings about BBC EastEnders exit as actor admits ‘it’s my first real job’

VedVision HeadLines July 6, 2025
King Charles downs glass of whisky as he steps out on final day of Scottish tour
  • Current Affairs

King Charles downs glass of whisky as he steps out on final day of Scottish tour

VedVision HeadLines July 6, 2025
World economy could get carved up into these 3 trading blocs
  • Uncategorized

World economy could get carved up into these 3 trading blocs

VedVision HeadLines July 6, 2025
Kapil Sharma calls Archana Puran Singh ‘Anaconda’, accuses her of locking Navjot Singh Sidhu; her reply wins hearts | Television News
  • Current Affairs

Kapil Sharma calls Archana Puran Singh ‘Anaconda’, accuses her of locking Navjot Singh Sidhu; her reply wins hearts | Television News

VedVision HeadLines July 6, 2025
Copyright © All rights reserved. | MoreNews by AF themes.