* . *
  • About
  • Advertise
  • Privacy & Policy
  • Contact
Monday, June 30, 2025
Earth-News
  • Home
  • Business
  • Entertainment
    Susquehanna Raises Penn Entertainment Inc. (PENN) Price Target. – Yahoo Finance

    Susquehanna Raises Price Target for Penn Entertainment Inc. (PENN)

    George Lopez is coming to Spokane – KXLY.com

    George Lopez is coming to Spokane – KXLY.com

    Netflix unveils Dallas immersive venue for fans of hit shows like ‘Squid Game,’ ‘Stranger Things’ – Houston Chronicle

    Step Inside Netflix’s New Dallas Immersive Experience Featuring Hits Like ‘Squid Game’ and ‘Stranger Things

    ‘Puttin’ on the Ritz’: Civic Players bring ‘Young Frankenstein’ to life – Yahoo

    Civic Players Deliver a Hilarious and Unforgettable Performance of ‘Young Frankenstein

    ‘Wheel of Fortune’: Amputee Wins $60,000 After Breaking Incredible ‘Curse’ – Hastings Tribune

    Wheel of Fortune’ Amputee Breaks Incredible ‘Curse’ to Win $60,000!

    North Star Sports & Entertainment Network: Coming soon – KTTC News

    North Star Sports & Entertainment Network: Coming soon – KTTC News

  • General
  • Health
  • News

    Cracking the Code: Why China’s Economic Challenges Aren’t Shaking Markets, Unlike America’s” – Bloomberg

    Trump’s Narrow Window to Spread the Truth About Harris

    Trump’s Narrow Window to Spread the Truth About Harris

    Israel-Gaza war live updates: Hamas leader Ismail Haniyeh assassinated in Iran, group says

    Israel-Gaza war live updates: Hamas leader Ismail Haniyeh assassinated in Iran, group says

    PAP Boss to Niger Delta Youths, Stay Away from the Protest

    PAP Boss to Niger Delta Youths, Stay Away from the Protest

    Court Restricts Protests In Lagos To Freedom, Peace Park

    Court Restricts Protests In Lagos To Freedom, Peace Park

    Fans React to Jazz Jennings’ Inspiring Weight Loss Journey

    Fans React to Jazz Jennings’ Inspiring Weight Loss Journey

    Trending Tags

    • Trump Inauguration
    • United Stated
    • White House
    • Market Stories
    • Election Results
  • Science
  • Sports
  • Technology
    Mirrors in space and underwater curtains: can technology buy us enough time to save the Arctic ice caps? – The Guardian

    Can Technology Like Space Mirrors and Underwater Curtains Buy Us Time to Save the Arctic Ice Caps?

    Naples restaurant owner prepares for hurricane season with new flood technology – Fox4Now.com

    Naples restaurant owner prepares for hurricane season with new flood technology – Fox4Now.com

    Emerging Memory and Storage Technology Market Analysis Report 2025-2034 | AI and HPC Boom Fuels Surging Demand for Fast, Low-Power Memory Devices – Yahoo Finance

    How AI and HPC Are Driving Explosive Growth in Fast, Low-Power Memory Technologies Through 2034

    Ostin Technology (OST): Volatility’s Warning or Contrarian Opportunity? – AInvest

    Ostin Technology (OST): Navigating Market Volatility – Red Flag or Hidden Opportunity?

    St. Francis Medical Center brings advanced robotic surgery technology to Northeast Louisiana – KNOE

    St. Francis Medical Center brings advanced robotic surgery technology to Northeast Louisiana – KNOE

    Wayve Expands Engineering Leadership to Power Next-Gen Autonomous Driving Technology – Silicon Canals

    Wayve Boosts Engineering Leadership to Accelerate Next-Gen Autonomous Driving Innovation

    Trending Tags

    • Nintendo Switch
    • CES 2017
    • Playstation 4 Pro
    • Mark Zuckerberg
No Result
View All Result
  • Home
  • Business
  • Entertainment
    Susquehanna Raises Penn Entertainment Inc. (PENN) Price Target. – Yahoo Finance

    Susquehanna Raises Price Target for Penn Entertainment Inc. (PENN)

    George Lopez is coming to Spokane – KXLY.com

    George Lopez is coming to Spokane – KXLY.com

    Netflix unveils Dallas immersive venue for fans of hit shows like ‘Squid Game,’ ‘Stranger Things’ – Houston Chronicle

    Step Inside Netflix’s New Dallas Immersive Experience Featuring Hits Like ‘Squid Game’ and ‘Stranger Things

    ‘Puttin’ on the Ritz’: Civic Players bring ‘Young Frankenstein’ to life – Yahoo

    Civic Players Deliver a Hilarious and Unforgettable Performance of ‘Young Frankenstein

    ‘Wheel of Fortune’: Amputee Wins $60,000 After Breaking Incredible ‘Curse’ – Hastings Tribune

    Wheel of Fortune’ Amputee Breaks Incredible ‘Curse’ to Win $60,000!

    North Star Sports & Entertainment Network: Coming soon – KTTC News

    North Star Sports & Entertainment Network: Coming soon – KTTC News

  • General
  • Health
  • News

    Cracking the Code: Why China’s Economic Challenges Aren’t Shaking Markets, Unlike America’s” – Bloomberg

    Trump’s Narrow Window to Spread the Truth About Harris

    Trump’s Narrow Window to Spread the Truth About Harris

    Israel-Gaza war live updates: Hamas leader Ismail Haniyeh assassinated in Iran, group says

    Israel-Gaza war live updates: Hamas leader Ismail Haniyeh assassinated in Iran, group says

    PAP Boss to Niger Delta Youths, Stay Away from the Protest

    PAP Boss to Niger Delta Youths, Stay Away from the Protest

    Court Restricts Protests In Lagos To Freedom, Peace Park

    Court Restricts Protests In Lagos To Freedom, Peace Park

    Fans React to Jazz Jennings’ Inspiring Weight Loss Journey

    Fans React to Jazz Jennings’ Inspiring Weight Loss Journey

    Trending Tags

    • Trump Inauguration
    • United Stated
    • White House
    • Market Stories
    • Election Results
  • Science
  • Sports
  • Technology
    Mirrors in space and underwater curtains: can technology buy us enough time to save the Arctic ice caps? – The Guardian

    Can Technology Like Space Mirrors and Underwater Curtains Buy Us Time to Save the Arctic Ice Caps?

    Naples restaurant owner prepares for hurricane season with new flood technology – Fox4Now.com

    Naples restaurant owner prepares for hurricane season with new flood technology – Fox4Now.com

    Emerging Memory and Storage Technology Market Analysis Report 2025-2034 | AI and HPC Boom Fuels Surging Demand for Fast, Low-Power Memory Devices – Yahoo Finance

    How AI and HPC Are Driving Explosive Growth in Fast, Low-Power Memory Technologies Through 2034

    Ostin Technology (OST): Volatility’s Warning or Contrarian Opportunity? – AInvest

    Ostin Technology (OST): Navigating Market Volatility – Red Flag or Hidden Opportunity?

    St. Francis Medical Center brings advanced robotic surgery technology to Northeast Louisiana – KNOE

    St. Francis Medical Center brings advanced robotic surgery technology to Northeast Louisiana – KNOE

    Wayve Expands Engineering Leadership to Power Next-Gen Autonomous Driving Technology – Silicon Canals

    Wayve Boosts Engineering Leadership to Accelerate Next-Gen Autonomous Driving Innovation

    Trending Tags

    • Nintendo Switch
    • CES 2017
    • Playstation 4 Pro
    • Mark Zuckerberg
No Result
View All Result
Earth-News
No Result
View All Result
Home Technology

Meta researchers distill System 2 thinking into LLMs, improving performance on complex reasoning

July 13, 2024
in Technology
Meta researchers distill System 2 thinking into LLMs, improving performance on complex reasoning
Share on FacebookShare on Twitter

July 12, 2024 9:20 PM

A robot thinking fast and slow

Image credit: VentureBeat with DALL-E 3

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More

Large language models (LLMs) are very good at answering simple questions but require special prompting techniques to handle complex tasks that need reasoning and planning. Often referred to as “System 2” techniques, these prompting schemes enhance the reasoning capabilities of LLMs by forcing them to generate intermediate steps toward solving a problem.

While effective, System 2 techniques make LLM applications slow and computationally expensive. In a new paper, researchers at Meta FAIR present “System 2 distillation,” a technique that teaches LLMs complex tasks without requiring intermediate steps. 

System 1 and System 2 in cognitive science and LLMs

In cognitive science, System 1 and System 2 refer to two distinct modes of thinking. System 1 thinking is fast, intuitive and automatic. It is what we use when recognizing patterns, making quick judgments, or understanding familiar symbols. For example, we use System 1 thinking to identify traffic signs, recognize faces, and associate basic symbols with their meanings.

System 2 thinking, on the other hand, is slow, deliberate and analytical. It requires conscious effort and is used for complex problem-solving, such as manipulating abstract symbols, solving mathematical equations or planning a trip. 

LLMs are usually considered analogous to System 1 thinking. They can generate text very quickly, but they struggle with tasks that require deliberate reasoning and planning. 

In recent years, AI researchers have shown that LLMs can be made to mimic System 2 thinking by prompting them to generate intermediate reasoning steps before providing their final answer. For example, “Chain of Thought” is a prompting technique that instructs the LLM to explain its reasoning process step by step, which often leads to more accurate results for logical reasoning tasks. Several System 2 prompting techniques are tailored for different tasks.

“Many of these methods are shown to produce more accurate results due to this explicit reasoning, but typically do so at much higher inference cost and latency for a response,” the Meta AI researchers write. “Due to the latter, many of these approaches are not used in production systems, which mostly use System 1 generations.”

System 2 distillation

An interesting observation about System 2 thinking in humans is that when we repeatedly perform a task that requires deliberate effort, it gradually becomes ingrained in our System 1. For example, when you learn to drive, you use a lot of conscious effort to control the car, follow traffic rules and navigate. But as you gain more experience, driving becomes second nature. You no longer need to think about each step, and you can perform them intuitively and automatically.

This phenomenon inspired the Meta AI researchers to develop “System 2 distillation” for LLMs. 

Distillation is a common technique in machine learning (ML), where a larger model, referred to as the “teacher,” is used to train a smaller model, or the “student.” For example, developers often use frontier models such as GPT-4 and Claude to generate training examples for smaller models such as Llama-2 7B.

However, System 2 distillation does not use a separate teacher model. Instead, the researchers found a way to distill the knowledge gained from the model’s own System 2 reasoning capabilities into its fast-paced and compute-efficient System 1 generation.

System 2 distillationSystem 2 distillation (source: arxiv)

The process starts by prompting the LLM to solve a problem using System 2 prompting techniques. The responses are then verified for correctness through an unsupervised mechanism. For example, they use “self-consistency,” where the model is given the same prompt multiple times. Its answers are then compared, and the one that shows up most often is considered the correct answer and is chosen for the distillation dataset. If the answers are too inconsistent, then the example and its answers are discarded.

Next, they discard the intermediate steps generated by System 2 reasoning and only keep the final answers. Finally, they fine-tuned the model on the initial question and the answer. This allows the model to skip the reasoning steps and jump straight to the answer.

System 2 distillation in action

The researchers evaluated their method on a range of reasoning tasks and four different System 2 prompting techniques. For the base model, they used Llama-2-70B, which is large enough to have the capacity for internalizing new knowledge.

The System 2 approaches they used in their experiments include Chain-of-Thought, System 2 Attention, Rephrase and Respond and Branch-Solve-Merge. Some of these techniques require the model to be prompted several times, which makes them both slow and expensive. For example, Rephrase and Respond first prompts the model to rephrase the original query with elaboration, and then it re-prompts the model with the rephrased question. Branch-Solve-Merge is even more complicated and requires multiple back-and-forths with the model.

The results show that System 2 distillation can significantly improve the performance of LLMs on complex reasoning tasks, often matching or exceeding the accuracy of the original System 2 methods. Additionally, the distilled models can generate responses much faster and with less compute because they don’t have to go through the intermediate reasoning steps.

For example, they found that distillation was successful for tasks that use System 2 Attention to deal with biased opinions or irrelevant information. It also showed impressive results in some reasoning tasks, where Rephrase and Respond is used to clarify and improve responses, and for fine-grained evaluation and processing of tasks through Branch-Solve-Merge.

“We have shown that in many cases it is possible to distill this System 2 reasoning into the outputs of the LLM without intermediate generations while maintaining, or sometimes even improving, performance,” the researchers write. 

However, the researchers also found that, like humans, LLMs can’t distill all types of reasoning skills into their fast-paced inference mechanism. For example, they were unable to successfully distill complex math reasoning tasks that required Chain-of-Thought prompting. This suggests that some tasks might always require deliberate reasoning.

There is much more to be learned about System 2 distillation, such as how well it works on smaller models and how distillation affects the model’s broader performance on tasks that were not included in the distillation training dataset. It is also worth noting that LLM benchmarks are often prone to contamination, where the model already has some kind of knowledge of the test examples, resulting in bloated outcomes on test sets. 

However, distillation will surely be a powerful optimization tool for mature LLM pipelines that perform specific tasks at each step.

“Looking forward, systems that can distill useful tasks in this way free up more time to spend on reasoning about the tasks that they cannot yet do well, just as humans do,” the researchers write.

VB Daily

Stay in the know! Get the latest news in your inbox daily

By subscribing, you agree to VentureBeat’s Terms of Service.

Thanks for subscribing. Check out more VB newsletters here.

An error occured.

>>> Read full article>>>
Copyright for syndicated content belongs to the linked Source : VentureBeat – https://venturebeat.com/ai/meta-researchers-distill-system-2-thinking-into-llms-improving-performance-on-complex-reasoning/

Tags: distillResearcherstechnology
Previous Post

DeepMind’s PEER scales language models with millions of tiny experts

Next Post

Ripple (XRP) Price Prediction: XRP on the Rise – Will the Uptrend Hold?

More than 1,800 National Science Foundation workers abruptly kicked out of agency headquarters – Space

Over 1,800 National Science Foundation Employees Suddenly Evicted from Agency Headquarters

June 30, 2025
Scientists Retrace 30,000-Year-Old Sea Voyage, in a Hollowed-Out Log – The New York Times

Scientists Successfully Recreate Epic 30,000-Year-Old Sea Voyage in a Hollowed-Out Log

June 30, 2025
Tom Brady Ditches Sneakers Lifestyle For Jeff Bezos & Lauren Sánchez’s Wedding – Yahoo

Tom Brady Swaps Sneakers Lifestyle for Glamorous Jeff Bezos and Lauren Sánchez Wedding Celebration

June 30, 2025
Chivu Confirms France Star Fit For Inter Milan Vs Fluminense Club World Cup Clash – Yahoo Sports

Chivu Confirms France Star Ready for Inter Milan’s Club World Cup Showdown Against Fluminense

June 30, 2025
Trump’s tariffs damage the US economy more if they drive investors away from American assets – Peterson Institute for International Economics

How Trump’s Tariffs Could Backfire by Driving Investors Away from American Assets

June 30, 2025
Surprising New Study Links Daytime Napping to Hidden Health Risks – Prevention

Surprising New Research Uncovers Hidden Health Risks of Daytime Napping

June 30, 2025
The 1975’s Matty Healy Says ‘We Don’t Need More Politics’ at Glastonbury – Variety

The 1975’s Matty Healy Declares “We Don’t Need More Politics” at Glastonbury

June 30, 2025
Mirrors in space and underwater curtains: can technology buy us enough time to save the Arctic ice caps? – The Guardian

Can Technology Like Space Mirrors and Underwater Curtains Buy Us Time to Save the Arctic Ice Caps?

June 29, 2025
Guirassy latest candidate for Milan attack with 70m price tag – Yahoo Sports

Guirassy Emerges as Milan’s Top Striker Target with €70 Million Price Tag

June 29, 2025
‘Half the tree of life’: ecologists’ horror as nature reserves are emptied of insects – The Guardian

Half the Tree of Life at Risk: Ecologists Warn as Insect Populations Plummet in Nature Reserves

June 29, 2025

Categories

Archives

June 2025
MTWTFSS
 1
2345678
9101112131415
16171819202122
23242526272829
30 
« May    
Earth-News.info

The Earth News is an independent English-language daily published Website from all around the World News

Browse by Category

  • Business (20,132)
  • Ecology (700)
  • Economy (724)
  • Entertainment (21,613)
  • General (15,628)
  • Health (9,763)
  • Lifestyle (729)
  • News (22,149)
  • People (725)
  • Politics (730)
  • Science (15,941)
  • Sports (21,220)
  • Technology (15,708)
  • World (704)

Recent News

More than 1,800 National Science Foundation workers abruptly kicked out of agency headquarters – Space

Over 1,800 National Science Foundation Employees Suddenly Evicted from Agency Headquarters

June 30, 2025
Scientists Retrace 30,000-Year-Old Sea Voyage, in a Hollowed-Out Log – The New York Times

Scientists Successfully Recreate Epic 30,000-Year-Old Sea Voyage in a Hollowed-Out Log

June 30, 2025
  • About
  • Advertise
  • Privacy & Policy
  • Contact

© 2023 earth-news.info

No Result
View All Result

© 2023 earth-news.info

No Result
View All Result

© 2023 earth-news.info

Go to mobile version