* . *
  • About
  • Advertise
  • Privacy & Policy
  • Contact
Saturday, September 6, 2025
Earth-News
  • Home
  • Business
  • Entertainment
  • General
  • Health
  • News

    Trending Tags

    • Trump Inauguration
    • United Stated
    • White House
    • Market Stories
    • Election Results
  • Science
  • Sports
  • Technology
    Coherent Joins LLNL’s STARFIRE Diode Technology Working Group to Advance Inertial Fusion Energy – GlobeNewswire

    Coherent Partners with LLNL’s STARFIRE Team to Drive Breakthroughs in Inertial Fusion Energy

    Gene Associated With Deadly Heart Disease in Golden Retrievers Identified – Technology Networks

    Breakthrough Discovery Uncovers Gene Behind Deadly Heart Disease in Golden Retrievers

    Monkey Island LNG Picks ConocoPhillips’ Liquefaction Technology – Hart Energy

    Monkey Island LNG Selects ConocoPhillips’ Advanced Liquefaction Technology for Next-Gen Energy Solutions

    Credo Technology Group Holding Ltd. (CRDO) Surpasses Q1 Earnings and Revenue Estimates – Yahoo Finance

    Credo Technology Group Surpasses Q1 Earnings and Revenue Expectations

    The Economist is hiring a science and technology correspondent – The Economist

    Exciting Opportunity: Become Our Next Science and Technology Correspondent!

    Blockchain lender Figure Technology seeks to raise up to $526M in IPO (FIGR:Pending) – Seeking Alpha

    Blockchain Lender Figure Technology Sets Sights on $526M in Thrilling IPO Launch

    Trending Tags

    • Nintendo Switch
    • CES 2017
    • Playstation 4 Pro
    • Mark Zuckerberg
No Result
View All Result
  • Home
  • Business
  • Entertainment
  • General
  • Health
  • News

    Trending Tags

    • Trump Inauguration
    • United Stated
    • White House
    • Market Stories
    • Election Results
  • Science
  • Sports
  • Technology
    Coherent Joins LLNL’s STARFIRE Diode Technology Working Group to Advance Inertial Fusion Energy – GlobeNewswire

    Coherent Partners with LLNL’s STARFIRE Team to Drive Breakthroughs in Inertial Fusion Energy

    Gene Associated With Deadly Heart Disease in Golden Retrievers Identified – Technology Networks

    Breakthrough Discovery Uncovers Gene Behind Deadly Heart Disease in Golden Retrievers

    Monkey Island LNG Picks ConocoPhillips’ Liquefaction Technology – Hart Energy

    Monkey Island LNG Selects ConocoPhillips’ Advanced Liquefaction Technology for Next-Gen Energy Solutions

    Credo Technology Group Holding Ltd. (CRDO) Surpasses Q1 Earnings and Revenue Estimates – Yahoo Finance

    Credo Technology Group Surpasses Q1 Earnings and Revenue Expectations

    The Economist is hiring a science and technology correspondent – The Economist

    Exciting Opportunity: Become Our Next Science and Technology Correspondent!

    Blockchain lender Figure Technology seeks to raise up to $526M in IPO (FIGR:Pending) – Seeking Alpha

    Blockchain Lender Figure Technology Sets Sights on $526M in Thrilling IPO Launch

    Trending Tags

    • Nintendo Switch
    • CES 2017
    • Playstation 4 Pro
    • Mark Zuckerberg
No Result
View All Result
Earth-News
No Result
View All Result
Home Technology

Researchers upend AI status quo by eliminating matrix multiplication in LLMs

June 26, 2024
in Technology
Researchers upend AI status quo by eliminating matrix multiplication in LLMs
Share on FacebookShare on Twitter

Illustration of a brain inside of a light bulb.

Enlarge / Illustration of a brain inside of a light bulb.

Researchers claim to have developed a new way to run AI language models more efficiently by eliminating matrix multiplication from the process. This fundamentally redesigns neural network operations that are currently accelerated by GPU chips. The findings, detailed in a recent preprint paper from researchers at the University of California Santa Cruz, UC Davis, LuxiTech, and Soochow University, could have deep implications for the environmental impact and operational costs of AI systems.

Matrix multiplication (often abbreviated to “MatMul”) is at the center of most neural network computational tasks today, and GPUs are particularly good at executing the math quickly because they can perform large numbers of multiplication operations in parallel. That ability momentarily made Nvidia the most valuable company in the world last week; the company currently holds an estimated 98 percent market share for data center GPUs, which are commonly used to power AI systems like ChatGPT and Google Gemini.

In the new paper, titled “Scalable MatMul-free Language Modeling,” the researchers describe creating a custom 2.7 billion parameter model without using MatMul that features similar performance to conventional large language models (LLMs). They also demonstrate running a 1.3 billion parameter model at 23.8 tokens per second on a GPU that was accelerated by a custom-programmed FPGA chip that uses about 13 watts of power (not counting the GPU’s power draw). The implication is that a more efficient FPGA “paves the way for the development of more efficient and hardware-friendly architectures,” they write.

The paper doesn’t provide power estimates for conventional LLMs, but this post from UC Santa Cruz estimates about 700 watts for a conventional model. However, in our experience, you can run a 2.7B parameter version of Llama 2 competently on a home PC with an RTX 3060 (that uses about 200 watts peak) powered by a 500-watt power supply. So, if you could theoretically completely run an LLM in only 13 watts on an FPGA (without a GPU), that would be a 38-fold decrease in power usage.

The technique has not yet been peer-reviewed, but the researchers—Rui-Jie Zhu, Yu Zhang, Ethan Sifferman, Tyler Sheaves, Yiqiao Wang, Dustin Richmond, Peng Zhou, and Jason Eshraghian—claim that their work challenges the prevailing paradigm that matrix multiplication operations are indispensable for building high-performing language models. They argue that their approach could make large language models more accessible, efficient, and sustainable, particularly for deployment on resource-constrained hardware like smartphones.

Doing away with matrix math

In the paper, the researchers mention BitNet (the so-called “1-bit” transformer technique that made the rounds as a preprint in October) as an important precursor to their work. According to the authors, BitNet demonstrated the viability of using binary and ternary weights in language models, successfully scaling up to 3 billion parameters while maintaining competitive performance.

However, they note that BitNet still relied on matrix multiplications in its self-attention mechanism. Limitations of BitNet served as a motivation for the current study, pushing them to develop a completely “MatMul-free” architecture that could maintain performance while eliminating matrix multiplications even in the attention mechanism.

>>> Read full article>>>
Copyright for syndicated content belongs to the linked Source : Ars Technica – https://arstechnica.com/?p=2033314

Tags: Researcherstechnologyupend
Previous Post

OpenAI’s ChatGPT for Mac is now available to all users

Next Post

Star Wars behind the scenes: Creating the unique aesthetic of The Acolyte

Ecology issues $738K penalty against downtown Walla Walla Chevron owner – Union-Bulletin

Ecology issues $738K penalty against downtown Walla Walla Chevron owner – Union-Bulletin

September 6, 2025
What the science says about acetaminophen, pregnant mothers and autism – NBC News

What the science says about acetaminophen, pregnant mothers and autism – NBC News

September 6, 2025
RFK Jr accused of ‘reckless disregard for science and the truth’ in Senate hearing – The Guardian

RFK Jr. Faces Fierce Backlash for ‘Reckless Disregard for Science and Truth’ in Heated Senate Hearing

September 6, 2025
Start The Go-Go Years Today: Ignite Your Early Retirement Lifestyle – Forbes

Start The Go-Go Years Today: Ignite Your Early Retirement Lifestyle – Forbes

September 6, 2025
When Sports Teach More Than Skills – American Enterprise Institute

When Sports Teach More Than Skills – American Enterprise Institute

September 6, 2025
Abby Dow and England near record territory as Australia search for upset – The Guardian

Abby Dow and England Close in on Record as Australia Hunts for an Upset

September 6, 2025
America is getting the economy we voted for – Noah Smith | Substack

America Is Finally Experiencing the Economy We Created

September 6, 2025
Pendulum Announce Homecoming 2026 Australian Tour – yahoo.com

Pendulum Announces Thrilling Homecoming Tour Across Australia in 2026

September 6, 2025
HHS responds to report about autism and acetaminophen : Shots – Health News – NPR

HHS Addresses New Findings on Autism and Acetaminophen Use

September 6, 2025
September 2, 2025: Trump administration news – CNN

September 2, 2025: Trump administration news – CNN

September 6, 2025

Categories

Archives

September 2025
MTWTFSS
1234567
891011121314
15161718192021
22232425262728
2930 
« Aug    
Earth-News.info

The Earth News is an independent English-language daily published Website from all around the World News

Browse by Category

  • Business (20,132)
  • Ecology (810)
  • Economy (828)
  • Entertainment (21,705)
  • General (16,889)
  • Health (9,869)
  • Lifestyle (841)
  • News (22,149)
  • People (830)
  • Politics (835)
  • Science (16,039)
  • Sports (21,327)
  • Technology (15,808)
  • World (809)

Recent News

Ecology issues $738K penalty against downtown Walla Walla Chevron owner – Union-Bulletin

Ecology issues $738K penalty against downtown Walla Walla Chevron owner – Union-Bulletin

September 6, 2025
What the science says about acetaminophen, pregnant mothers and autism – NBC News

What the science says about acetaminophen, pregnant mothers and autism – NBC News

September 6, 2025
  • About
  • Advertise
  • Privacy & Policy
  • Contact

© 2023 earth-news.info

No Result
View All Result

© 2023 earth-news.info

No Result
View All Result

© 2023 earth-news.info

Go to mobile version