Close Menu
    Trending
    • Blazing-Fast ML Model Serving with FastAPI + Redis (Boost 10x Speed!) | by Sarayavalasaravikiran | AI Simplified in Plain English | Jul, 2025
    • AI Knowledge Bases vs. Traditional Support: Who Wins in 2025?
    • Why Your Finance Team Needs an AI Strategy, Now
    • How to Access NASA’s Climate Data — And How It’s Powering the Fight Against Climate Change Pt. 1
    • From Training to Drift Monitoring: End-to-End Fraud Detection in Python | by Aakash Chavan Ravindranath, Ph.D | Jul, 2025
    • Using Graph Databases to Model Patient Journeys and Clinical Relationships
    • Cuba’s Energy Crisis: A Systemic Breakdown
    • AI Startup TML From Ex-OpenAI Exec Mira Murati Pays $500,000
    AIBS News
    • Home
    • Artificial Intelligence
    • Machine Learning
    • AI Technology
    • Data Science
    • More
      • Technology
      • Business
    AIBS News
    Home»Data Science»Multiverse Says It Compresses Llama Models by 80%
    Data Science

    Multiverse Says It Compresses Llama Models by 80%

    Team_AIBS NewsBy Team_AIBS NewsApril 9, 2025No Comments3 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Donostia, Spain – April 8, 2025 – Multiverse Computing at the moment launched two new AI fashions compressed by CompactifAI, Multiverse’s AI compressor: 80 p.c compressed variations of Llama 3.1-8B and Llama 3.3-70B.

    Each fashions have 60 p.c fewer parameters than the unique fashions, 84 p.c better power effi ciency, 40 p.c sooner inference, and yield a 50 p.c price discount with out sacrifi cing accuracy, in response to Multiverse. “AI builders can instantly plug the fashions into any utility – edge, on-premise, or cloud,” the corporate stated.

    Multiverse will launch variations of the highest LLMs compressed by CompactifAI over the approaching months.

    “Meta’s Llama 4 launch underscores a serious shift in AI: smaller, extra highly effective, and multimodal fashions are not non-obligatory — they’re the brand new default,” stated Dmitry Zakharchenko, chief software program workplace at Blaize, a U.S. edge AI chip firm. “As AI strikes from cloud to edge, success is determined by fashions which are environment friendly, inexpensive, and absolutely programmable.”

    Multiverse stated CompactifAI is the fi rst compressor of its form, utilizing quantum-inspired tensor networks to make AI methods extra effi cient and moveable, decreasing dimension as much as 93 p.c with solely a 2-3 p.c drop in accuracy—an astounding feat when in comparison with an industry-standard 20-30% accuracy loss with 50-60 p.c compression methods.

    “CompactifAI is altering the economics of AI processing and opening up new use circumstances for AI fashions,” stated Enrique Lizaso Olmos, CEO of Multiverse Computing. “Eff orts to curb unwieldy fashions have come up quick. Our novel strategy to compression grounded in quantum-inspired methods makes it attainable to pair efficiency with processing effi ciency and offers us an enormous edge on LLM suppliers.”

    Multiverse Computing was based in 2019 by pioneers in quantum-inspired software program to develop novel options to advanced enterprise issues. In 2023 the corporate started making use of its core expertise to deal with the AI power disaster with CompactifAI.

    LLM suppliers have turned to methods reminiscent of pruning and quantization to compress fashions however have but to eradicate the tradeoff between dimension and efficiency. As an illustration, Llama3.1-8B Slim by CompactifAI requires 300x fewer coaching tokens than Meta’s CAI Llama3, and 3x fewer coaching tokens than Nvidia’s Llama3.1-Minitron whereas outperforming throughout benchmarks. For Llama3.3-70B Slim by CompactifAI, comparative benchmarks present a rise in reasoning capabilities whereas sustaining unique precision.

    “We’re quickly delivering compressed variations of probably the most highly effective LLMs on the earth,” stated Sam Mugel, Chief Expertise Offi cer at Multiverse. “The superior capabilities of those two huge fashions can now fi t into smartphones, laptops, and automobiles, or real-world machines like oil rigs and satellites. Our aggressive roadmap to roll out dozens of compressed, main LLMs might dramatically speed up the affect of AI in the actual world.”





    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleSuicide forum is subject of first Ofcom probe, BBC understands
    Next Article [ML Paper Note][2024]MLoRA: Multi-Domain Low-Rank Adaptive Network for Click-Through Rate Prediction | by wh | Apr, 2025
    Team_AIBS News
    • Website

    Related Posts

    Data Science

    AI Knowledge Bases vs. Traditional Support: Who Wins in 2025?

    July 2, 2025
    Data Science

    Using Graph Databases to Model Patient Journeys and Clinical Relationships

    July 1, 2025
    Data Science

    The New Career Crisis: AI Is Breaking the Entry-Level Path for Gen Z

    July 1, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Blazing-Fast ML Model Serving with FastAPI + Redis (Boost 10x Speed!) | by Sarayavalasaravikiran | AI Simplified in Plain English | Jul, 2025

    July 2, 2025

    I Tried Buying a Car Through Amazon: Here Are the Pros, Cons

    December 10, 2024

    Amazon and eBay to pay ‘fair share’ for e-waste recycling

    December 10, 2024

    Artificial Intelligence Concerns & Predictions For 2025

    December 10, 2024

    Barbara Corcoran: Entrepreneurs Must ‘Embrace Change’

    December 10, 2024
    Categories
    • AI Technology
    • Artificial Intelligence
    • Business
    • Data Science
    • Machine Learning
    • Technology
    Most Popular

    Learnings from a Machine Learning Engineer — Part 4: The Model

    February 14, 2025

    The Real Machine Learning Loop: From Problem to Production (And Back Again) | by Julieta D. Rubis | May, 2025

    May 25, 2025

    Deploy agentic AI faster with DataRobot and NVIDIA

    March 18, 2025
    Our Picks

    Blazing-Fast ML Model Serving with FastAPI + Redis (Boost 10x Speed!) | by Sarayavalasaravikiran | AI Simplified in Plain English | Jul, 2025

    July 2, 2025

    AI Knowledge Bases vs. Traditional Support: Who Wins in 2025?

    July 2, 2025

    Why Your Finance Team Needs an AI Strategy, Now

    July 2, 2025
    Categories
    • AI Technology
    • Artificial Intelligence
    • Business
    • Data Science
    • Machine Learning
    • Technology
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2024 Aibsnews.comAll Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.