Close Menu
Breaking News in Technology & Business – Tech Geekwire

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    IEEE Spectrum: Flagship Publication of the IEEE

    July 4, 2025

    GOP Opposition Mounts Against AI Provision in Reconciliation Bill

    July 4, 2025

    Navigation Help

    July 4, 2025
    Facebook X (Twitter) Instagram
    Breaking News in Technology & Business – Tech GeekwireBreaking News in Technology & Business – Tech Geekwire
    • New
      • Amazon
      • Digital Health Technology
      • Microsoft
      • Startup
    • AI
    • Corporation
    • Crypto
    • Event
    Facebook X (Twitter) Instagram
    Breaking News in Technology & Business – Tech Geekwire
    Home » Pruna AI Opens Optimization Framework to Streamline AI Model Efficiency
    AI

    Pruna AI Opens Optimization Framework to Streamline AI Model Efficiency

    techgeekwireBy techgeekwireMarch 27, 2025No Comments3 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email

    Pruna AI Launches Open-Source Framework for AI Model Optimization

    European startup Pruna AI is releasing its AI model optimization framework to the open-source community this Thursday. The framework focuses on compression algorithms to improve the efficiency of AI models. Pruna AI’s framework incorporates several methods such as caching, pruning, quantization, and distillation to optimize AI models.

    “We also standardize saving and loading the compressed models, applying combinations of these compression methods, and also evaluating your compressed model after you compress it,” explained John Rachwan, co-founder and CTO of Pruna AI, in an interview with TechCrunch. The framework assesses the quality impact and performance benefits after compression.

    Rachwan went on to explain, “If I were to use a metaphor, we are similar to how Hugging Face standardized transformers and diffusers — how to call them, how to save them, load them, etc. We are doing the same, but for efficiency methods.”

    A Comprehensive Approach to AI Model Optimization

    Major AI labs are already leveraging various compression techniques. For example, OpenAI employs distillation to create faster versions of its flagship models. This method likely enabled the creation of GPT-4 Turbo, an accelerated version of GPT-4. Similarly, the Flux.1-schnell image generation model is a distilled version from Black Forest Labs.

    Distillation is a technique where knowledge is extracted from a large AI model, the “teacher,” to train a smaller “student” model. Developers feed requests to the teacher model, record its outputs, and assess their accuracy against a dataset. These outputs then train the student model to mimic the teacher’s behavior.

    According to Rachwan, “For big companies, what they usually do is that they build this stuff in-house. And what you can find in the open source world is usually based on single methods. For example, let’s say one quantization method for LLMs, or one caching method for diffusion models. But you cannot find a tool that aggregates all of them, makes them all easy to use and combine together. And this is the big value that Pruna is bringing right now.”

    Pruna AI co-founders
    Left to right: Rayan Nait Mazi, Bertrand Charpentier, John Rachwan, Stephan Günnemann

    While the framework supports diverse model types, Pruna AI currently focuses on image and video generation models. Notable users include Scenario and PhotoRoom. Besides the open-source edition, Pruna AI offers an enterprise version with advanced features, including an optimization agent.

    Rachwan added, “The most exciting feature that we are releasing soon will be a compression agent. Basically, you give it your model, you say: ‘I want more speed but don’t drop my accuracy by more than 2%.’ And then, the agent will just do its magic. It will find the best combination for you, return it for you. You don’t have to do anything as a developer.”

    Pruna AI’s pro version utilizes an hourly pricing model, similar to cloud-based GPU rentals, according to Rachwan. Optimized models can lead to significant cost savings in inference, especially for models critical to AI infrastructure. With its compression framework, Pruna AI has reduced the size of a Llama model eightfold without significant accuracy loss.

    Pruna AI views its compression framework as a cost-effective investment. The company completed a $6.5 million seed funding round recently, with investments from EQT Ventures, Daphni, Motier Ventures, and Kima Ventures.

    AI compression open-source optimization startups
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    techgeekwire
    • Website

    Related Posts

    IEEE Spectrum: Flagship Publication of the IEEE

    July 4, 2025

    GOP Opposition Mounts Against AI Provision in Reconciliation Bill

    July 4, 2025

    Navigation Help

    July 4, 2025

    Andreessen Horowitz Backs Controversial Startup Cluely Despite ‘Rage-Bait’ Marketing

    July 4, 2025

    Invesco QQQ ETF Hits All-Time High as Tech Stocks Continue to Soar

    July 4, 2025

    ContractPodAi Partners with Microsoft to Advance Legal AI Automation

    July 4, 2025
    Leave A Reply Cancel Reply

    Top Reviews
    Editors Picks

    IEEE Spectrum: Flagship Publication of the IEEE

    July 4, 2025

    GOP Opposition Mounts Against AI Provision in Reconciliation Bill

    July 4, 2025

    Navigation Help

    July 4, 2025

    Andreessen Horowitz Backs Controversial Startup Cluely Despite ‘Rage-Bait’ Marketing

    July 4, 2025
    Advertisement
    Demo
    About Us
    About Us

    A rich source of news about the latest technologies in the world. Compiled in the most detailed and accurate manner in the fastest way globally. Please follow us to receive the earliest notification

    We're accepting new partnerships right now.

    Email Us: info@example.com
    Contact: +1-320-0123-451

    Our Picks

    IEEE Spectrum: Flagship Publication of the IEEE

    July 4, 2025

    GOP Opposition Mounts Against AI Provision in Reconciliation Bill

    July 4, 2025

    Navigation Help

    July 4, 2025
    Categories
    • AI (2,696)
    • Amazon (1,056)
    • Corporation (990)
    • Crypto (1,130)
    • Digital Health Technology (1,079)
    • Event (523)
    • Microsoft (1,230)
    • New (9,568)
    • Startup (1,164)
    © 2025 TechGeekWire. Designed by TechGeekWire.
    • Home

    Type above and press Enter to search. Press Esc to cancel.