Training a large artificial intelligence model is expensive, not just in dollars, but in time, energy, and computational ...
Multiverse Computing S.L. said today it has raised $215 million in funding to accelerate the deployment of its quantum computing-inspired artificial intelligence model compression technology, which ...
Effective compression is about finding patterns to make data smaller without losing information. When an algorithm or model can accurately guess the next piece of data in a sequence, it shows it’s ...
Pruna AI, a European startup that has been working on compression algorithms for AI models, is making its optimization framework open source on Thursday. Pruna AI has been creating a framework that ...
Intel has disclosed a maximum severity vulnerability in some versions of its Intel Neural Compressor software for AI model compression. The bug, designated as CVE-2024-22476, provides an ...
Nvidia researchers have introduced a new technique that dramatically reduces how much memory large language models need to track conversation history — by as much as 20x — without modifying the model ...
Chinese artificial intelligence startup DeepSeek has introduced DeepSeek-OCR, an open-source model accompanied by a research paper that pioneers a novel "optical compression" method aimed at reducing ...
Small changes in the large language models (LLMs) at the heart of AI applications can result in substantial energy savings, according to a report released by the United Nations Educational, Scientific ...
Large language models (LLMs) such as GPT-4o and other modern state-of-the-art generative models like Anthropic’s Claude, Google's PaLM and Meta's Llama have been dominating the AI field recently.
Model distillation is one of the technology trends that has reached a level of maturity identified in Gartner’s 2025 Hype Cycle for artificial intelligence (AI) as “the slope of enlightenment”.