venturebeat
Huawei's new open source technique shrinks LLMs to make them run on less powerful, less expensive hardware

Huawei’s Computing Systems Lab in Zurich has introduced a new open-source quantization method for large language models (LLMs) aimed at reducing memory demands without sacrificing output quality. The technique, called SINQ (Sinkhorn-Normalized Quantization), is designed to be fast, calibration-free, and easy to integrate into existing model workflows. The code for performing it has been made available by the Huawei research team on Github and Hugging Face under a permissive, enterprise-friendly Apache 2.0 license, allowing organizations to take and use it, modify it, and deploy it commercially — all for free.Across models of different sizes, SINQ cuts memory usage by 60–70%, depending on architecture and bit-width. This enables models that would previously require >60 GB of memory [...]

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

Destination
US judge rules Huawei must answer criminal charges about alleged Iran deal

A US judge has ruled that Huawei must stand trial following a 16-count indictment from 2019 accusing the Chinese telecommunications company of trying to steal trade secrets from its US rivals and sell [...]

Match Score: 94.93

venturebeat
Nvidia says it can shrink LLM memory 20x without changing model weights

Nvidia researchers have introduced a new technique that dramatically reduces how much memory large language models need to track conversation history — by as much as 20x — without modifying the mo [...]

Match Score: 85.25

venturebeat
Moonshot's Kimi K2 Thinking emerges as leading open source AI, outperforming GPT-5, Claude Sonnet 4.5 on key benchmarks

Even as concern and skepticism grows over U.S. AI startup OpenAI's buildout strategy and high spending commitments, Chinese open source AI providers are escalating their competition and one has e [...]

Match Score: 67.20

venturebeat
Mistral launches powerful Devstral 2 coding model including open source, laptop-friendly version

French AI startup Mistral has weathered a rocky period of public questioning over the last year to emerge, now here in December 2025, with new, crowd-pleasing models for enterprise and indie developer [...]

Match Score: 65.65

venturebeat
'Western Qwen': IBM wows with Granite 4 LLM launch and hybrid Mamba/Transformer architecture

IBM today announced the release of Granite 4.0, the newest generation of its homemade family of open source large language models (LLMs) designed to balance high performance with lower memory and cost [...]

Match Score: 64.10

venturebeat
Baseten takes on hyperscalers with new AI training platform that lets you own your model weights

Baseten, the AI infrastructure company recently valued at $2.15 billion, is making its most significant product pivot yet: a full-scale push into model training that could reshape how enterprises wean [...]

Match Score: 63.50

venturebeat
DeepSeek-V4 arrives with near state-of-the-art intelligence at 1/6th the cost of Opus 4.7, GPT-5.5

The whale has resurfaced. DeepSeek, the Chinese AI startup offshoot of High-Flyer Capital Management quantitative analysis firm, became a near-overnight sensation globally in January 2025 with the rel [...]

Match Score: 63.35

venturebeat
CrowdStrike & NVIDIA’s open source AI gives enterprises the edge against machine-speed attacks

Every SOC leader knows the feeling: drowning in alerts, blind to the real threat, stuck playing defense in a war waged at the speed of AI. Now CrowdStrike and NVIDIA are flipping the script. Armed wit [...]

Match Score: 62.26

Destination
Huawei reveals a wide-ass 16:10 foldable with a DeepSeek-powered AI assistant

Because of sanctions that will prevent Huawei’s latest foldable from going on sale in the US, many folks who are interested in the handset will never lay eyes on it in person. Still, you might want [...]

Match Score: 61.40