AI is no longer the exclusive domain of tech giants! Tether launches QVAC—has the moment arrived for everyone to have their own LLM?

Stablecoin issuer Tether announced today (17th) a major technological breakthrough for its AI infrastructure QVAC Fabric: the world’s first cross-platform supported BitNet LoRA fine-tuning framework. This allows large language models (LLMs), which previously required enterprise-grade GPUs and cloud computing power, to now be trained and inferred on consumer-grade hardware, including smartphones.

Smartphones can now train LLMs: 1B models completed within 1 hour

According to data released by Tether, the framework has successfully achieved BitNet model fine-tuning on various devices, including the common Samsung S25 and iPhone 16.

Samsung S25 (Adreno GPU):

  • 125 million parameters model: about 10 minutes for fine-tuning
  • 1 billion parameters model: about 1 hour 18 minutes

iPhone 16 (Apple GPU):

  • 1 billion parameters: about 1 hour 45 minutes

Extreme testing can handle models up to 13 billion parameters for fine-tuning

Previously, AI training tasks were performed using high-end NVIDIA GPUs, but now they have been compressed to edge devices like smartphones.

Key Technologies: BitNet + LoRA — Cutting AI costs drastically

The core of this breakthrough lies in the combination of two technologies:

BitNet (1-bit LLM):

  • Compresses traditional high-precision weights into just -1, 0, 1, significantly reducing memory and computation requirements.

LoRA (Low-Rank Adaptation):

  • Trains only a small number of parameters (reducing training volume by up to 99%), greatly lowering fine-tuning costs.

Together, these enable models to operate in extremely low-resource environments.

Real-world tests show that BitNet-1B uses 77.8% less VRAM than Gemma-3-1B and 65.6% less than Qwen3-0.6B. On the same hardware, this allows running models approximately twice as large.

GPU unlocking for mobile AI: performance boosted up to 11 times

Another key breakthrough by QVAC is enabling BitNet to run on “non-NVIDIA” ecosystems. It supports GPUs from AMD, Intel, Apple Silicon, and even mobile GPUs like Adreno, Mali, and Apple Bionic.

Large language models are no longer the exclusive domain of tech giants; AI can now be decentralized

Tether CEO Paolo Ardoino stated: “Intelligence will be a key factor in future societal development. It has the potential to enhance social stability, serve as a bridge connecting society, or further empower a select few. The future of AI should be accessible, usable, and available to everyone, not monopolized by a handful of cloud service providers with enormous resources.”

Traditional AI development heavily relies on cloud and large GPU clusters, which are costly and concentrated among a few tech giants. Tether’s QVAC platform supports meaningful large-scale model training on consumer hardware, including smartphones, demonstrating that advanced AI can be decentralized and highly inclusive. In the coming months, they will continue investing significant resources and funding to ensure AI is accessible anytime, anywhere on local devices.

This article: AI is no longer a patent of tech giants! Tether launches QVAC — the era of everyone having an LLM? Originally published on Chain News ABMedia.

View Original
Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.
Comment
0/400
No comments